commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 10
2.94k
| new_contents
stringlengths 21
3.18k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43k
| ndiff
stringlengths 51
3.32k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
8ac492be603f958a29bbc6bb5215d79ec469d269 | tests/test_init.py | tests/test_init.py | """Test the checkers"""
from nose.tools import ok_, eq_
from preflyt import check
CHECKERS = [
{"checker": "env", "name": "USER"}
]
BAD_CHECKERS = [
{"checker": "env", "name": "USER1231342dhkfgjhk2394dv09324jk12039csdfg01231"}
]
def test_everything():
"""Test the check method."""
good, results = check(CHECKERS)
ok_(good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
def test_everything_failure():
"""Test the check method."""
good, results = check(BAD_CHECKERS)
ok_(not good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
| """Test the checkers"""
from nose.tools import ok_, eq_
from preflyt import check
CHECKERS = [
{"checker": "env", "name": "PATH"}
]
BAD_CHECKERS = [
{"checker": "env", "name": "PATH1231342dhkfgjhk2394dv09324jk12039csdfg01231"}
]
def test_everything():
"""Test the check method."""
good, results = check(CHECKERS)
ok_(good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
def test_everything_failure():
"""Test the check method."""
good, results = check(BAD_CHECKERS)
ok_(not good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
| Update environment test to have cross platform support | Update environment test to have cross platform support
| Python | mit | humangeo/preflyt | """Test the checkers"""
from nose.tools import ok_, eq_
from preflyt import check
CHECKERS = [
- {"checker": "env", "name": "USER"}
+ {"checker": "env", "name": "PATH"}
]
BAD_CHECKERS = [
- {"checker": "env", "name": "USER1231342dhkfgjhk2394dv09324jk12039csdfg01231"}
+ {"checker": "env", "name": "PATH1231342dhkfgjhk2394dv09324jk12039csdfg01231"}
]
def test_everything():
"""Test the check method."""
good, results = check(CHECKERS)
ok_(good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
def test_everything_failure():
"""Test the check method."""
good, results = check(BAD_CHECKERS)
ok_(not good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
| Update environment test to have cross platform support | ## Code Before:
"""Test the checkers"""
from nose.tools import ok_, eq_
from preflyt import check
CHECKERS = [
{"checker": "env", "name": "USER"}
]
BAD_CHECKERS = [
{"checker": "env", "name": "USER1231342dhkfgjhk2394dv09324jk12039csdfg01231"}
]
def test_everything():
"""Test the check method."""
good, results = check(CHECKERS)
ok_(good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
def test_everything_failure():
"""Test the check method."""
good, results = check(BAD_CHECKERS)
ok_(not good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
## Instruction:
Update environment test to have cross platform support
## Code After:
"""Test the checkers"""
from nose.tools import ok_, eq_
from preflyt import check
CHECKERS = [
{"checker": "env", "name": "PATH"}
]
BAD_CHECKERS = [
{"checker": "env", "name": "PATH1231342dhkfgjhk2394dv09324jk12039csdfg01231"}
]
def test_everything():
"""Test the check method."""
good, results = check(CHECKERS)
ok_(good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
def test_everything_failure():
"""Test the check method."""
good, results = check(BAD_CHECKERS)
ok_(not good, results)
eq_(len(results), 1)
for field_name in ('check', 'success', 'message'):
ok_(field_name in results[0])
|
131fb74b0f399ad3abff5dcc2b09621cac1226e7 | config/nox_routing.py | config/nox_routing.py | from experiment_config_lib import ControllerConfig
from sts.control_flow import Fuzzer
from sts.input_traces.input_logger import InputLogger
from sts.invariant_checker import InvariantChecker
from sts.simulation_state import SimulationConfig
# Use NOX as our controller
command_line = "./nox_core -i ptcp:6633 routing"
controllers = [ControllerConfig(command_line, cwd="nox_classic/build/src", address="127.0.0.1", port=6633)]
dataplane_trace = "dataplane_traces/ping_pong_fat_tree.trace"
simulation_config = SimulationConfig(controller_configs=controllers,
dataplane_trace=dataplane_trace)
# Use a Fuzzer (already the default)
control_flow = Fuzzer(simulation_config, input_logger=InputLogger(),
check_interval=80,
invariant_check=InvariantChecker.check_connectivity)
| from experiment_config_lib import ControllerConfig
from sts.control_flow import Fuzzer
from sts.input_traces.input_logger import InputLogger
from sts.invariant_checker import InvariantChecker
from sts.simulation_state import SimulationConfig
from sts.topology import MeshTopology
# Use NOX as our controller
command_line = "./nox_core -v -i ptcp:6633 sample_routing"
controllers = [ControllerConfig(command_line, cwd="nox_classic/build/src", address="127.0.0.1", port=6633)]
topology_class = MeshTopology
topology_params = "num_switches=4"
dataplane_trace = "dataplane_traces/ping_pong_same_subnet_4_switches.trace"
# dataplane_trace = "dataplane_traces/ping_pong_fat_tree.trace"
simulation_config = SimulationConfig(controller_configs=controllers,
topology_class=topology_class,
topology_params=topology_params,
dataplane_trace=dataplane_trace)
#simulation_config = SimulationConfig(controller_configs=controllers,
# dataplane_trace=dataplane_trace)
# Use a Fuzzer (already the default)
control_flow = Fuzzer(simulation_config, input_logger=InputLogger(),
check_interval=80,
invariant_check=InvariantChecker.check_connectivity)
| Update NOX config to use sample_routing | Update NOX config to use sample_routing
| Python | apache-2.0 | ucb-sts/sts,jmiserez/sts,jmiserez/sts,ucb-sts/sts | from experiment_config_lib import ControllerConfig
from sts.control_flow import Fuzzer
from sts.input_traces.input_logger import InputLogger
from sts.invariant_checker import InvariantChecker
from sts.simulation_state import SimulationConfig
+ from sts.topology import MeshTopology
# Use NOX as our controller
- command_line = "./nox_core -i ptcp:6633 routing"
+ command_line = "./nox_core -v -i ptcp:6633 sample_routing"
controllers = [ControllerConfig(command_line, cwd="nox_classic/build/src", address="127.0.0.1", port=6633)]
+
+ topology_class = MeshTopology
+ topology_params = "num_switches=4"
+ dataplane_trace = "dataplane_traces/ping_pong_same_subnet_4_switches.trace"
- dataplane_trace = "dataplane_traces/ping_pong_fat_tree.trace"
+ # dataplane_trace = "dataplane_traces/ping_pong_fat_tree.trace"
simulation_config = SimulationConfig(controller_configs=controllers,
+ topology_class=topology_class,
+ topology_params=topology_params,
dataplane_trace=dataplane_trace)
+
+ #simulation_config = SimulationConfig(controller_configs=controllers,
+ # dataplane_trace=dataplane_trace)
# Use a Fuzzer (already the default)
control_flow = Fuzzer(simulation_config, input_logger=InputLogger(),
check_interval=80,
invariant_check=InvariantChecker.check_connectivity)
| Update NOX config to use sample_routing | ## Code Before:
from experiment_config_lib import ControllerConfig
from sts.control_flow import Fuzzer
from sts.input_traces.input_logger import InputLogger
from sts.invariant_checker import InvariantChecker
from sts.simulation_state import SimulationConfig
# Use NOX as our controller
command_line = "./nox_core -i ptcp:6633 routing"
controllers = [ControllerConfig(command_line, cwd="nox_classic/build/src", address="127.0.0.1", port=6633)]
dataplane_trace = "dataplane_traces/ping_pong_fat_tree.trace"
simulation_config = SimulationConfig(controller_configs=controllers,
dataplane_trace=dataplane_trace)
# Use a Fuzzer (already the default)
control_flow = Fuzzer(simulation_config, input_logger=InputLogger(),
check_interval=80,
invariant_check=InvariantChecker.check_connectivity)
## Instruction:
Update NOX config to use sample_routing
## Code After:
from experiment_config_lib import ControllerConfig
from sts.control_flow import Fuzzer
from sts.input_traces.input_logger import InputLogger
from sts.invariant_checker import InvariantChecker
from sts.simulation_state import SimulationConfig
from sts.topology import MeshTopology
# Use NOX as our controller
command_line = "./nox_core -v -i ptcp:6633 sample_routing"
controllers = [ControllerConfig(command_line, cwd="nox_classic/build/src", address="127.0.0.1", port=6633)]
topology_class = MeshTopology
topology_params = "num_switches=4"
dataplane_trace = "dataplane_traces/ping_pong_same_subnet_4_switches.trace"
# dataplane_trace = "dataplane_traces/ping_pong_fat_tree.trace"
simulation_config = SimulationConfig(controller_configs=controllers,
topology_class=topology_class,
topology_params=topology_params,
dataplane_trace=dataplane_trace)
#simulation_config = SimulationConfig(controller_configs=controllers,
# dataplane_trace=dataplane_trace)
# Use a Fuzzer (already the default)
control_flow = Fuzzer(simulation_config, input_logger=InputLogger(),
check_interval=80,
invariant_check=InvariantChecker.check_connectivity)
|
96e86fb389d67d55bf6b4e0f3f0f318e75b532dd | kirppu/management/commands/accounting_data.py | kirppu/management/commands/accounting_data.py | from django.core.management.base import BaseCommand
from django.utils.translation import activate
from kirppu.accounting import accounting_receipt
class Command(BaseCommand):
help = 'Dump accounting CSV to standard output'
def add_arguments(self, parser):
parser.add_argument('--lang', type=str, help="Change language, for example: en")
def handle(self, *args, **options):
if "lang" in options:
activate(options["lang"])
accounting_receipt(self.stdout)
| from django.core.management.base import BaseCommand
from django.utils.translation import activate
from kirppu.accounting import accounting_receipt
class Command(BaseCommand):
help = 'Dump accounting CSV to standard output'
def add_arguments(self, parser):
parser.add_argument('--lang', type=str, help="Change language, for example: en")
parser.add_argument('event', type=str, help="Event slug to dump data for")
def handle(self, *args, **options):
if "lang" in options:
activate(options["lang"])
from kirppu.models import Event
event = Event.objects.get(slug=options["event"])
accounting_receipt(self.stdout, event)
| Fix accounting data dump command. | Fix accounting data dump command.
| Python | mit | jlaunonen/kirppu,jlaunonen/kirppu,jlaunonen/kirppu,jlaunonen/kirppu | from django.core.management.base import BaseCommand
from django.utils.translation import activate
from kirppu.accounting import accounting_receipt
class Command(BaseCommand):
help = 'Dump accounting CSV to standard output'
def add_arguments(self, parser):
parser.add_argument('--lang', type=str, help="Change language, for example: en")
+ parser.add_argument('event', type=str, help="Event slug to dump data for")
def handle(self, *args, **options):
if "lang" in options:
activate(options["lang"])
- accounting_receipt(self.stdout)
+ from kirppu.models import Event
+ event = Event.objects.get(slug=options["event"])
+ accounting_receipt(self.stdout, event)
+ | Fix accounting data dump command. | ## Code Before:
from django.core.management.base import BaseCommand
from django.utils.translation import activate
from kirppu.accounting import accounting_receipt
class Command(BaseCommand):
help = 'Dump accounting CSV to standard output'
def add_arguments(self, parser):
parser.add_argument('--lang', type=str, help="Change language, for example: en")
def handle(self, *args, **options):
if "lang" in options:
activate(options["lang"])
accounting_receipt(self.stdout)
## Instruction:
Fix accounting data dump command.
## Code After:
from django.core.management.base import BaseCommand
from django.utils.translation import activate
from kirppu.accounting import accounting_receipt
class Command(BaseCommand):
help = 'Dump accounting CSV to standard output'
def add_arguments(self, parser):
parser.add_argument('--lang', type=str, help="Change language, for example: en")
parser.add_argument('event', type=str, help="Event slug to dump data for")
def handle(self, *args, **options):
if "lang" in options:
activate(options["lang"])
from kirppu.models import Event
event = Event.objects.get(slug=options["event"])
accounting_receipt(self.stdout, event)
|
ca75631f513c433c6024a2f00d045f304703a85d | webapp/tests/test_browser.py | webapp/tests/test_browser.py | import os
from django.core.urlresolvers import reverse
from django.test import TestCase
from django.test.utils import override_settings
from . import DATA_DIR
class BrowserTest(TestCase):
def test_browser(self):
url = reverse('graphite.browser.views.browser')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser')
def test_header(self):
url = reverse('graphite.browser.views.header')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser Header')
@override_settings(INDEX_FILE=os.path.join(DATA_DIR, 'index'))
def test_search(self):
url = reverse('graphite.browser.views.search')
response = self.client.post(url)
self.assertEqual(response.content, '')
# simple query
response = self.client.post(url, {'query': 'collectd'})
self.assertEqual(response.content.split(',')[0],
'collectd.test.df-root.df_complex-free')
# No match
response = self.client.post(url, {'query': 'other'})
self.assertEqual(response.content, '')
# Multiple terms (OR)
response = self.client.post(url, {'query': 'midterm shortterm'})
self.assertEqual(response.content.split(','),
['collectd.test.load.load.midterm',
'collectd.test.load.load.shortterm'])
| import os
from django.contrib.auth.models import User
from django.core.urlresolvers import reverse
from django.test import TestCase
from django.test.utils import override_settings
from . import DATA_DIR
class BrowserTest(TestCase):
def test_browser(self):
url = reverse('graphite.browser.views.browser')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser')
def test_header(self):
self.assertEqual(User.objects.count(), 0)
url = reverse('graphite.browser.views.header')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser Header')
# Graphite has created a default user
self.assertEqual(User.objects.get().username, 'default')
@override_settings(INDEX_FILE=os.path.join(DATA_DIR, 'index'))
def test_search(self):
url = reverse('graphite.browser.views.search')
response = self.client.post(url)
self.assertEqual(response.content, '')
# simple query
response = self.client.post(url, {'query': 'collectd'})
self.assertEqual(response.content.split(',')[0],
'collectd.test.df-root.df_complex-free')
# No match
response = self.client.post(url, {'query': 'other'})
self.assertEqual(response.content, '')
# Multiple terms (OR)
response = self.client.post(url, {'query': 'midterm shortterm'})
self.assertEqual(response.content.split(','),
['collectd.test.load.load.midterm',
'collectd.test.load.load.shortterm'])
| Add tests for making sure a user is dynamically created | Add tests for making sure a user is dynamically created
| Python | apache-2.0 | EinsamHauer/graphite-web-iow,bpaquet/graphite-web,Skyscanner/graphite-web,disqus/graphite-web,synedge/graphite-web,gwaldo/graphite-web,blacked/graphite-web,cosm0s/graphite-web,brutasse/graphite-web,blacked/graphite-web,bbc/graphite-web,gwaldo/graphite-web,Invoca/graphite-web,atnak/graphite-web,deniszh/graphite-web,dhtech/graphite-web,DanCech/graphite-web,Skyscanner/graphite-web,nkhuyu/graphite-web,jssjr/graphite-web,esnet/graphite-web,cgvarela/graphite-web,JeanFred/graphite-web,nkhuyu/graphite-web,graphite-project/graphite-web,bruce-lyft/graphite-web,bmhatfield/graphite-web,phreakocious/graphite-web,redice/graphite-web,EinsamHauer/graphite-web-iow,zBMNForks/graphite-web,kkdk5535/graphite-web,pu239ppy/graphite-web,krux/graphite-web,Invoca/graphite-web,jssjr/graphite-web,krux/graphite-web,edwardmlyte/graphite-web,Squarespace/graphite-web,EinsamHauer/graphite-web-iow,DanCech/graphite-web,bpaquet/graphite-web,phreakocious/graphite-web,bpaquet/graphite-web,cosm0s/graphite-web,axibase/graphite-web,lyft/graphite-web,atnak/graphite-web,graphite-server/graphite-web,axibase/graphite-web,Skyscanner/graphite-web,Invoca/graphite-web,kkdk5535/graphite-web,axibase/graphite-web,piotr1212/graphite-web,synedge/graphite-web,DanCech/graphite-web,SEJeff/graphite-web,SEJeff/graphite-web,graphite-server/graphite-web,pu239ppy/graphite-web,lfckop/graphite-web,cgvarela/graphite-web,DanCech/graphite-web,johnseekins/graphite-web,brutasse/graphite-web,Aloomaio/graphite-web,markolson/graphite-web,redice/graphite-web,penpen/graphite-web,axibase/graphite-web,cbowman0/graphite-web,bbc/graphite-web,mcoolive/graphite-web,lyft/graphite-web,nkhuyu/graphite-web,section-io/graphite-web,dhtech/graphite-web,ZelunZhang/graphite-web,bmhatfield/graphite-web,bpaquet/graphite-web,atnak/graphite-web,cgvarela/graphite-web,mcoolive/graphite-web,bmhatfield/graphite-web,edwardmlyte/graphite-web,nkhuyu/graphite-web,cosm0s/graphite-web,lyft/graphite-web,phreakocious/graphite-web,Squarespace/graphite-web,bbc/graphite-web,esnet/graphite-web,AICIDNN/graphite-web,dbn/graphite-web,JeanFred/graphite-web,JeanFred/graphite-web,SEJeff/graphite-web,markolson/graphite-web,Invoca/graphite-web,gwaldo/graphite-web,johnseekins/graphite-web,redice/graphite-web,goir/graphite-web,piotr1212/graphite-web,obfuscurity/graphite-web,cgvarela/graphite-web,bruce-lyft/graphite-web,jssjr/graphite-web,drax68/graphite-web,cosm0s/graphite-web,Invoca/graphite-web,mcoolive/graphite-web,g76r/graphite-web,kkdk5535/graphite-web,cybem/graphite-web-iow,cybem/graphite-web-iow,gwaldo/graphite-web,ZelunZhang/graphite-web,dhtech/graphite-web,Squarespace/graphite-web,phreakocious/graphite-web,criteo-forks/graphite-web,bmhatfield/graphite-web,Aloomaio/graphite-web,lfckop/graphite-web,johnseekins/graphite-web,atnak/graphite-web,SEJeff/graphite-web,section-io/graphite-web,edwardmlyte/graphite-web,bruce-lyft/graphite-web,drax68/graphite-web,phreakocious/graphite-web,redice/graphite-web,nkhuyu/graphite-web,deniszh/graphite-web,bbc/graphite-web,redice/graphite-web,Skyscanner/graphite-web,EinsamHauer/graphite-web-iow,JeanFred/graphite-web,johnseekins/graphite-web,goir/graphite-web,blacked/graphite-web,graphite-project/graphite-web,brutasse/graphite-web,graphite-project/graphite-web,zBMNForks/graphite-web,bpaquet/graphite-web,obfuscurity/graphite-web,piotr1212/graphite-web,synedge/graphite-web,criteo-forks/graphite-web,DanCech/graphite-web,deniszh/graphite-web,bpaquet/graphite-web,disqus/graphite-web,markolson/graphite-web,g76r/graphite-web,axibase/graphite-web,edwardmlyte/graphite-web,jssjr/graphite-web,graphite-server/graphite-web,criteo-forks/graphite-web,zBMNForks/graphite-web,dhtech/graphite-web,phreakocious/graphite-web,cybem/graphite-web-iow,AICIDNN/graphite-web,esnet/graphite-web,g76r/graphite-web,synedge/graphite-web,dbn/graphite-web,criteo-forks/graphite-web,Skyscanner/graphite-web,section-io/graphite-web,johnseekins/graphite-web,obfuscurity/graphite-web,JeanFred/graphite-web,DanCech/graphite-web,pu239ppy/graphite-web,lyft/graphite-web,section-io/graphite-web,krux/graphite-web,cosm0s/graphite-web,cbowman0/graphite-web,drax68/graphite-web,lfckop/graphite-web,lyft/graphite-web,goir/graphite-web,Squarespace/graphite-web,Aloomaio/graphite-web,blacked/graphite-web,lfckop/graphite-web,graphite-project/graphite-web,goir/graphite-web,synedge/graphite-web,Invoca/graphite-web,atnak/graphite-web,dbn/graphite-web,zBMNForks/graphite-web,bbc/graphite-web,goir/graphite-web,bmhatfield/graphite-web,g76r/graphite-web,disqus/graphite-web,krux/graphite-web,deniszh/graphite-web,dhtech/graphite-web,atnak/graphite-web,cybem/graphite-web-iow,dbn/graphite-web,disqus/graphite-web,johnseekins/graphite-web,mcoolive/graphite-web,AICIDNN/graphite-web,cgvarela/graphite-web,piotr1212/graphite-web,penpen/graphite-web,synedge/graphite-web,mcoolive/graphite-web,criteo-forks/graphite-web,zBMNForks/graphite-web,bruce-lyft/graphite-web,markolson/graphite-web,ZelunZhang/graphite-web,cbowman0/graphite-web,g76r/graphite-web,goir/graphite-web,penpen/graphite-web,lyft/graphite-web,JeanFred/graphite-web,cybem/graphite-web-iow,lfckop/graphite-web,ZelunZhang/graphite-web,cbowman0/graphite-web,gwaldo/graphite-web,cybem/graphite-web-iow,Squarespace/graphite-web,AICIDNN/graphite-web,SEJeff/graphite-web,kkdk5535/graphite-web,graphite-project/graphite-web,brutasse/graphite-web,piotr1212/graphite-web,cbowman0/graphite-web,krux/graphite-web,AICIDNN/graphite-web,esnet/graphite-web,Squarespace/graphite-web,deniszh/graphite-web,graphite-server/graphite-web,ZelunZhang/graphite-web,Skyscanner/graphite-web,blacked/graphite-web,dbn/graphite-web,kkdk5535/graphite-web,mcoolive/graphite-web,graphite-project/graphite-web,criteo-forks/graphite-web,EinsamHauer/graphite-web-iow,edwardmlyte/graphite-web,disqus/graphite-web,bruce-lyft/graphite-web,g76r/graphite-web,Aloomaio/graphite-web,drax68/graphite-web,pu239ppy/graphite-web,edwardmlyte/graphite-web,dbn/graphite-web,deniszh/graphite-web,drax68/graphite-web,lfckop/graphite-web,penpen/graphite-web,nkhuyu/graphite-web,section-io/graphite-web,pu239ppy/graphite-web,ZelunZhang/graphite-web,piotr1212/graphite-web,penpen/graphite-web,zBMNForks/graphite-web,obfuscurity/graphite-web,bruce-lyft/graphite-web,graphite-server/graphite-web,section-io/graphite-web,bmhatfield/graphite-web,jssjr/graphite-web,kkdk5535/graphite-web,graphite-server/graphite-web,esnet/graphite-web,cosm0s/graphite-web,obfuscurity/graphite-web,axibase/graphite-web,penpen/graphite-web,markolson/graphite-web,jssjr/graphite-web,Aloomaio/graphite-web,Aloomaio/graphite-web,cbowman0/graphite-web,pu239ppy/graphite-web,krux/graphite-web,disqus/graphite-web,brutasse/graphite-web,obfuscurity/graphite-web,redice/graphite-web,EinsamHauer/graphite-web-iow,AICIDNN/graphite-web,blacked/graphite-web,gwaldo/graphite-web,brutasse/graphite-web,cgvarela/graphite-web,drax68/graphite-web | import os
+ from django.contrib.auth.models import User
from django.core.urlresolvers import reverse
from django.test import TestCase
from django.test.utils import override_settings
from . import DATA_DIR
class BrowserTest(TestCase):
def test_browser(self):
url = reverse('graphite.browser.views.browser')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser')
def test_header(self):
+ self.assertEqual(User.objects.count(), 0)
url = reverse('graphite.browser.views.header')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser Header')
+
+ # Graphite has created a default user
+ self.assertEqual(User.objects.get().username, 'default')
@override_settings(INDEX_FILE=os.path.join(DATA_DIR, 'index'))
def test_search(self):
url = reverse('graphite.browser.views.search')
response = self.client.post(url)
self.assertEqual(response.content, '')
# simple query
response = self.client.post(url, {'query': 'collectd'})
self.assertEqual(response.content.split(',')[0],
'collectd.test.df-root.df_complex-free')
# No match
response = self.client.post(url, {'query': 'other'})
self.assertEqual(response.content, '')
# Multiple terms (OR)
response = self.client.post(url, {'query': 'midterm shortterm'})
self.assertEqual(response.content.split(','),
['collectd.test.load.load.midterm',
'collectd.test.load.load.shortterm'])
| Add tests for making sure a user is dynamically created | ## Code Before:
import os
from django.core.urlresolvers import reverse
from django.test import TestCase
from django.test.utils import override_settings
from . import DATA_DIR
class BrowserTest(TestCase):
def test_browser(self):
url = reverse('graphite.browser.views.browser')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser')
def test_header(self):
url = reverse('graphite.browser.views.header')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser Header')
@override_settings(INDEX_FILE=os.path.join(DATA_DIR, 'index'))
def test_search(self):
url = reverse('graphite.browser.views.search')
response = self.client.post(url)
self.assertEqual(response.content, '')
# simple query
response = self.client.post(url, {'query': 'collectd'})
self.assertEqual(response.content.split(',')[0],
'collectd.test.df-root.df_complex-free')
# No match
response = self.client.post(url, {'query': 'other'})
self.assertEqual(response.content, '')
# Multiple terms (OR)
response = self.client.post(url, {'query': 'midterm shortterm'})
self.assertEqual(response.content.split(','),
['collectd.test.load.load.midterm',
'collectd.test.load.load.shortterm'])
## Instruction:
Add tests for making sure a user is dynamically created
## Code After:
import os
from django.contrib.auth.models import User
from django.core.urlresolvers import reverse
from django.test import TestCase
from django.test.utils import override_settings
from . import DATA_DIR
class BrowserTest(TestCase):
def test_browser(self):
url = reverse('graphite.browser.views.browser')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser')
def test_header(self):
self.assertEqual(User.objects.count(), 0)
url = reverse('graphite.browser.views.header')
response = self.client.get(url)
self.assertContains(response, 'Graphite Browser Header')
# Graphite has created a default user
self.assertEqual(User.objects.get().username, 'default')
@override_settings(INDEX_FILE=os.path.join(DATA_DIR, 'index'))
def test_search(self):
url = reverse('graphite.browser.views.search')
response = self.client.post(url)
self.assertEqual(response.content, '')
# simple query
response = self.client.post(url, {'query': 'collectd'})
self.assertEqual(response.content.split(',')[0],
'collectd.test.df-root.df_complex-free')
# No match
response = self.client.post(url, {'query': 'other'})
self.assertEqual(response.content, '')
# Multiple terms (OR)
response = self.client.post(url, {'query': 'midterm shortterm'})
self.assertEqual(response.content.split(','),
['collectd.test.load.load.midterm',
'collectd.test.load.load.shortterm'])
|
20bd5c16d5850f988e92c39db3ff041c37c83b73 | contract_sale_generation/models/abstract_contract.py | contract_sale_generation/models/abstract_contract.py |
from odoo import api, fields, models
class ContractAbstractContract(models.AbstractModel):
_inherit = "contract.abstract.contract"
sale_autoconfirm = fields.Boolean(string="Sale Autoconfirm")
@api.model
def _get_generation_type_selection(self):
res = super()._get_generation_type_selection()
res.append(("sale", "Sale"))
return res
|
from odoo import api, fields, models
class ContractAbstractContract(models.AbstractModel):
_inherit = "contract.abstract.contract"
sale_autoconfirm = fields.Boolean(string="Sale Autoconfirm")
@api.model
def _selection_generation_type(self):
res = super()._selection_generation_type()
res.append(("sale", "Sale"))
return res
| Align method on Odoo conventions | [14.0][IMP] contract_sale_generation: Align method on Odoo conventions
| Python | agpl-3.0 | OCA/contract,OCA/contract,OCA/contract |
from odoo import api, fields, models
class ContractAbstractContract(models.AbstractModel):
_inherit = "contract.abstract.contract"
sale_autoconfirm = fields.Boolean(string="Sale Autoconfirm")
@api.model
- def _get_generation_type_selection(self):
+ def _selection_generation_type(self):
- res = super()._get_generation_type_selection()
+ res = super()._selection_generation_type()
res.append(("sale", "Sale"))
return res
| Align method on Odoo conventions | ## Code Before:
from odoo import api, fields, models
class ContractAbstractContract(models.AbstractModel):
_inherit = "contract.abstract.contract"
sale_autoconfirm = fields.Boolean(string="Sale Autoconfirm")
@api.model
def _get_generation_type_selection(self):
res = super()._get_generation_type_selection()
res.append(("sale", "Sale"))
return res
## Instruction:
Align method on Odoo conventions
## Code After:
from odoo import api, fields, models
class ContractAbstractContract(models.AbstractModel):
_inherit = "contract.abstract.contract"
sale_autoconfirm = fields.Boolean(string="Sale Autoconfirm")
@api.model
def _selection_generation_type(self):
res = super()._selection_generation_type()
res.append(("sale", "Sale"))
return res
|
6422f6057d43dfb5259028291991f39c5b81b446 | spreadflow_core/flow.py | spreadflow_core/flow.py | from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from collections import defaultdict
class Flowmap(dict):
def __init__(self):
super(Flowmap, self).__init__()
self.decorators = []
self.annotations = {}
def graph(self):
result = defaultdict(set)
backlog = set()
processed = set()
for port_out, port_in in self.iteritems():
result[port_out].add(port_in)
backlog.add(port_in)
while len(backlog):
node = backlog.pop()
if node in processed:
continue
else:
processed.add(node)
try:
arcs = tuple(node.dependencies)
except AttributeError:
continue
for port_out, port_in in arcs:
result[port_out].add(port_in)
backlog.add(port_out)
backlog.add(port_in)
return result
| from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from collections import defaultdict, MutableMapping
class Flowmap(MutableMapping):
def __init__(self):
super(Flowmap, self).__init__()
self.annotations = {}
self.connections = {}
self.decorators = []
def __getitem__(self, key):
return self.connections[key]
def __setitem__(self, key, value):
self.connections[key] = value
def __delitem__(self, key):
del self.connections[key]
def __iter__(self):
return iter(self.connections)
def __len__(self):
return len(self.connections)
def graph(self):
result = defaultdict(set)
backlog = set()
processed = set()
for port_out, port_in in self.iteritems():
result[port_out].add(port_in)
backlog.add(port_in)
while len(backlog):
node = backlog.pop()
if node in processed:
continue
else:
processed.add(node)
try:
arcs = tuple(node.dependencies)
except AttributeError:
continue
for port_out, port_in in arcs:
result[port_out].add(port_in)
backlog.add(port_out)
backlog.add(port_in)
return result
| Refactor Flowmap into a MutableMapping | Refactor Flowmap into a MutableMapping
| Python | mit | spreadflow/spreadflow-core,znerol/spreadflow-core | from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
- from collections import defaultdict
+ from collections import defaultdict, MutableMapping
- class Flowmap(dict):
+ class Flowmap(MutableMapping):
def __init__(self):
super(Flowmap, self).__init__()
+ self.annotations = {}
+ self.connections = {}
self.decorators = []
- self.annotations = {}
+
+ def __getitem__(self, key):
+ return self.connections[key]
+
+ def __setitem__(self, key, value):
+ self.connections[key] = value
+
+ def __delitem__(self, key):
+ del self.connections[key]
+
+ def __iter__(self):
+ return iter(self.connections)
+
+ def __len__(self):
+ return len(self.connections)
def graph(self):
result = defaultdict(set)
backlog = set()
processed = set()
for port_out, port_in in self.iteritems():
result[port_out].add(port_in)
backlog.add(port_in)
while len(backlog):
node = backlog.pop()
if node in processed:
continue
else:
processed.add(node)
try:
arcs = tuple(node.dependencies)
except AttributeError:
continue
for port_out, port_in in arcs:
result[port_out].add(port_in)
backlog.add(port_out)
backlog.add(port_in)
return result
| Refactor Flowmap into a MutableMapping | ## Code Before:
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from collections import defaultdict
class Flowmap(dict):
def __init__(self):
super(Flowmap, self).__init__()
self.decorators = []
self.annotations = {}
def graph(self):
result = defaultdict(set)
backlog = set()
processed = set()
for port_out, port_in in self.iteritems():
result[port_out].add(port_in)
backlog.add(port_in)
while len(backlog):
node = backlog.pop()
if node in processed:
continue
else:
processed.add(node)
try:
arcs = tuple(node.dependencies)
except AttributeError:
continue
for port_out, port_in in arcs:
result[port_out].add(port_in)
backlog.add(port_out)
backlog.add(port_in)
return result
## Instruction:
Refactor Flowmap into a MutableMapping
## Code After:
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from collections import defaultdict, MutableMapping
class Flowmap(MutableMapping):
def __init__(self):
super(Flowmap, self).__init__()
self.annotations = {}
self.connections = {}
self.decorators = []
def __getitem__(self, key):
return self.connections[key]
def __setitem__(self, key, value):
self.connections[key] = value
def __delitem__(self, key):
del self.connections[key]
def __iter__(self):
return iter(self.connections)
def __len__(self):
return len(self.connections)
def graph(self):
result = defaultdict(set)
backlog = set()
processed = set()
for port_out, port_in in self.iteritems():
result[port_out].add(port_in)
backlog.add(port_in)
while len(backlog):
node = backlog.pop()
if node in processed:
continue
else:
processed.add(node)
try:
arcs = tuple(node.dependencies)
except AttributeError:
continue
for port_out, port_in in arcs:
result[port_out].add(port_in)
backlog.add(port_out)
backlog.add(port_in)
return result
|
5beb443d4c9cf834be03ff33a2fb01605f8feb80 | pyof/v0x01/symmetric/hello.py | pyof/v0x01/symmetric/hello.py | """Defines Hello message."""
# System imports
# Third-party imports
from pyof.foundation.base import GenericMessage
from pyof.v0x01.common.header import Header, Type
__all__ = ('Hello',)
# Classes
class Hello(GenericMessage):
"""OpenFlow Hello Message.
This message does not contain a body beyond the OpenFlow Header.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
| """Defines Hello message."""
# System imports
# Third-party imports
from pyof.foundation.base import GenericMessage
from pyof.foundation.basic_types import BinaryData
from pyof.v0x01.common.header import Header, Type
__all__ = ('Hello',)
# Classes
class Hello(GenericMessage):
"""OpenFlow Hello Message.
This message does not contain a body beyond the OpenFlow Header.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
elements = BinaryData()
| Add optional elements in v0x01 Hello | Add optional elements in v0x01 Hello
For spec compliance. Ignore the elements as they're not used.
Fix #379
| Python | mit | kytos/python-openflow | """Defines Hello message."""
# System imports
# Third-party imports
from pyof.foundation.base import GenericMessage
+ from pyof.foundation.basic_types import BinaryData
from pyof.v0x01.common.header import Header, Type
__all__ = ('Hello',)
# Classes
class Hello(GenericMessage):
"""OpenFlow Hello Message.
This message does not contain a body beyond the OpenFlow Header.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
+ elements = BinaryData()
| Add optional elements in v0x01 Hello | ## Code Before:
"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.foundation.base import GenericMessage
from pyof.v0x01.common.header import Header, Type
__all__ = ('Hello',)
# Classes
class Hello(GenericMessage):
"""OpenFlow Hello Message.
This message does not contain a body beyond the OpenFlow Header.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
## Instruction:
Add optional elements in v0x01 Hello
## Code After:
"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.foundation.base import GenericMessage
from pyof.foundation.basic_types import BinaryData
from pyof.v0x01.common.header import Header, Type
__all__ = ('Hello',)
# Classes
class Hello(GenericMessage):
"""OpenFlow Hello Message.
This message does not contain a body beyond the OpenFlow Header.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
elements = BinaryData()
|
015d536e591d5af7e93f299e84504fe8a17f76b3 | tests.py | tests.py | import logging
import unittest
from StringIO import StringIO
class TestArgParsing(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
from script import parseargs
self.parseargs = parseargs
def test_parseargs(self):
opts, args = self.parseargs(["foo"])
self.assertEqual(opts.silent, False)
self.assertEqual(args, [])
class TestMain(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
root = logging.getLogger()
buffer = logging.handlers.BufferingHandler(100)
root.addHandler(buffer)
self.buffer = buffer.buffer
self.out = StringIO()
self.err = StringIO()
def main(self, *args, **kwargs):
from script import main
_kwargs = {
"out": self.out,
"err": self.err,
}
_kwargs.update(kwargs)
return main(*args, **_kwargs)
def test_main(self):
result = self.main(["foo"])
self.assertEqual(result, None)
self.assertEqual(self.buffer, [])
def test_main_verbose(self):
result = self.main(["foo", "-vv"])
self.assertEqual(result, None)
self.assertEqual(len(self.buffer), 1)
self.assertEqual(self.buffer[0].msg, "Ready to run")
| import logging
import unittest
from StringIO import StringIO
class TestArgParsing(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
from script import parseargs
self.parseargs = parseargs
def test_parseargs(self):
opts, args = self.parseargs(["foo"])
self.assertEqual(opts.silent, False)
self.assertEqual(args, [])
class TestMain(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
root = logging.getLogger()
buffer = logging.handlers.BufferingHandler(100)
root.addHandler(buffer)
self.buffer = buffer.buffer
self.out = StringIO()
self.err = StringIO()
def main(self, *args, **kwargs):
from script import main
_kwargs = {
"out": self.out,
"err": self.err,
}
_kwargs.update(kwargs)
return main(*args, **_kwargs)
def test_main(self):
result = self.main(["foo"])
self.assertEqual(result, None)
self.assertEqual(self.buffer, [])
def test_main_verbose(self):
result = self.main(["foo", "-vv"])
self.assertEqual(result, None)
self.assertEqual(len(self.buffer), 1)
self.assertEqual(self.buffer[0].msg, "Ready to run")
self.assertTrue("Ready to run" in self.err.getvalue())
| Test that log messages go to stderr. | Test that log messages go to stderr.
| Python | isc | whilp/python-script,whilp/python-script | import logging
import unittest
from StringIO import StringIO
class TestArgParsing(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
from script import parseargs
self.parseargs = parseargs
def test_parseargs(self):
opts, args = self.parseargs(["foo"])
self.assertEqual(opts.silent, False)
self.assertEqual(args, [])
class TestMain(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
root = logging.getLogger()
buffer = logging.handlers.BufferingHandler(100)
root.addHandler(buffer)
self.buffer = buffer.buffer
self.out = StringIO()
self.err = StringIO()
def main(self, *args, **kwargs):
from script import main
_kwargs = {
"out": self.out,
"err": self.err,
}
_kwargs.update(kwargs)
return main(*args, **_kwargs)
def test_main(self):
result = self.main(["foo"])
self.assertEqual(result, None)
self.assertEqual(self.buffer, [])
def test_main_verbose(self):
result = self.main(["foo", "-vv"])
self.assertEqual(result, None)
self.assertEqual(len(self.buffer), 1)
self.assertEqual(self.buffer[0].msg, "Ready to run")
+ self.assertTrue("Ready to run" in self.err.getvalue())
| Test that log messages go to stderr. | ## Code Before:
import logging
import unittest
from StringIO import StringIO
class TestArgParsing(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
from script import parseargs
self.parseargs = parseargs
def test_parseargs(self):
opts, args = self.parseargs(["foo"])
self.assertEqual(opts.silent, False)
self.assertEqual(args, [])
class TestMain(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
root = logging.getLogger()
buffer = logging.handlers.BufferingHandler(100)
root.addHandler(buffer)
self.buffer = buffer.buffer
self.out = StringIO()
self.err = StringIO()
def main(self, *args, **kwargs):
from script import main
_kwargs = {
"out": self.out,
"err": self.err,
}
_kwargs.update(kwargs)
return main(*args, **_kwargs)
def test_main(self):
result = self.main(["foo"])
self.assertEqual(result, None)
self.assertEqual(self.buffer, [])
def test_main_verbose(self):
result = self.main(["foo", "-vv"])
self.assertEqual(result, None)
self.assertEqual(len(self.buffer), 1)
self.assertEqual(self.buffer[0].msg, "Ready to run")
## Instruction:
Test that log messages go to stderr.
## Code After:
import logging
import unittest
from StringIO import StringIO
class TestArgParsing(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
from script import parseargs
self.parseargs = parseargs
def test_parseargs(self):
opts, args = self.parseargs(["foo"])
self.assertEqual(opts.silent, False)
self.assertEqual(args, [])
class TestMain(unittest.TestCase):
def setUp(self):
unittest.TestCase.setUp(self)
root = logging.getLogger()
buffer = logging.handlers.BufferingHandler(100)
root.addHandler(buffer)
self.buffer = buffer.buffer
self.out = StringIO()
self.err = StringIO()
def main(self, *args, **kwargs):
from script import main
_kwargs = {
"out": self.out,
"err": self.err,
}
_kwargs.update(kwargs)
return main(*args, **_kwargs)
def test_main(self):
result = self.main(["foo"])
self.assertEqual(result, None)
self.assertEqual(self.buffer, [])
def test_main_verbose(self):
result = self.main(["foo", "-vv"])
self.assertEqual(result, None)
self.assertEqual(len(self.buffer), 1)
self.assertEqual(self.buffer[0].msg, "Ready to run")
self.assertTrue("Ready to run" in self.err.getvalue())
|
bc6001d6c25bdb5d83830e5a65fe5aea9fc1eb99 | ume/cmd.py | ume/cmd.py | import logging as l
import argparse
from ume.utils import (
save_mat,
dynamic_load,
)
def parse_args():
p = argparse.ArgumentParser(
description='CLI interface UME')
p.add_argument('--config', dest='inifile', default='config.ini')
subparsers = p.add_subparsers(
dest='subparser_name',
help='sub-commands for instant action')
f_parser = subparsers.add_parser('feature')
f_parser.add_argument('-n', '--name', type=str, required=True)
subparsers.add_parser('validation')
subparsers.add_parser('prediction')
return p.parse_args()
def run_feature(args):
klass = dynamic_load(args.name)
result = klass()
save_mat(args.name, result)
def main():
l.basicConfig(format='%(asctime)s %(message)s', level=l.INFO)
args = parse_args()
if args.subparser_name == 'validate':
pass
elif args.subparser_name == 'predict':
pass
elif args.subparser_name == 'feature':
run_feature(args)
else:
raise RuntimeError("No such sub-command.")
| import logging as l
import argparse
import os
from ume.utils import (
save_mat,
dynamic_load,
)
def parse_args():
p = argparse.ArgumentParser(
description='CLI interface UME')
p.add_argument('--config', dest='inifile', default='config.ini')
subparsers = p.add_subparsers(
dest='subparser_name',
help='sub-commands for instant action')
f_parser = subparsers.add_parser('feature')
f_parser.add_argument('-n', '--name', type=str, required=True)
i_parser = subparsers.add_parser('init')
subparsers.add_parser('validation')
subparsers.add_parser('prediction')
return p.parse_args()
def run_feature(args):
klass = dynamic_load(args.name)
result = klass()
save_mat(args.name, result)
def run_initialize(args):
pwd = os.getcwd()
os.makedirs(os.path.join(pwd, "data/input"))
os.makedirs(os.path.join(pwd, "data/output"))
os.makedirs(os.path.join(pwd, "data/working"))
os.makedirs(os.path.join(pwd, "note"))
os.makedirs(os.path.join(pwd, "trunk"))
def main():
l.basicConfig(format='%(asctime)s %(message)s', level=l.INFO)
args = parse_args()
if args.subparser_name == 'validate':
pass
elif args.subparser_name == 'predict':
pass
elif args.subparser_name == 'feature':
run_feature(args)
elif args.subparser_name == 'init':
run_initialize(args)
else:
raise RuntimeError("No such sub-command.")
| Add init function to create directories | Add init function to create directories
| Python | mit | smly/ume,smly/ume,smly/ume,smly/ume | import logging as l
import argparse
+ import os
from ume.utils import (
save_mat,
dynamic_load,
)
def parse_args():
p = argparse.ArgumentParser(
description='CLI interface UME')
p.add_argument('--config', dest='inifile', default='config.ini')
subparsers = p.add_subparsers(
dest='subparser_name',
help='sub-commands for instant action')
f_parser = subparsers.add_parser('feature')
f_parser.add_argument('-n', '--name', type=str, required=True)
+ i_parser = subparsers.add_parser('init')
subparsers.add_parser('validation')
subparsers.add_parser('prediction')
return p.parse_args()
def run_feature(args):
klass = dynamic_load(args.name)
result = klass()
save_mat(args.name, result)
+ def run_initialize(args):
+ pwd = os.getcwd()
+ os.makedirs(os.path.join(pwd, "data/input"))
+ os.makedirs(os.path.join(pwd, "data/output"))
+ os.makedirs(os.path.join(pwd, "data/working"))
+ os.makedirs(os.path.join(pwd, "note"))
+ os.makedirs(os.path.join(pwd, "trunk"))
+
+
def main():
l.basicConfig(format='%(asctime)s %(message)s', level=l.INFO)
args = parse_args()
if args.subparser_name == 'validate':
pass
elif args.subparser_name == 'predict':
pass
elif args.subparser_name == 'feature':
run_feature(args)
+ elif args.subparser_name == 'init':
+ run_initialize(args)
else:
raise RuntimeError("No such sub-command.")
| Add init function to create directories | ## Code Before:
import logging as l
import argparse
from ume.utils import (
save_mat,
dynamic_load,
)
def parse_args():
p = argparse.ArgumentParser(
description='CLI interface UME')
p.add_argument('--config', dest='inifile', default='config.ini')
subparsers = p.add_subparsers(
dest='subparser_name',
help='sub-commands for instant action')
f_parser = subparsers.add_parser('feature')
f_parser.add_argument('-n', '--name', type=str, required=True)
subparsers.add_parser('validation')
subparsers.add_parser('prediction')
return p.parse_args()
def run_feature(args):
klass = dynamic_load(args.name)
result = klass()
save_mat(args.name, result)
def main():
l.basicConfig(format='%(asctime)s %(message)s', level=l.INFO)
args = parse_args()
if args.subparser_name == 'validate':
pass
elif args.subparser_name == 'predict':
pass
elif args.subparser_name == 'feature':
run_feature(args)
else:
raise RuntimeError("No such sub-command.")
## Instruction:
Add init function to create directories
## Code After:
import logging as l
import argparse
import os
from ume.utils import (
save_mat,
dynamic_load,
)
def parse_args():
p = argparse.ArgumentParser(
description='CLI interface UME')
p.add_argument('--config', dest='inifile', default='config.ini')
subparsers = p.add_subparsers(
dest='subparser_name',
help='sub-commands for instant action')
f_parser = subparsers.add_parser('feature')
f_parser.add_argument('-n', '--name', type=str, required=True)
i_parser = subparsers.add_parser('init')
subparsers.add_parser('validation')
subparsers.add_parser('prediction')
return p.parse_args()
def run_feature(args):
klass = dynamic_load(args.name)
result = klass()
save_mat(args.name, result)
def run_initialize(args):
pwd = os.getcwd()
os.makedirs(os.path.join(pwd, "data/input"))
os.makedirs(os.path.join(pwd, "data/output"))
os.makedirs(os.path.join(pwd, "data/working"))
os.makedirs(os.path.join(pwd, "note"))
os.makedirs(os.path.join(pwd, "trunk"))
def main():
l.basicConfig(format='%(asctime)s %(message)s', level=l.INFO)
args = parse_args()
if args.subparser_name == 'validate':
pass
elif args.subparser_name == 'predict':
pass
elif args.subparser_name == 'feature':
run_feature(args)
elif args.subparser_name == 'init':
run_initialize(args)
else:
raise RuntimeError("No such sub-command.")
|
1b75fe13249eba8d951735ad422dc512b1e28caf | test/test_all_stores.py | test/test_all_stores.py | import store_fixture
import groundstation.store
class TestGitStore(store_fixture.StoreTestCase):
storeClass = groundstation.store.git_store.GitStore
| import os
import store_fixture
import groundstation.store
class TestGitStore(store_fixture.StoreTestCase):
storeClass = groundstation.store.git_store.GitStore
def test_creates_required_dirs(self):
for d in groundstation.store.git_store.GitStore.required_dirs:
path = os.path.join(self.path, d)
self.assertTrue(os.path.exists(path))
self.assertTrue(os.path.isdir(path))
| Add testcase for database initialization | Add testcase for database initialization
| Python | mit | richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation | + import os
+
import store_fixture
import groundstation.store
class TestGitStore(store_fixture.StoreTestCase):
storeClass = groundstation.store.git_store.GitStore
+ def test_creates_required_dirs(self):
+ for d in groundstation.store.git_store.GitStore.required_dirs:
+ path = os.path.join(self.path, d)
+ self.assertTrue(os.path.exists(path))
+ self.assertTrue(os.path.isdir(path))
+ | Add testcase for database initialization | ## Code Before:
import store_fixture
import groundstation.store
class TestGitStore(store_fixture.StoreTestCase):
storeClass = groundstation.store.git_store.GitStore
## Instruction:
Add testcase for database initialization
## Code After:
import os
import store_fixture
import groundstation.store
class TestGitStore(store_fixture.StoreTestCase):
storeClass = groundstation.store.git_store.GitStore
def test_creates_required_dirs(self):
for d in groundstation.store.git_store.GitStore.required_dirs:
path = os.path.join(self.path, d)
self.assertTrue(os.path.exists(path))
self.assertTrue(os.path.isdir(path))
|
2f9a4029e909f71539f3b7326b867e27386c3378 | tests/interface_test.py | tests/interface_test.py | import unittest
import aiozmq
class ZmqTransportTests(unittest.TestCase):
def test_interface(self):
tr = aiozmq.ZmqTransport()
self.assertRaises(NotImplementedError, tr.write, [b'data'])
self.assertRaises(NotImplementedError, tr.abort)
self.assertRaises(NotImplementedError, tr.getsockopt, 1)
self.assertRaises(NotImplementedError, tr.setsockopt, 1, 2)
self.assertRaises(NotImplementedError, tr.set_write_buffer_limits)
self.assertRaises(NotImplementedError, tr.get_write_buffer_size)
self.assertRaises(NotImplementedError, tr.bind, 'endpoint')
self.assertRaises(NotImplementedError, tr.unbind, 'endpoint')
self.assertRaises(NotImplementedError, tr.bindings)
self.assertRaises(NotImplementedError, tr.connect, 'endpoint')
self.assertRaises(NotImplementedError, tr.disconnect, 'endpoint')
self.assertRaises(NotImplementedError, tr.connections)
self.assertRaises(NotImplementedError, tr.subscribe, b'filter')
self.assertRaises(NotImplementedError, tr.unsubscribe, b'filter')
self.assertRaises(NotImplementedError, tr.subscriptions)
class ZmqProtocolTests(unittest.TestCase):
def test_interface(self):
pr = aiozmq.ZmqProtocol()
self.assertIsNone(pr.msg_received((b'data',)))
| import unittest
import aiozmq
class ZmqTransportTests(unittest.TestCase):
def test_interface(self):
tr = aiozmq.ZmqTransport()
self.assertRaises(NotImplementedError, tr.write, [b'data'])
self.assertRaises(NotImplementedError, tr.abort)
self.assertRaises(NotImplementedError, tr.getsockopt, 1)
self.assertRaises(NotImplementedError, tr.setsockopt, 1, 2)
self.assertRaises(NotImplementedError, tr.set_write_buffer_limits)
self.assertRaises(NotImplementedError, tr.get_write_buffer_size)
self.assertRaises(NotImplementedError, tr.pause_reading)
self.assertRaises(NotImplementedError, tr.resume_reading)
self.assertRaises(NotImplementedError, tr.bind, 'endpoint')
self.assertRaises(NotImplementedError, tr.unbind, 'endpoint')
self.assertRaises(NotImplementedError, tr.bindings)
self.assertRaises(NotImplementedError, tr.connect, 'endpoint')
self.assertRaises(NotImplementedError, tr.disconnect, 'endpoint')
self.assertRaises(NotImplementedError, tr.connections)
self.assertRaises(NotImplementedError, tr.subscribe, b'filter')
self.assertRaises(NotImplementedError, tr.unsubscribe, b'filter')
self.assertRaises(NotImplementedError, tr.subscriptions)
class ZmqProtocolTests(unittest.TestCase):
def test_interface(self):
pr = aiozmq.ZmqProtocol()
self.assertIsNone(pr.msg_received((b'data',)))
| Add missing tests for interfaces | Add missing tests for interfaces
| Python | bsd-2-clause | MetaMemoryT/aiozmq,claws/aiozmq,asteven/aiozmq,aio-libs/aiozmq | import unittest
import aiozmq
class ZmqTransportTests(unittest.TestCase):
def test_interface(self):
tr = aiozmq.ZmqTransport()
self.assertRaises(NotImplementedError, tr.write, [b'data'])
self.assertRaises(NotImplementedError, tr.abort)
self.assertRaises(NotImplementedError, tr.getsockopt, 1)
self.assertRaises(NotImplementedError, tr.setsockopt, 1, 2)
self.assertRaises(NotImplementedError, tr.set_write_buffer_limits)
self.assertRaises(NotImplementedError, tr.get_write_buffer_size)
+ self.assertRaises(NotImplementedError, tr.pause_reading)
+ self.assertRaises(NotImplementedError, tr.resume_reading)
self.assertRaises(NotImplementedError, tr.bind, 'endpoint')
self.assertRaises(NotImplementedError, tr.unbind, 'endpoint')
self.assertRaises(NotImplementedError, tr.bindings)
self.assertRaises(NotImplementedError, tr.connect, 'endpoint')
self.assertRaises(NotImplementedError, tr.disconnect, 'endpoint')
self.assertRaises(NotImplementedError, tr.connections)
self.assertRaises(NotImplementedError, tr.subscribe, b'filter')
self.assertRaises(NotImplementedError, tr.unsubscribe, b'filter')
self.assertRaises(NotImplementedError, tr.subscriptions)
class ZmqProtocolTests(unittest.TestCase):
def test_interface(self):
pr = aiozmq.ZmqProtocol()
self.assertIsNone(pr.msg_received((b'data',)))
| Add missing tests for interfaces | ## Code Before:
import unittest
import aiozmq
class ZmqTransportTests(unittest.TestCase):
def test_interface(self):
tr = aiozmq.ZmqTransport()
self.assertRaises(NotImplementedError, tr.write, [b'data'])
self.assertRaises(NotImplementedError, tr.abort)
self.assertRaises(NotImplementedError, tr.getsockopt, 1)
self.assertRaises(NotImplementedError, tr.setsockopt, 1, 2)
self.assertRaises(NotImplementedError, tr.set_write_buffer_limits)
self.assertRaises(NotImplementedError, tr.get_write_buffer_size)
self.assertRaises(NotImplementedError, tr.bind, 'endpoint')
self.assertRaises(NotImplementedError, tr.unbind, 'endpoint')
self.assertRaises(NotImplementedError, tr.bindings)
self.assertRaises(NotImplementedError, tr.connect, 'endpoint')
self.assertRaises(NotImplementedError, tr.disconnect, 'endpoint')
self.assertRaises(NotImplementedError, tr.connections)
self.assertRaises(NotImplementedError, tr.subscribe, b'filter')
self.assertRaises(NotImplementedError, tr.unsubscribe, b'filter')
self.assertRaises(NotImplementedError, tr.subscriptions)
class ZmqProtocolTests(unittest.TestCase):
def test_interface(self):
pr = aiozmq.ZmqProtocol()
self.assertIsNone(pr.msg_received((b'data',)))
## Instruction:
Add missing tests for interfaces
## Code After:
import unittest
import aiozmq
class ZmqTransportTests(unittest.TestCase):
def test_interface(self):
tr = aiozmq.ZmqTransport()
self.assertRaises(NotImplementedError, tr.write, [b'data'])
self.assertRaises(NotImplementedError, tr.abort)
self.assertRaises(NotImplementedError, tr.getsockopt, 1)
self.assertRaises(NotImplementedError, tr.setsockopt, 1, 2)
self.assertRaises(NotImplementedError, tr.set_write_buffer_limits)
self.assertRaises(NotImplementedError, tr.get_write_buffer_size)
self.assertRaises(NotImplementedError, tr.pause_reading)
self.assertRaises(NotImplementedError, tr.resume_reading)
self.assertRaises(NotImplementedError, tr.bind, 'endpoint')
self.assertRaises(NotImplementedError, tr.unbind, 'endpoint')
self.assertRaises(NotImplementedError, tr.bindings)
self.assertRaises(NotImplementedError, tr.connect, 'endpoint')
self.assertRaises(NotImplementedError, tr.disconnect, 'endpoint')
self.assertRaises(NotImplementedError, tr.connections)
self.assertRaises(NotImplementedError, tr.subscribe, b'filter')
self.assertRaises(NotImplementedError, tr.unsubscribe, b'filter')
self.assertRaises(NotImplementedError, tr.subscriptions)
class ZmqProtocolTests(unittest.TestCase):
def test_interface(self):
pr = aiozmq.ZmqProtocol()
self.assertIsNone(pr.msg_received((b'data',)))
|
5b3d26b6c9256f869d3bc08dfa00bf9b8de58f85 | tests/test_cli_parse.py | tests/test_cli_parse.py |
import os
import pytest
import tempfile
from click.testing import CliRunner
import tbmodels
from tbmodels._cli import cli
from parameters import SAMPLES_DIR
@pytest.mark.parametrize('prefix', ['silicon', 'bi'])
def test_cli_parse(models_equal, prefix):
runner = CliRunner()
with tempfile.TemporaryDirectory() as d:
out_file = os.path.join(d, 'model_out.hdf5')
runner.invoke(cli, ['parse', '-o', out_file, '-f', SAMPLES_DIR, '-p', prefix])
model_res = tbmodels.Model.from_hdf5_file(out_file)
model_reference = tbmodels.Model.from_wannier_folder(folder=SAMPLES_DIR, prefix=prefix)
models_equal(model_res, model_reference)
|
import os
import pytest
import tempfile
from click.testing import CliRunner
import tbmodels
from tbmodels._cli import cli
from parameters import SAMPLES_DIR
@pytest.mark.parametrize('prefix', ['silicon', 'bi'])
def test_cli_parse(models_equal, prefix):
runner = CliRunner()
with tempfile.NamedTemporaryFile() as out_file:
runner.invoke(cli, ['parse', '-o', out_file.name, '-f', SAMPLES_DIR, '-p', prefix])
model_res = tbmodels.Model.from_hdf5_file(out_file.name)
model_reference = tbmodels.Model.from_wannier_folder(folder=SAMPLES_DIR, prefix=prefix)
models_equal(model_res, model_reference)
| Change from TemporaryDirectory to NamedTemporaryFile | Change from TemporaryDirectory to NamedTemporaryFile
| Python | apache-2.0 | Z2PackDev/TBmodels,Z2PackDev/TBmodels |
import os
import pytest
import tempfile
from click.testing import CliRunner
import tbmodels
from tbmodels._cli import cli
from parameters import SAMPLES_DIR
@pytest.mark.parametrize('prefix', ['silicon', 'bi'])
def test_cli_parse(models_equal, prefix):
runner = CliRunner()
- with tempfile.TemporaryDirectory() as d:
+ with tempfile.NamedTemporaryFile() as out_file:
- out_file = os.path.join(d, 'model_out.hdf5')
- runner.invoke(cli, ['parse', '-o', out_file, '-f', SAMPLES_DIR, '-p', prefix])
+ runner.invoke(cli, ['parse', '-o', out_file.name, '-f', SAMPLES_DIR, '-p', prefix])
- model_res = tbmodels.Model.from_hdf5_file(out_file)
+ model_res = tbmodels.Model.from_hdf5_file(out_file.name)
model_reference = tbmodels.Model.from_wannier_folder(folder=SAMPLES_DIR, prefix=prefix)
models_equal(model_res, model_reference)
| Change from TemporaryDirectory to NamedTemporaryFile | ## Code Before:
import os
import pytest
import tempfile
from click.testing import CliRunner
import tbmodels
from tbmodels._cli import cli
from parameters import SAMPLES_DIR
@pytest.mark.parametrize('prefix', ['silicon', 'bi'])
def test_cli_parse(models_equal, prefix):
runner = CliRunner()
with tempfile.TemporaryDirectory() as d:
out_file = os.path.join(d, 'model_out.hdf5')
runner.invoke(cli, ['parse', '-o', out_file, '-f', SAMPLES_DIR, '-p', prefix])
model_res = tbmodels.Model.from_hdf5_file(out_file)
model_reference = tbmodels.Model.from_wannier_folder(folder=SAMPLES_DIR, prefix=prefix)
models_equal(model_res, model_reference)
## Instruction:
Change from TemporaryDirectory to NamedTemporaryFile
## Code After:
import os
import pytest
import tempfile
from click.testing import CliRunner
import tbmodels
from tbmodels._cli import cli
from parameters import SAMPLES_DIR
@pytest.mark.parametrize('prefix', ['silicon', 'bi'])
def test_cli_parse(models_equal, prefix):
runner = CliRunner()
with tempfile.NamedTemporaryFile() as out_file:
runner.invoke(cli, ['parse', '-o', out_file.name, '-f', SAMPLES_DIR, '-p', prefix])
model_res = tbmodels.Model.from_hdf5_file(out_file.name)
model_reference = tbmodels.Model.from_wannier_folder(folder=SAMPLES_DIR, prefix=prefix)
models_equal(model_res, model_reference)
|
62cee7d5a625bb3515eddaddbe940239a41ba31c | rest_framework_msgpack/parsers.py | rest_framework_msgpack/parsers.py | import decimal
import msgpack
from dateutil.parser import parse
from rest_framework.parsers import BaseParser
from rest_framework.exceptions import ParseError
class MessagePackDecoder(object):
def decode(self, obj):
if '__class__' in obj:
decode_func = getattr(self, 'decode_%s' % obj['__class__'])
return decode_func(obj)
return obj
def decode_datetime(self, obj):
return parse(obj['as_str'])
def decode_date(self, obj):
return parse(obj['as_str']).date()
def decode_time(self, obj):
return parse(obj['as_str']).time()
def decode_decimal(self, obj):
return decimal.Decimal(obj['as_str'])
class MessagePackParser(BaseParser):
"""
Parses MessagePack-serialized data.
"""
media_type = 'application/msgpack'
def parse(self, stream, media_type=None, parser_context=None):
try:
return msgpack.load(stream,
use_list=True,
encoding="utf-8",
object_hook=MessagePackDecoder().decode)
except Exception as exc:
raise ParseError('MessagePack parse error - %s' % unicode(exc))
| import decimal
import msgpack
from dateutil.parser import parse
from django.utils.six import text_type
from rest_framework.parsers import BaseParser
from rest_framework.exceptions import ParseError
class MessagePackDecoder(object):
def decode(self, obj):
if '__class__' in obj:
decode_func = getattr(self, 'decode_%s' % obj['__class__'])
return decode_func(obj)
return obj
def decode_datetime(self, obj):
return parse(obj['as_str'])
def decode_date(self, obj):
return parse(obj['as_str']).date()
def decode_time(self, obj):
return parse(obj['as_str']).time()
def decode_decimal(self, obj):
return decimal.Decimal(obj['as_str'])
class MessagePackParser(BaseParser):
"""
Parses MessagePack-serialized data.
"""
media_type = 'application/msgpack'
def parse(self, stream, media_type=None, parser_context=None):
try:
return msgpack.load(stream,
use_list=True,
encoding="utf-8",
object_hook=MessagePackDecoder().decode)
except Exception as exc:
raise ParseError('MessagePack parse error - %s' % text_type(exc))
| Use six.text_type for python3 compat | Use six.text_type for python3 compat | Python | bsd-3-clause | juanriaza/django-rest-framework-msgpack | import decimal
import msgpack
from dateutil.parser import parse
+ from django.utils.six import text_type
+
from rest_framework.parsers import BaseParser
from rest_framework.exceptions import ParseError
class MessagePackDecoder(object):
def decode(self, obj):
if '__class__' in obj:
decode_func = getattr(self, 'decode_%s' % obj['__class__'])
return decode_func(obj)
return obj
def decode_datetime(self, obj):
return parse(obj['as_str'])
def decode_date(self, obj):
return parse(obj['as_str']).date()
def decode_time(self, obj):
return parse(obj['as_str']).time()
def decode_decimal(self, obj):
return decimal.Decimal(obj['as_str'])
class MessagePackParser(BaseParser):
"""
Parses MessagePack-serialized data.
"""
media_type = 'application/msgpack'
def parse(self, stream, media_type=None, parser_context=None):
try:
return msgpack.load(stream,
use_list=True,
encoding="utf-8",
object_hook=MessagePackDecoder().decode)
except Exception as exc:
- raise ParseError('MessagePack parse error - %s' % unicode(exc))
+ raise ParseError('MessagePack parse error - %s' % text_type(exc))
| Use six.text_type for python3 compat | ## Code Before:
import decimal
import msgpack
from dateutil.parser import parse
from rest_framework.parsers import BaseParser
from rest_framework.exceptions import ParseError
class MessagePackDecoder(object):
def decode(self, obj):
if '__class__' in obj:
decode_func = getattr(self, 'decode_%s' % obj['__class__'])
return decode_func(obj)
return obj
def decode_datetime(self, obj):
return parse(obj['as_str'])
def decode_date(self, obj):
return parse(obj['as_str']).date()
def decode_time(self, obj):
return parse(obj['as_str']).time()
def decode_decimal(self, obj):
return decimal.Decimal(obj['as_str'])
class MessagePackParser(BaseParser):
"""
Parses MessagePack-serialized data.
"""
media_type = 'application/msgpack'
def parse(self, stream, media_type=None, parser_context=None):
try:
return msgpack.load(stream,
use_list=True,
encoding="utf-8",
object_hook=MessagePackDecoder().decode)
except Exception as exc:
raise ParseError('MessagePack parse error - %s' % unicode(exc))
## Instruction:
Use six.text_type for python3 compat
## Code After:
import decimal
import msgpack
from dateutil.parser import parse
from django.utils.six import text_type
from rest_framework.parsers import BaseParser
from rest_framework.exceptions import ParseError
class MessagePackDecoder(object):
def decode(self, obj):
if '__class__' in obj:
decode_func = getattr(self, 'decode_%s' % obj['__class__'])
return decode_func(obj)
return obj
def decode_datetime(self, obj):
return parse(obj['as_str'])
def decode_date(self, obj):
return parse(obj['as_str']).date()
def decode_time(self, obj):
return parse(obj['as_str']).time()
def decode_decimal(self, obj):
return decimal.Decimal(obj['as_str'])
class MessagePackParser(BaseParser):
"""
Parses MessagePack-serialized data.
"""
media_type = 'application/msgpack'
def parse(self, stream, media_type=None, parser_context=None):
try:
return msgpack.load(stream,
use_list=True,
encoding="utf-8",
object_hook=MessagePackDecoder().decode)
except Exception as exc:
raise ParseError('MessagePack parse error - %s' % text_type(exc))
|
19bae697bc6e017a97eef77d1425d1ccfbe27ff6 | vprof/__main__.py | vprof/__main__.py | """Visual profiler for Python."""
import argparse
import functools
import json
import profile
import stats_server
import subprocess
import sys
_MODULE_DESC = 'Python visual profiler.'
_HOST = 'localhost'
_PORT = 8000
def main():
parser = argparse.ArgumentParser(description=_MODULE_DESC)
parser.add_argument('source', metavar='src', nargs=1,
help='Python program to profile.')
args = parser.parse_args()
sys.argv[:] = args.source
print('Collecting profile stats...')
program_info = profile.CProfile(args.source[0]).run()
partial_handler = functools.partial(
stats_server.StatsHandler, profile_json=json.dumps(program_info))
subprocess.call(['open', 'http://%s:%s' % (_HOST, _PORT)])
stats_server.start(_HOST, _PORT, partial_handler)
if __name__ == "__main__":
main()
| """Visual profiler for Python."""
import argparse
import functools
import json
import profile
import stats_server
import subprocess
import sys
_MODULE_DESC = 'Python visual profiler.'
_HOST = 'localhost'
_PORT = 8000
_PROFILE_MAP = {
'c': profile.CProfile
}
def main():
parser = argparse.ArgumentParser(description=_MODULE_DESC)
parser.add_argument('profilers', metavar='opts',
help='Profilers configuration')
parser.add_argument('source', metavar='src', nargs=1,
help='Python program to profile.')
args = parser.parse_args()
sys.argv[:] = args.source
program_name = args.source[0]
if len(args.profilers) > len(set(args.profilers)):
print('Profiler configuration is ambiguous. Remove duplicates.')
sys.exit(1)
for prof_option in args.profilers:
if prof_option not in _PROFILE_MAP:
print('Unrecognized option: %s' % prof_option)
sys.exit(2)
print('Collecting profile stats...')
prof_option = args.profilers[0]
profiler = _PROFILE_MAP[prof_option]
program_info = profile.CProfile(args.source[0]).run()
partial_handler = functools.partial(
stats_server.StatsHandler, profile_json=json.dumps(program_info))
subprocess.call(['open', 'http://%s:%s' % (_HOST, _PORT)])
stats_server.start(_HOST, _PORT, partial_handler)
if __name__ == "__main__":
main()
| Add profilers selection as CLI option. | Add profilers selection as CLI option.
| Python | bsd-2-clause | nvdv/vprof,nvdv/vprof,nvdv/vprof | """Visual profiler for Python."""
import argparse
import functools
import json
import profile
import stats_server
import subprocess
import sys
_MODULE_DESC = 'Python visual profiler.'
_HOST = 'localhost'
_PORT = 8000
+ _PROFILE_MAP = {
+ 'c': profile.CProfile
+ }
def main():
parser = argparse.ArgumentParser(description=_MODULE_DESC)
+ parser.add_argument('profilers', metavar='opts',
+ help='Profilers configuration')
parser.add_argument('source', metavar='src', nargs=1,
help='Python program to profile.')
args = parser.parse_args()
+
sys.argv[:] = args.source
+ program_name = args.source[0]
+
+ if len(args.profilers) > len(set(args.profilers)):
+ print('Profiler configuration is ambiguous. Remove duplicates.')
+ sys.exit(1)
+
+ for prof_option in args.profilers:
+ if prof_option not in _PROFILE_MAP:
+ print('Unrecognized option: %s' % prof_option)
+ sys.exit(2)
print('Collecting profile stats...')
+
+ prof_option = args.profilers[0]
+ profiler = _PROFILE_MAP[prof_option]
program_info = profile.CProfile(args.source[0]).run()
partial_handler = functools.partial(
stats_server.StatsHandler, profile_json=json.dumps(program_info))
subprocess.call(['open', 'http://%s:%s' % (_HOST, _PORT)])
stats_server.start(_HOST, _PORT, partial_handler)
if __name__ == "__main__":
main()
| Add profilers selection as CLI option. | ## Code Before:
"""Visual profiler for Python."""
import argparse
import functools
import json
import profile
import stats_server
import subprocess
import sys
_MODULE_DESC = 'Python visual profiler.'
_HOST = 'localhost'
_PORT = 8000
def main():
parser = argparse.ArgumentParser(description=_MODULE_DESC)
parser.add_argument('source', metavar='src', nargs=1,
help='Python program to profile.')
args = parser.parse_args()
sys.argv[:] = args.source
print('Collecting profile stats...')
program_info = profile.CProfile(args.source[0]).run()
partial_handler = functools.partial(
stats_server.StatsHandler, profile_json=json.dumps(program_info))
subprocess.call(['open', 'http://%s:%s' % (_HOST, _PORT)])
stats_server.start(_HOST, _PORT, partial_handler)
if __name__ == "__main__":
main()
## Instruction:
Add profilers selection as CLI option.
## Code After:
"""Visual profiler for Python."""
import argparse
import functools
import json
import profile
import stats_server
import subprocess
import sys
_MODULE_DESC = 'Python visual profiler.'
_HOST = 'localhost'
_PORT = 8000
_PROFILE_MAP = {
'c': profile.CProfile
}
def main():
parser = argparse.ArgumentParser(description=_MODULE_DESC)
parser.add_argument('profilers', metavar='opts',
help='Profilers configuration')
parser.add_argument('source', metavar='src', nargs=1,
help='Python program to profile.')
args = parser.parse_args()
sys.argv[:] = args.source
program_name = args.source[0]
if len(args.profilers) > len(set(args.profilers)):
print('Profiler configuration is ambiguous. Remove duplicates.')
sys.exit(1)
for prof_option in args.profilers:
if prof_option not in _PROFILE_MAP:
print('Unrecognized option: %s' % prof_option)
sys.exit(2)
print('Collecting profile stats...')
prof_option = args.profilers[0]
profiler = _PROFILE_MAP[prof_option]
program_info = profile.CProfile(args.source[0]).run()
partial_handler = functools.partial(
stats_server.StatsHandler, profile_json=json.dumps(program_info))
subprocess.call(['open', 'http://%s:%s' % (_HOST, _PORT)])
stats_server.start(_HOST, _PORT, partial_handler)
if __name__ == "__main__":
main()
|
5e7cce09a6e6a847dad1714973fddb53d60c4c3f | yawf_sample/simple/models.py | yawf_sample/simple/models.py | from django.db import models
import reversion
from yawf.revision import RevisionModelMixin
class WINDOW_OPEN_STATUS:
MINIMIZED = 'minimized'
MAXIMIZED = 'maximized'
NORMAL = 'normal'
types = (MINIMIZED, MAXIMIZED, NORMAL)
choices = zip(types, types)
@reversion.register
class Window(RevisionModelMixin, models.Model):
title = models.CharField(max_length=255)
width = models.IntegerField()
height = models.IntegerField()
workflow_type = 'simple'
open_status = models.CharField(
max_length=32,
choices=WINDOW_OPEN_STATUS.choices,
default='init',
editable=False)
| from django.db import models
import reversion
from yawf.revision import RevisionModelMixin
class WINDOW_OPEN_STATUS:
MINIMIZED = 'minimized'
MAXIMIZED = 'maximized'
NORMAL = 'normal'
types = (MINIMIZED, MAXIMIZED, NORMAL)
choices = zip(types, types)
class Window(RevisionModelMixin, models.Model):
title = models.CharField(max_length=255)
width = models.IntegerField()
height = models.IntegerField()
workflow_type = 'simple'
open_status = models.CharField(
max_length=32,
choices=WINDOW_OPEN_STATUS.choices,
default='init',
editable=False)
reversion.register(Window)
| Fix reversion register in sample app | Fix reversion register in sample app
| Python | mit | freevoid/yawf | from django.db import models
import reversion
from yawf.revision import RevisionModelMixin
class WINDOW_OPEN_STATUS:
MINIMIZED = 'minimized'
MAXIMIZED = 'maximized'
NORMAL = 'normal'
types = (MINIMIZED, MAXIMIZED, NORMAL)
choices = zip(types, types)
- @reversion.register
class Window(RevisionModelMixin, models.Model):
title = models.CharField(max_length=255)
width = models.IntegerField()
height = models.IntegerField()
workflow_type = 'simple'
open_status = models.CharField(
max_length=32,
choices=WINDOW_OPEN_STATUS.choices,
default='init',
editable=False)
+ reversion.register(Window)
+ | Fix reversion register in sample app | ## Code Before:
from django.db import models
import reversion
from yawf.revision import RevisionModelMixin
class WINDOW_OPEN_STATUS:
MINIMIZED = 'minimized'
MAXIMIZED = 'maximized'
NORMAL = 'normal'
types = (MINIMIZED, MAXIMIZED, NORMAL)
choices = zip(types, types)
@reversion.register
class Window(RevisionModelMixin, models.Model):
title = models.CharField(max_length=255)
width = models.IntegerField()
height = models.IntegerField()
workflow_type = 'simple'
open_status = models.CharField(
max_length=32,
choices=WINDOW_OPEN_STATUS.choices,
default='init',
editable=False)
## Instruction:
Fix reversion register in sample app
## Code After:
from django.db import models
import reversion
from yawf.revision import RevisionModelMixin
class WINDOW_OPEN_STATUS:
MINIMIZED = 'minimized'
MAXIMIZED = 'maximized'
NORMAL = 'normal'
types = (MINIMIZED, MAXIMIZED, NORMAL)
choices = zip(types, types)
class Window(RevisionModelMixin, models.Model):
title = models.CharField(max_length=255)
width = models.IntegerField()
height = models.IntegerField()
workflow_type = 'simple'
open_status = models.CharField(
max_length=32,
choices=WINDOW_OPEN_STATUS.choices,
default='init',
editable=False)
reversion.register(Window)
|
d15f6df74b8fe188a7a80c3491aeec62c35ce415 | policy.py | policy.py | from __future__ import unicode_literals
class PolicyError(RuntimeError):
def __init__(self, message, url):
self.message = message
self.url = url
def __str__(self):
return "{}: {}".format(self.message, self.url)
class RedirectLimitPolicy(object):
def __init__(self, max_redirects):
self.max_redirects = max_redirects
self.redirects = 0
def __call__(self, url):
if self.redirects >= self.max_redirects:
raise PolicyError('too many redirects', url)
self.redirects += 1
class VersionCheckPolicy(object):
def __init__(self, os_family, version_string):
self.version_string = version_string
self.os_family = os_family
self.url_version_string = self._version_transform(version_string)
def _version_transform(self, version):
if self.os_family == 'centos':
return version.replace('.', '_')
return version
def __call__(self, url):
if self.os_family == 'centos':
ver = self.url_version_string
if ver not in url:
message = 'version "{}" not found in url'.format(ver)
raise PolicyError(message, url)
| from __future__ import unicode_literals
class PolicyError(RuntimeError):
def __init__(self, message, url):
self.message = message
self.url = url
def __str__(self):
return "{}: {}".format(self.message, self.url)
class RedirectLimitPolicy(object):
def __init__(self, max_redirects):
self.max_redirects = max_redirects
self.redirects = 0
def __call__(self, url):
if self.redirects >= self.max_redirects:
raise PolicyError('too many redirects', url)
self.redirects += 1
class VersionCheckPolicy(object):
def __init__(self, os_family, version_string):
self.version_string = version_string
self.os_family = os_family
self.url_version_string = self._version_transform(version_string)
def _version_transform(self, version):
if self.os_family == 'centos':
return version[:version.rindex('.')]
return version
def __call__(self, url):
if self.os_family == 'centos':
ver = self.url_version_string
if ver not in url:
message = 'version "{}" not found in url'.format(ver)
raise PolicyError(message, url)
| Check for the first part of the version string | Check for the first part of the version string
Older CentOS Linux images only used the xxxx part of the full xxxx.yy
version string from Atlas.
| Python | mit | lpancescu/atlas-lint | from __future__ import unicode_literals
class PolicyError(RuntimeError):
def __init__(self, message, url):
self.message = message
self.url = url
def __str__(self):
return "{}: {}".format(self.message, self.url)
class RedirectLimitPolicy(object):
def __init__(self, max_redirects):
self.max_redirects = max_redirects
self.redirects = 0
def __call__(self, url):
if self.redirects >= self.max_redirects:
raise PolicyError('too many redirects', url)
self.redirects += 1
class VersionCheckPolicy(object):
def __init__(self, os_family, version_string):
self.version_string = version_string
self.os_family = os_family
self.url_version_string = self._version_transform(version_string)
def _version_transform(self, version):
if self.os_family == 'centos':
- return version.replace('.', '_')
+ return version[:version.rindex('.')]
return version
def __call__(self, url):
if self.os_family == 'centos':
ver = self.url_version_string
if ver not in url:
message = 'version "{}" not found in url'.format(ver)
raise PolicyError(message, url)
| Check for the first part of the version string | ## Code Before:
from __future__ import unicode_literals
class PolicyError(RuntimeError):
def __init__(self, message, url):
self.message = message
self.url = url
def __str__(self):
return "{}: {}".format(self.message, self.url)
class RedirectLimitPolicy(object):
def __init__(self, max_redirects):
self.max_redirects = max_redirects
self.redirects = 0
def __call__(self, url):
if self.redirects >= self.max_redirects:
raise PolicyError('too many redirects', url)
self.redirects += 1
class VersionCheckPolicy(object):
def __init__(self, os_family, version_string):
self.version_string = version_string
self.os_family = os_family
self.url_version_string = self._version_transform(version_string)
def _version_transform(self, version):
if self.os_family == 'centos':
return version.replace('.', '_')
return version
def __call__(self, url):
if self.os_family == 'centos':
ver = self.url_version_string
if ver not in url:
message = 'version "{}" not found in url'.format(ver)
raise PolicyError(message, url)
## Instruction:
Check for the first part of the version string
## Code After:
from __future__ import unicode_literals
class PolicyError(RuntimeError):
def __init__(self, message, url):
self.message = message
self.url = url
def __str__(self):
return "{}: {}".format(self.message, self.url)
class RedirectLimitPolicy(object):
def __init__(self, max_redirects):
self.max_redirects = max_redirects
self.redirects = 0
def __call__(self, url):
if self.redirects >= self.max_redirects:
raise PolicyError('too many redirects', url)
self.redirects += 1
class VersionCheckPolicy(object):
def __init__(self, os_family, version_string):
self.version_string = version_string
self.os_family = os_family
self.url_version_string = self._version_transform(version_string)
def _version_transform(self, version):
if self.os_family == 'centos':
return version[:version.rindex('.')]
return version
def __call__(self, url):
if self.os_family == 'centos':
ver = self.url_version_string
if ver not in url:
message = 'version "{}" not found in url'.format(ver)
raise PolicyError(message, url)
|
2965891b46e89e0d7222ec16a2327f2bdef86f52 | chemex/util.py | chemex/util.py | """The util module contains a variety of utility functions."""
import configparser
import sys
def read_cfg_file(filename):
"""Read and parse the experiment configuration file with configparser."""
config = configparser.ConfigParser(inline_comment_prefixes=("#", ";"))
config.optionxform = str
try:
out = config.read(str(filename))
if not out and filename is not None:
exit(f"\nERROR: The file '{filename}' is empty or does not exist!\n")
except configparser.MissingSectionHeaderError:
exit(f"\nERROR: You are missing a section heading in {filename:s}\n")
except configparser.ParsingError:
exit(
"\nERROR: Having trouble reading your parameter file, did you"
" forget '=' signs?\n{:s}".format(sys.exc_info()[1])
)
return config
def normalize_path(working_dir, filename):
"""Normalize the path of a filename relative to a specific directory."""
path = filename
if not path.is_absolute():
path = working_dir / path
return path.resolve()
def header1(string):
"""Print a formatted heading."""
print(("\n".join(["", "", string, "=" * len(string), ""])))
def header2(string):
"""Print a formatted subheading."""
print(("\n".join(["", string, "-" * len(string), ""])))
| """The util module contains a variety of utility functions."""
import configparser
import sys
def listfloat(text):
return [float(val) for val in text.strip("[]").split(",")]
def read_cfg_file(filename=None):
"""Read and parse the experiment configuration file with configparser."""
config = configparser.ConfigParser(
comment_prefixes="#", inline_comment_prefixes="#", converters=listfloat
)
config.optionxform = str
try:
result = config.read(str(filename))
if not result and filename is not None:
exit(f"\nERROR: The file '{filename}' is empty or does not exist!\n")
except configparser.MissingSectionHeaderError:
exit(f"\nERROR: You are missing a section heading in {filename:s}\n")
except configparser.ParsingError:
exit(
"\nERROR: Having trouble reading your parameter file, did you"
" forget '=' signs?\n{:s}".format(sys.exc_info()[1])
)
return config
def normalize_path(working_dir, filename):
"""Normalize the path of a filename relative to a specific directory."""
path = filename
if not path.is_absolute():
path = working_dir / path
return path.resolve()
def header1(string):
"""Print a formatted heading."""
print(("\n".join(["", "", string, "=" * len(string), ""])))
def header2(string):
"""Print a formatted subheading."""
print(("\n".join(["", string, "-" * len(string), ""])))
| Update settings for reading config files | Update settings for reading config files
Update the definition of comments, now only allowing the use of "#"
for comments. Add a converter function to parse list of floats,
such as:
list_of_floats = [1.0, 2.0, 3.0]
| Python | bsd-3-clause | gbouvignies/chemex | """The util module contains a variety of utility functions."""
import configparser
import sys
+ def listfloat(text):
+ return [float(val) for val in text.strip("[]").split(",")]
+
+
- def read_cfg_file(filename):
+ def read_cfg_file(filename=None):
"""Read and parse the experiment configuration file with configparser."""
- config = configparser.ConfigParser(inline_comment_prefixes=("#", ";"))
+ config = configparser.ConfigParser(
+ comment_prefixes="#", inline_comment_prefixes="#", converters=listfloat
+ )
config.optionxform = str
try:
- out = config.read(str(filename))
+ result = config.read(str(filename))
- if not out and filename is not None:
+ if not result and filename is not None:
exit(f"\nERROR: The file '{filename}' is empty or does not exist!\n")
except configparser.MissingSectionHeaderError:
exit(f"\nERROR: You are missing a section heading in {filename:s}\n")
except configparser.ParsingError:
exit(
"\nERROR: Having trouble reading your parameter file, did you"
" forget '=' signs?\n{:s}".format(sys.exc_info()[1])
)
return config
def normalize_path(working_dir, filename):
"""Normalize the path of a filename relative to a specific directory."""
path = filename
if not path.is_absolute():
path = working_dir / path
return path.resolve()
def header1(string):
"""Print a formatted heading."""
print(("\n".join(["", "", string, "=" * len(string), ""])))
def header2(string):
"""Print a formatted subheading."""
print(("\n".join(["", string, "-" * len(string), ""])))
| Update settings for reading config files | ## Code Before:
"""The util module contains a variety of utility functions."""
import configparser
import sys
def read_cfg_file(filename):
"""Read and parse the experiment configuration file with configparser."""
config = configparser.ConfigParser(inline_comment_prefixes=("#", ";"))
config.optionxform = str
try:
out = config.read(str(filename))
if not out and filename is not None:
exit(f"\nERROR: The file '{filename}' is empty or does not exist!\n")
except configparser.MissingSectionHeaderError:
exit(f"\nERROR: You are missing a section heading in {filename:s}\n")
except configparser.ParsingError:
exit(
"\nERROR: Having trouble reading your parameter file, did you"
" forget '=' signs?\n{:s}".format(sys.exc_info()[1])
)
return config
def normalize_path(working_dir, filename):
"""Normalize the path of a filename relative to a specific directory."""
path = filename
if not path.is_absolute():
path = working_dir / path
return path.resolve()
def header1(string):
"""Print a formatted heading."""
print(("\n".join(["", "", string, "=" * len(string), ""])))
def header2(string):
"""Print a formatted subheading."""
print(("\n".join(["", string, "-" * len(string), ""])))
## Instruction:
Update settings for reading config files
## Code After:
"""The util module contains a variety of utility functions."""
import configparser
import sys
def listfloat(text):
return [float(val) for val in text.strip("[]").split(",")]
def read_cfg_file(filename=None):
"""Read and parse the experiment configuration file with configparser."""
config = configparser.ConfigParser(
comment_prefixes="#", inline_comment_prefixes="#", converters=listfloat
)
config.optionxform = str
try:
result = config.read(str(filename))
if not result and filename is not None:
exit(f"\nERROR: The file '{filename}' is empty or does not exist!\n")
except configparser.MissingSectionHeaderError:
exit(f"\nERROR: You are missing a section heading in {filename:s}\n")
except configparser.ParsingError:
exit(
"\nERROR: Having trouble reading your parameter file, did you"
" forget '=' signs?\n{:s}".format(sys.exc_info()[1])
)
return config
def normalize_path(working_dir, filename):
"""Normalize the path of a filename relative to a specific directory."""
path = filename
if not path.is_absolute():
path = working_dir / path
return path.resolve()
def header1(string):
"""Print a formatted heading."""
print(("\n".join(["", "", string, "=" * len(string), ""])))
def header2(string):
"""Print a formatted subheading."""
print(("\n".join(["", string, "-" * len(string), ""])))
|
b6e9e37350a4b435df00a54b2ccd9da70a4db788 | nogotofail/mitm/util/ip.py | nogotofail/mitm/util/ip.py | r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
'''
import subprocess
import re
def get_interface_addresses():
"""Get all ip addresses assigned to interfaces.
Returns a tuple of (v4 addresses, v6 addresses)
"""
try:
output = subprocess.check_output("ifconfig")
except subprocess.CalledProcessError:
# Couldn't call ifconfig. Best guess it.
return (["127.0.0.1"], [])
# Parse out the results.
v4 = re.findall("inet addr:([^ ]*)", output)
v6 = re.findall("inet6 addr: ([^ ]*)", output)
return v4, v6
| r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
'''
import subprocess
import re
def get_interface_addresses():
"""Get all ip addresses assigned to interfaces.
Returns a tuple of (v4 addresses, v6 addresses)
"""
try:
output = subprocess.check_output("ifconfig")
except subprocess.CalledProcessError:
# Couldn't call ifconfig. Best guess it.
return (["127.0.0.1"], [])
# Parse out the results.
v4 = re.findall("inet (addr:)?([^ ]*)", output)
v6 = re.findall("inet6 (addr: )?([^ ]*)", output)
v4 = [e[1] for e in v4]
v6 = [e[1] for e in v6]
return v4, v6
| Fix local interface addr parsing | Fix local interface addr parsing
On Fedora 21 the format of ifconfig is a little different.
Fixes #17
| Python | apache-2.0 | google/nogotofail,leasual/nogotofail,mkenne11/nogotofail,joshcooper/nogotofail,digideskio/nogotofail,mkenne11/nogotofail-pii,joshcooper/nogotofail,google/nogotofail,mkenne11/nogotofail,digideskio/nogotofail,leasual/nogotofail,mkenne11/nogotofail-pii | r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
'''
import subprocess
import re
def get_interface_addresses():
"""Get all ip addresses assigned to interfaces.
Returns a tuple of (v4 addresses, v6 addresses)
"""
try:
output = subprocess.check_output("ifconfig")
except subprocess.CalledProcessError:
# Couldn't call ifconfig. Best guess it.
return (["127.0.0.1"], [])
# Parse out the results.
- v4 = re.findall("inet addr:([^ ]*)", output)
+ v4 = re.findall("inet (addr:)?([^ ]*)", output)
- v6 = re.findall("inet6 addr: ([^ ]*)", output)
+ v6 = re.findall("inet6 (addr: )?([^ ]*)", output)
+ v4 = [e[1] for e in v4]
+ v6 = [e[1] for e in v6]
return v4, v6
| Fix local interface addr parsing | ## Code Before:
r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
'''
import subprocess
import re
def get_interface_addresses():
"""Get all ip addresses assigned to interfaces.
Returns a tuple of (v4 addresses, v6 addresses)
"""
try:
output = subprocess.check_output("ifconfig")
except subprocess.CalledProcessError:
# Couldn't call ifconfig. Best guess it.
return (["127.0.0.1"], [])
# Parse out the results.
v4 = re.findall("inet addr:([^ ]*)", output)
v6 = re.findall("inet6 addr: ([^ ]*)", output)
return v4, v6
## Instruction:
Fix local interface addr parsing
## Code After:
r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
'''
import subprocess
import re
def get_interface_addresses():
"""Get all ip addresses assigned to interfaces.
Returns a tuple of (v4 addresses, v6 addresses)
"""
try:
output = subprocess.check_output("ifconfig")
except subprocess.CalledProcessError:
# Couldn't call ifconfig. Best guess it.
return (["127.0.0.1"], [])
# Parse out the results.
v4 = re.findall("inet (addr:)?([^ ]*)", output)
v6 = re.findall("inet6 (addr: )?([^ ]*)", output)
v4 = [e[1] for e in v4]
v6 = [e[1] for e in v6]
return v4, v6
|
91f5db6ddf6e26cec27917109689c200498dc85f | statsmodels/formula/try_formula.py | statsmodels/formula/try_formula.py | import statsmodels.api as sm
import numpy as np
star98 = sm.datasets.star98.load_pandas().data
formula = 'SUCCESS ~ LOWINC + PERASIAN + PERBLACK + PERHISP + PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
dta = star98[["NABOVE", "NBELOW", "LOWINC", "PERASIAN", "PERBLACK", "PERHISP",
"PCTCHRT", "PCTYRRND", "PERMINTE", "AVYRSEXP", "AVSALK",
"PERSPENK", "PTRATIO", "PCTAF"]]
endog = dta["NABOVE"]/(dta["NABOVE"] + dta.pop("NBELOW"))
del dta["NABOVE"]
dta["SUCCESS"] = endog
endog = dta.pop("SUCCESS")
exog = dta
mod = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
| import statsmodels.api as sm
import numpy as np
star98 = sm.datasets.star98.load_pandas().data
formula = 'SUCCESS ~ LOWINC + PERASIAN + PERBLACK + PERHISP + PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
dta = star98[["NABOVE", "NBELOW", "LOWINC", "PERASIAN", "PERBLACK", "PERHISP",
"PCTCHRT", "PCTYRRND", "PERMINTE", "AVYRSEXP", "AVSALK",
"PERSPENK", "PTRATIO", "PCTAF"]]
endog = dta["NABOVE"]/(dta["NABOVE"] + dta.pop("NBELOW"))
del dta["NABOVE"]
dta["SUCCESS"] = endog
endog = dta.pop("SUCCESS")
exog = dta
mod = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
# try passing a formula object, using user-injected code
def double_it(x):
return 2*x
# What is the correct entry point for this? Should users be able to inject
# code into default_env or similar? I don't see a way to do this yet using
# the approach I have been using, it should be an argument to Desc
from charlton.builtins import builtins
builtins['double_it'] = double_it
formula = 'SUCCESS ~ double_it(LOWINC) + PERASIAN + PERBLACK + PERHISP + '
formula += 'PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
mod2 = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
| Add example for injecting user transform | ENH: Add example for injecting user transform
| Python | bsd-3-clause | bert9bert/statsmodels,YihaoLu/statsmodels,cbmoore/statsmodels,hlin117/statsmodels,bavardage/statsmodels,detrout/debian-statsmodels,hainm/statsmodels,statsmodels/statsmodels,jstoxrocky/statsmodels,statsmodels/statsmodels,astocko/statsmodels,DonBeo/statsmodels,bzero/statsmodels,Averroes/statsmodels,gef756/statsmodels,josef-pkt/statsmodels,alekz112/statsmodels,josef-pkt/statsmodels,kiyoto/statsmodels,kiyoto/statsmodels,saketkc/statsmodels,bavardage/statsmodels,cbmoore/statsmodels,saketkc/statsmodels,wdurhamh/statsmodels,yl565/statsmodels,YihaoLu/statsmodels,ChadFulton/statsmodels,wdurhamh/statsmodels,phobson/statsmodels,ChadFulton/statsmodels,wkfwkf/statsmodels,wkfwkf/statsmodels,jstoxrocky/statsmodels,bert9bert/statsmodels,gef756/statsmodels,astocko/statsmodels,wdurhamh/statsmodels,wkfwkf/statsmodels,wzbozon/statsmodels,alekz112/statsmodels,wzbozon/statsmodels,ChadFulton/statsmodels,phobson/statsmodels,wkfwkf/statsmodels,huongttlan/statsmodels,yl565/statsmodels,rgommers/statsmodels,yl565/statsmodels,nvoron23/statsmodels,edhuckle/statsmodels,nguyentu1602/statsmodels,bsipocz/statsmodels,nvoron23/statsmodels,statsmodels/statsmodels,waynenilsen/statsmodels,wwf5067/statsmodels,wwf5067/statsmodels,edhuckle/statsmodels,josef-pkt/statsmodels,wzbozon/statsmodels,jstoxrocky/statsmodels,musically-ut/statsmodels,gef756/statsmodels,phobson/statsmodels,wwf5067/statsmodels,huongttlan/statsmodels,waynenilsen/statsmodels,adammenges/statsmodels,DonBeo/statsmodels,yl565/statsmodels,waynenilsen/statsmodels,gef756/statsmodels,gef756/statsmodels,bashtage/statsmodels,musically-ut/statsmodels,bavardage/statsmodels,wzbozon/statsmodels,wkfwkf/statsmodels,Averroes/statsmodels,bzero/statsmodels,edhuckle/statsmodels,cbmoore/statsmodels,ChadFulton/statsmodels,saketkc/statsmodels,adammenges/statsmodels,ChadFulton/statsmodels,YihaoLu/statsmodels,josef-pkt/statsmodels,saketkc/statsmodels,jseabold/statsmodels,josef-pkt/statsmodels,phobson/statsmodels,kiyoto/statsmodels,wzbozon/statsmodels,alekz112/statsmodels,nvoron23/statsmodels,DonBeo/statsmodels,bert9bert/statsmodels,musically-ut/statsmodels,wdurhamh/statsmodels,yl565/statsmodels,statsmodels/statsmodels,bsipocz/statsmodels,yarikoptic/pystatsmodels,DonBeo/statsmodels,jseabold/statsmodels,nguyentu1602/statsmodels,bashtage/statsmodels,nvoron23/statsmodels,wdurhamh/statsmodels,rgommers/statsmodels,adammenges/statsmodels,alekz112/statsmodels,detrout/debian-statsmodels,YihaoLu/statsmodels,musically-ut/statsmodels,bert9bert/statsmodels,jseabold/statsmodels,detrout/debian-statsmodels,bavardage/statsmodels,kiyoto/statsmodels,rgommers/statsmodels,statsmodels/statsmodels,Averroes/statsmodels,cbmoore/statsmodels,wwf5067/statsmodels,hainm/statsmodels,huongttlan/statsmodels,bert9bert/statsmodels,detrout/debian-statsmodels,hlin117/statsmodels,Averroes/statsmodels,phobson/statsmodels,bzero/statsmodels,ChadFulton/statsmodels,saketkc/statsmodels,edhuckle/statsmodels,hainm/statsmodels,astocko/statsmodels,jseabold/statsmodels,bashtage/statsmodels,waynenilsen/statsmodels,statsmodels/statsmodels,bashtage/statsmodels,astocko/statsmodels,rgommers/statsmodels,hainm/statsmodels,kiyoto/statsmodels,bsipocz/statsmodels,bsipocz/statsmodels,nvoron23/statsmodels,jstoxrocky/statsmodels,huongttlan/statsmodels,yarikoptic/pystatsmodels,bashtage/statsmodels,jseabold/statsmodels,DonBeo/statsmodels,hlin117/statsmodels,nguyentu1602/statsmodels,adammenges/statsmodels,rgommers/statsmodels,bavardage/statsmodels,josef-pkt/statsmodels,bashtage/statsmodels,edhuckle/statsmodels,cbmoore/statsmodels,bzero/statsmodels,yarikoptic/pystatsmodels,YihaoLu/statsmodels,bzero/statsmodels,hlin117/statsmodels,nguyentu1602/statsmodels | import statsmodels.api as sm
import numpy as np
star98 = sm.datasets.star98.load_pandas().data
formula = 'SUCCESS ~ LOWINC + PERASIAN + PERBLACK + PERHISP + PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
dta = star98[["NABOVE", "NBELOW", "LOWINC", "PERASIAN", "PERBLACK", "PERHISP",
"PCTCHRT", "PCTYRRND", "PERMINTE", "AVYRSEXP", "AVSALK",
"PERSPENK", "PTRATIO", "PCTAF"]]
endog = dta["NABOVE"]/(dta["NABOVE"] + dta.pop("NBELOW"))
del dta["NABOVE"]
dta["SUCCESS"] = endog
endog = dta.pop("SUCCESS")
exog = dta
mod = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
+ # try passing a formula object, using user-injected code
+
+ def double_it(x):
+ return 2*x
+
+ # What is the correct entry point for this? Should users be able to inject
+ # code into default_env or similar? I don't see a way to do this yet using
+ # the approach I have been using, it should be an argument to Desc
+ from charlton.builtins import builtins
+ builtins['double_it'] = double_it
+
+ formula = 'SUCCESS ~ double_it(LOWINC) + PERASIAN + PERBLACK + PERHISP + '
+ formula += 'PCTCHRT '
+ formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
+ mod2 = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
+ | Add example for injecting user transform | ## Code Before:
import statsmodels.api as sm
import numpy as np
star98 = sm.datasets.star98.load_pandas().data
formula = 'SUCCESS ~ LOWINC + PERASIAN + PERBLACK + PERHISP + PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
dta = star98[["NABOVE", "NBELOW", "LOWINC", "PERASIAN", "PERBLACK", "PERHISP",
"PCTCHRT", "PCTYRRND", "PERMINTE", "AVYRSEXP", "AVSALK",
"PERSPENK", "PTRATIO", "PCTAF"]]
endog = dta["NABOVE"]/(dta["NABOVE"] + dta.pop("NBELOW"))
del dta["NABOVE"]
dta["SUCCESS"] = endog
endog = dta.pop("SUCCESS")
exog = dta
mod = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
## Instruction:
Add example for injecting user transform
## Code After:
import statsmodels.api as sm
import numpy as np
star98 = sm.datasets.star98.load_pandas().data
formula = 'SUCCESS ~ LOWINC + PERASIAN + PERBLACK + PERHISP + PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
dta = star98[["NABOVE", "NBELOW", "LOWINC", "PERASIAN", "PERBLACK", "PERHISP",
"PCTCHRT", "PCTYRRND", "PERMINTE", "AVYRSEXP", "AVSALK",
"PERSPENK", "PTRATIO", "PCTAF"]]
endog = dta["NABOVE"]/(dta["NABOVE"] + dta.pop("NBELOW"))
del dta["NABOVE"]
dta["SUCCESS"] = endog
endog = dta.pop("SUCCESS")
exog = dta
mod = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
# try passing a formula object, using user-injected code
def double_it(x):
return 2*x
# What is the correct entry point for this? Should users be able to inject
# code into default_env or similar? I don't see a way to do this yet using
# the approach I have been using, it should be an argument to Desc
from charlton.builtins import builtins
builtins['double_it'] = double_it
formula = 'SUCCESS ~ double_it(LOWINC) + PERASIAN + PERBLACK + PERHISP + '
formula += 'PCTCHRT '
formula += '+ PCTYRRND + PERMINTE*AVYRSEXP*AVSALK + PERSPENK*PTRATIO*PCTAF'
mod2 = sm.GLM(endog, exog, formula=formula, family=sm.families.Binomial()).fit()
|
5e7c99844a0687125e34104cf2c7ee87ca69c0de | mopidy_soundcloud/__init__.py | mopidy_soundcloud/__init__.py | import os
from mopidy import config, ext
from mopidy.exceptions import ExtensionError
__version__ = "2.1.0"
class Extension(ext.Extension):
dist_name = "Mopidy-SoundCloud"
ext_name = "soundcloud"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["explore_songs"] = config.Integer(optional=True)
schema["auth_token"] = config.Secret()
schema["explore"] = config.Deprecated()
schema["explore_pages"] = config.Deprecated()
return schema
def validate_config(self, config): # no_coverage
if not config.getboolean("soundcloud", "enabled"):
return
if not config.get("soundcloud", "auth_token"):
raise ExtensionError(
"In order to use SoundCloud extension you must provide an "
"auth token. For more information refer to "
"https://github.com/mopidy/mopidy-soundcloud/"
)
def setup(self, registry):
from .actor import SoundCloudBackend
registry.add("backend", SoundCloudBackend)
| import pathlib
from mopidy import config, ext
from mopidy.exceptions import ExtensionError
__version__ = "2.1.0"
class Extension(ext.Extension):
dist_name = "Mopidy-SoundCloud"
ext_name = "soundcloud"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["explore_songs"] = config.Integer(optional=True)
schema["auth_token"] = config.Secret()
schema["explore"] = config.Deprecated()
schema["explore_pages"] = config.Deprecated()
return schema
def validate_config(self, config): # no_coverage
if not config.getboolean("soundcloud", "enabled"):
return
if not config.get("soundcloud", "auth_token"):
raise ExtensionError(
"In order to use SoundCloud extension you must provide an "
"auth token. For more information refer to "
"https://github.com/mopidy/mopidy-soundcloud/"
)
def setup(self, registry):
from .actor import SoundCloudBackend
registry.add("backend", SoundCloudBackend)
| Use pathlib to read ext.conf | Use pathlib to read ext.conf
| Python | mit | mopidy/mopidy-soundcloud | - import os
+ import pathlib
from mopidy import config, ext
from mopidy.exceptions import ExtensionError
__version__ = "2.1.0"
class Extension(ext.Extension):
dist_name = "Mopidy-SoundCloud"
ext_name = "soundcloud"
version = __version__
def get_default_config(self):
+ return config.read(pathlib.Path(__file__).parent / "ext.conf")
- conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
- return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["explore_songs"] = config.Integer(optional=True)
schema["auth_token"] = config.Secret()
schema["explore"] = config.Deprecated()
schema["explore_pages"] = config.Deprecated()
return schema
def validate_config(self, config): # no_coverage
if not config.getboolean("soundcloud", "enabled"):
return
if not config.get("soundcloud", "auth_token"):
raise ExtensionError(
"In order to use SoundCloud extension you must provide an "
"auth token. For more information refer to "
"https://github.com/mopidy/mopidy-soundcloud/"
)
def setup(self, registry):
from .actor import SoundCloudBackend
registry.add("backend", SoundCloudBackend)
| Use pathlib to read ext.conf | ## Code Before:
import os
from mopidy import config, ext
from mopidy.exceptions import ExtensionError
__version__ = "2.1.0"
class Extension(ext.Extension):
dist_name = "Mopidy-SoundCloud"
ext_name = "soundcloud"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["explore_songs"] = config.Integer(optional=True)
schema["auth_token"] = config.Secret()
schema["explore"] = config.Deprecated()
schema["explore_pages"] = config.Deprecated()
return schema
def validate_config(self, config): # no_coverage
if not config.getboolean("soundcloud", "enabled"):
return
if not config.get("soundcloud", "auth_token"):
raise ExtensionError(
"In order to use SoundCloud extension you must provide an "
"auth token. For more information refer to "
"https://github.com/mopidy/mopidy-soundcloud/"
)
def setup(self, registry):
from .actor import SoundCloudBackend
registry.add("backend", SoundCloudBackend)
## Instruction:
Use pathlib to read ext.conf
## Code After:
import pathlib
from mopidy import config, ext
from mopidy.exceptions import ExtensionError
__version__ = "2.1.0"
class Extension(ext.Extension):
dist_name = "Mopidy-SoundCloud"
ext_name = "soundcloud"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["explore_songs"] = config.Integer(optional=True)
schema["auth_token"] = config.Secret()
schema["explore"] = config.Deprecated()
schema["explore_pages"] = config.Deprecated()
return schema
def validate_config(self, config): # no_coverage
if not config.getboolean("soundcloud", "enabled"):
return
if not config.get("soundcloud", "auth_token"):
raise ExtensionError(
"In order to use SoundCloud extension you must provide an "
"auth token. For more information refer to "
"https://github.com/mopidy/mopidy-soundcloud/"
)
def setup(self, registry):
from .actor import SoundCloudBackend
registry.add("backend", SoundCloudBackend)
|
94ec7d4816d2a243b8a2b0a0f2dc8a55a7347122 | tests/saltunittest.py | tests/saltunittest.py |
# Import python libs
import os
import sys
# support python < 2.7 via unittest2
if sys.version_info[0:2] < (2, 7):
try:
from unittest2 import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
except ImportError:
print("You need to install unittest2 to run the salt tests")
sys.exit(1)
else:
from unittest import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
# Set up paths
TEST_DIR = os.path.dirname(os.path.normpath(os.path.abspath(__file__)))
SALT_LIBS = os.path.dirname(TEST_DIR)
for dir_ in [TEST_DIR, SALT_LIBS]:
if not dir_ in sys.path:
sys.path.insert(0, dir_)
|
# Import python libs
import os
import sys
# support python < 2.7 via unittest2
if sys.version_info[0:2] < (2, 7):
try:
from unittest2 import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
except ImportError:
raise SystemExit("You need to install unittest2 to run the salt tests")
else:
from unittest import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
# Set up paths
TEST_DIR = os.path.dirname(os.path.normpath(os.path.abspath(__file__)))
SALT_LIBS = os.path.dirname(TEST_DIR)
for dir_ in [TEST_DIR, SALT_LIBS]:
if not dir_ in sys.path:
sys.path.insert(0, dir_)
| Make an error go to stderr and remove net 1 LOC | Make an error go to stderr and remove net 1 LOC
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt |
# Import python libs
import os
import sys
# support python < 2.7 via unittest2
if sys.version_info[0:2] < (2, 7):
try:
from unittest2 import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
except ImportError:
- print("You need to install unittest2 to run the salt tests")
+ raise SystemExit("You need to install unittest2 to run the salt tests")
- sys.exit(1)
else:
from unittest import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
# Set up paths
TEST_DIR = os.path.dirname(os.path.normpath(os.path.abspath(__file__)))
SALT_LIBS = os.path.dirname(TEST_DIR)
for dir_ in [TEST_DIR, SALT_LIBS]:
if not dir_ in sys.path:
sys.path.insert(0, dir_)
| Make an error go to stderr and remove net 1 LOC | ## Code Before:
# Import python libs
import os
import sys
# support python < 2.7 via unittest2
if sys.version_info[0:2] < (2, 7):
try:
from unittest2 import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
except ImportError:
print("You need to install unittest2 to run the salt tests")
sys.exit(1)
else:
from unittest import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
# Set up paths
TEST_DIR = os.path.dirname(os.path.normpath(os.path.abspath(__file__)))
SALT_LIBS = os.path.dirname(TEST_DIR)
for dir_ in [TEST_DIR, SALT_LIBS]:
if not dir_ in sys.path:
sys.path.insert(0, dir_)
## Instruction:
Make an error go to stderr and remove net 1 LOC
## Code After:
# Import python libs
import os
import sys
# support python < 2.7 via unittest2
if sys.version_info[0:2] < (2, 7):
try:
from unittest2 import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
except ImportError:
raise SystemExit("You need to install unittest2 to run the salt tests")
else:
from unittest import TestLoader, TextTestRunner,\
TestCase, expectedFailure, \
TestSuite, skipIf
# Set up paths
TEST_DIR = os.path.dirname(os.path.normpath(os.path.abspath(__file__)))
SALT_LIBS = os.path.dirname(TEST_DIR)
for dir_ in [TEST_DIR, SALT_LIBS]:
if not dir_ in sys.path:
sys.path.insert(0, dir_)
|
b6737b91938d527872eff1d645a205cacf94e15d | tests/test_gobject.py | tests/test_gobject.py |
import unittest
import gobject
import testhelper
class TestGObjectAPI(unittest.TestCase):
def testGObjectModule(self):
obj = gobject.GObject()
self.assertEquals(obj.__module__,
'gobject._gobject')
self.assertEquals(obj.__grefcount__, 1)
class TestFloating(unittest.TestCase):
def testFloatingWithSinkFunc(self):
obj = testhelper.FloatingWithSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithoutSinkFunc(self):
obj = testhelper.FloatingWithoutSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithoutSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
|
import unittest
import gobject
import testhelper
class TestGObjectAPI(unittest.TestCase):
def testGObjectModule(self):
obj = gobject.GObject()
self.assertEquals(obj.__module__,
'gobject._gobject')
class TestReferenceCounting(unittest.TestCase):
def testRegularObject(self):
obj = gobject.GObject()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(gobject.GObject)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithSinkFunc(self):
obj = testhelper.FloatingWithSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithoutSinkFunc(self):
obj = testhelper.FloatingWithoutSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithoutSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
| Add a test to check for regular object reference count | Add a test to check for regular object reference count
https://bugzilla.gnome.org/show_bug.cgi?id=639949
| Python | lgpl-2.1 | alexef/pygobject,Distrotech/pygobject,davidmalcolm/pygobject,davibe/pygobject,jdahlin/pygobject,choeger/pygobject-cmake,MathieuDuponchelle/pygobject,choeger/pygobject-cmake,davidmalcolm/pygobject,davibe/pygobject,pexip/pygobject,Distrotech/pygobject,alexef/pygobject,sfeltman/pygobject,jdahlin/pygobject,GNOME/pygobject,GNOME/pygobject,thiblahute/pygobject,sfeltman/pygobject,thiblahute/pygobject,GNOME/pygobject,Distrotech/pygobject,MathieuDuponchelle/pygobject,nzjrs/pygobject,Distrotech/pygobject,nzjrs/pygobject,jdahlin/pygobject,MathieuDuponchelle/pygobject,davibe/pygobject,pexip/pygobject,alexef/pygobject,pexip/pygobject,sfeltman/pygobject,choeger/pygobject-cmake,thiblahute/pygobject,davidmalcolm/pygobject,nzjrs/pygobject,davibe/pygobject |
import unittest
import gobject
import testhelper
class TestGObjectAPI(unittest.TestCase):
def testGObjectModule(self):
obj = gobject.GObject()
self.assertEquals(obj.__module__,
'gobject._gobject')
+
+
+ class TestReferenceCounting(unittest.TestCase):
+ def testRegularObject(self):
+ obj = gobject.GObject()
self.assertEquals(obj.__grefcount__, 1)
+ obj = gobject.new(gobject.GObject)
+ self.assertEquals(obj.__grefcount__, 1)
- class TestFloating(unittest.TestCase):
def testFloatingWithSinkFunc(self):
obj = testhelper.FloatingWithSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithoutSinkFunc(self):
obj = testhelper.FloatingWithoutSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithoutSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
| Add a test to check for regular object reference count | ## Code Before:
import unittest
import gobject
import testhelper
class TestGObjectAPI(unittest.TestCase):
def testGObjectModule(self):
obj = gobject.GObject()
self.assertEquals(obj.__module__,
'gobject._gobject')
self.assertEquals(obj.__grefcount__, 1)
class TestFloating(unittest.TestCase):
def testFloatingWithSinkFunc(self):
obj = testhelper.FloatingWithSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithoutSinkFunc(self):
obj = testhelper.FloatingWithoutSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithoutSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
## Instruction:
Add a test to check for regular object reference count
## Code After:
import unittest
import gobject
import testhelper
class TestGObjectAPI(unittest.TestCase):
def testGObjectModule(self):
obj = gobject.GObject()
self.assertEquals(obj.__module__,
'gobject._gobject')
class TestReferenceCounting(unittest.TestCase):
def testRegularObject(self):
obj = gobject.GObject()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(gobject.GObject)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithSinkFunc(self):
obj = testhelper.FloatingWithSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
def testFloatingWithoutSinkFunc(self):
obj = testhelper.FloatingWithoutSinkFunc()
self.assertEquals(obj.__grefcount__, 1)
obj = gobject.new(testhelper.FloatingWithoutSinkFunc)
self.assertEquals(obj.__grefcount__, 1)
|
5cf0b19d67a667d4e0d48a12f0ee94f3387cfa37 | tests/test_helpers.py | tests/test_helpers.py |
import testtools
from talons import helpers
from tests import base
class TestHelpers(base.TestCase):
def test_bad_import(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('not.exist.function')
def test_no_function_in_module(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('sys.noexisting')
def test_not_callable(self):
with testtools.ExpectedException(TypeError):
helpers.import_function('sys.stdout')
|
import testtools
from talons import helpers
from tests import base
class TestHelpers(base.TestCase):
def test_bad_import(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('not.exist.function')
def test_no_function_in_module(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('sys.noexisting')
def test_not_callable(self):
with testtools.ExpectedException(TypeError):
helpers.import_function('sys.stdout')
def test_return_function(self):
fn = helpers.import_function('os.path.join')
self.assertEqual(callable(fn), True)
| Add test to ensure talons.helpers.import_function returns a callable | Add test to ensure talons.helpers.import_function returns a callable
| Python | apache-2.0 | talons/talons,jaypipes/talons |
import testtools
from talons import helpers
from tests import base
class TestHelpers(base.TestCase):
def test_bad_import(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('not.exist.function')
def test_no_function_in_module(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('sys.noexisting')
def test_not_callable(self):
with testtools.ExpectedException(TypeError):
helpers.import_function('sys.stdout')
+ def test_return_function(self):
+ fn = helpers.import_function('os.path.join')
+ self.assertEqual(callable(fn), True)
+ | Add test to ensure talons.helpers.import_function returns a callable | ## Code Before:
import testtools
from talons import helpers
from tests import base
class TestHelpers(base.TestCase):
def test_bad_import(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('not.exist.function')
def test_no_function_in_module(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('sys.noexisting')
def test_not_callable(self):
with testtools.ExpectedException(TypeError):
helpers.import_function('sys.stdout')
## Instruction:
Add test to ensure talons.helpers.import_function returns a callable
## Code After:
import testtools
from talons import helpers
from tests import base
class TestHelpers(base.TestCase):
def test_bad_import(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('not.exist.function')
def test_no_function_in_module(self):
with testtools.ExpectedException(ImportError):
helpers.import_function('sys.noexisting')
def test_not_callable(self):
with testtools.ExpectedException(TypeError):
helpers.import_function('sys.stdout')
def test_return_function(self):
fn = helpers.import_function('os.path.join')
self.assertEqual(callable(fn), True)
|
73dfb1fc4ff62705f37b1128e0684e88be416d8f | src/webmention/models.py | src/webmention/models.py | from django.db import models
class WebMentionResponse(models.Model):
response_body = models.TextField()
response_to = models.URLField()
source = models.URLField()
reviewed = models.BooleanField(default=False)
current = models.BooleanField(default=True)
date_created = models.DateTimeField(auto_now_add=True)
date_modified = models.DateTimeField(auto_now=True)
class Meta:
verbose_name = "webmention"
verbose_name_plural = "webmentions"
def __str__(self):
return self.source
def source_for_admin(self):
return '<a href="{href}">{href}</a>'.format(href=self.source)
source_for_admin.allow_tags = True
source_for_admin.short_description = "source"
def response_to_for_admin(self):
return '<a href="{href}">{href}</a>'.format(href=self.response_to)
response_to_for_admin.allow_tags = True
response_to_for_admin.short_description = "response to"
def invalidate(self):
if self.id:
self.current = False
self.save()
def update(self, source, target, response_body):
self.response_body = response_body
self.source = source
self.response_to = target
self.current = True
self.save()
| from django.db import models
class WebMentionResponse(models.Model):
response_body = models.TextField()
response_to = models.URLField()
source = models.URLField()
reviewed = models.BooleanField(default=False)
current = models.BooleanField(default=True)
date_created = models.DateTimeField(auto_now_add=True)
date_modified = models.DateTimeField(auto_now=True)
class Meta:
verbose_name = "webmention"
verbose_name_plural = "webmentions"
def __str__(self):
return self.source
def source_for_admin(self):
return format_html('<a href="{href}">{href}</a>'.format(href=self.source))
source_for_admin.short_description = "source"
def response_to_for_admin(self):
return format_html('<a href="{href}">{href}</a>'.format(href=self.response_to))
response_to_for_admin.short_description = "response to"
def invalidate(self):
if self.id:
self.current = False
self.save()
def update(self, source, target, response_body):
self.response_body = response_body
self.source = source
self.response_to = target
self.current = True
self.save()
| Update deprecated allow_tags to format_html | Update deprecated allow_tags to format_html
| Python | mit | easy-as-python/django-webmention | from django.db import models
class WebMentionResponse(models.Model):
response_body = models.TextField()
response_to = models.URLField()
source = models.URLField()
reviewed = models.BooleanField(default=False)
current = models.BooleanField(default=True)
date_created = models.DateTimeField(auto_now_add=True)
date_modified = models.DateTimeField(auto_now=True)
class Meta:
verbose_name = "webmention"
verbose_name_plural = "webmentions"
def __str__(self):
return self.source
def source_for_admin(self):
- return '<a href="{href}">{href}</a>'.format(href=self.source)
+ return format_html('<a href="{href}">{href}</a>'.format(href=self.source))
- source_for_admin.allow_tags = True
source_for_admin.short_description = "source"
def response_to_for_admin(self):
- return '<a href="{href}">{href}</a>'.format(href=self.response_to)
+ return format_html('<a href="{href}">{href}</a>'.format(href=self.response_to))
- response_to_for_admin.allow_tags = True
response_to_for_admin.short_description = "response to"
def invalidate(self):
if self.id:
self.current = False
self.save()
def update(self, source, target, response_body):
self.response_body = response_body
self.source = source
self.response_to = target
self.current = True
self.save()
| Update deprecated allow_tags to format_html | ## Code Before:
from django.db import models
class WebMentionResponse(models.Model):
response_body = models.TextField()
response_to = models.URLField()
source = models.URLField()
reviewed = models.BooleanField(default=False)
current = models.BooleanField(default=True)
date_created = models.DateTimeField(auto_now_add=True)
date_modified = models.DateTimeField(auto_now=True)
class Meta:
verbose_name = "webmention"
verbose_name_plural = "webmentions"
def __str__(self):
return self.source
def source_for_admin(self):
return '<a href="{href}">{href}</a>'.format(href=self.source)
source_for_admin.allow_tags = True
source_for_admin.short_description = "source"
def response_to_for_admin(self):
return '<a href="{href}">{href}</a>'.format(href=self.response_to)
response_to_for_admin.allow_tags = True
response_to_for_admin.short_description = "response to"
def invalidate(self):
if self.id:
self.current = False
self.save()
def update(self, source, target, response_body):
self.response_body = response_body
self.source = source
self.response_to = target
self.current = True
self.save()
## Instruction:
Update deprecated allow_tags to format_html
## Code After:
from django.db import models
class WebMentionResponse(models.Model):
response_body = models.TextField()
response_to = models.URLField()
source = models.URLField()
reviewed = models.BooleanField(default=False)
current = models.BooleanField(default=True)
date_created = models.DateTimeField(auto_now_add=True)
date_modified = models.DateTimeField(auto_now=True)
class Meta:
verbose_name = "webmention"
verbose_name_plural = "webmentions"
def __str__(self):
return self.source
def source_for_admin(self):
return format_html('<a href="{href}">{href}</a>'.format(href=self.source))
source_for_admin.short_description = "source"
def response_to_for_admin(self):
return format_html('<a href="{href}">{href}</a>'.format(href=self.response_to))
response_to_for_admin.short_description = "response to"
def invalidate(self):
if self.id:
self.current = False
self.save()
def update(self, source, target, response_body):
self.response_body = response_body
self.source = source
self.response_to = target
self.current = True
self.save()
|
e612a742caeedf5398522365c984a39c505e7872 | warehouse/exceptions.py | warehouse/exceptions.py | class FailedSynchronization(Exception):
pass
class SynchronizationTimeout(Exception):
"""
A synchronization for a particular project took longer than the timeout.
"""
| from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
class FailedSynchronization(Exception):
pass
class SynchronizationTimeout(Exception):
"""
A synchronization for a particular project took longer than the timeout.
"""
| Add the standard __future__ imports | Add the standard __future__ imports
| Python | bsd-2-clause | davidfischer/warehouse | + from __future__ import absolute_import
+ from __future__ import division
+ from __future__ import unicode_literals
+
+
class FailedSynchronization(Exception):
pass
class SynchronizationTimeout(Exception):
"""
A synchronization for a particular project took longer than the timeout.
"""
| Add the standard __future__ imports | ## Code Before:
class FailedSynchronization(Exception):
pass
class SynchronizationTimeout(Exception):
"""
A synchronization for a particular project took longer than the timeout.
"""
## Instruction:
Add the standard __future__ imports
## Code After:
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
class FailedSynchronization(Exception):
pass
class SynchronizationTimeout(Exception):
"""
A synchronization for a particular project took longer than the timeout.
"""
|
34c427200c6ab50fb64fa0d6116366a8fa9186a3 | netman/core/objects/bond.py | netman/core/objects/bond.py |
from netman.core.objects.interface import BaseInterface
class Bond(BaseInterface):
def __init__(self, number=None, link_speed=None, members=None, **interface):
super(Bond, self).__init__(**interface)
self.number = number
self.link_speed = link_speed
self.members = members or []
| import warnings
from netman.core.objects.interface import BaseInterface
class Bond(BaseInterface):
def __init__(self, number=None, link_speed=None, members=None, **interface):
super(Bond, self).__init__(**interface)
self.number = number
self.link_speed = link_speed
self.members = members or []
@property
def interface(self):
warnings.warn('Deprecated: Use directly the members of Bond instead.',
category=DeprecationWarning)
return self
| Support deprecated use of the interface property of Bond. | Support deprecated use of the interface property of Bond.
| Python | apache-2.0 | idjaw/netman,internaphosting/netman,internap/netman,godp1301/netman,mat128/netman,lindycoder/netman | + import warnings
from netman.core.objects.interface import BaseInterface
class Bond(BaseInterface):
def __init__(self, number=None, link_speed=None, members=None, **interface):
super(Bond, self).__init__(**interface)
self.number = number
self.link_speed = link_speed
self.members = members or []
+ @property
+ def interface(self):
+ warnings.warn('Deprecated: Use directly the members of Bond instead.',
+ category=DeprecationWarning)
+ return self
+ | Support deprecated use of the interface property of Bond. | ## Code Before:
from netman.core.objects.interface import BaseInterface
class Bond(BaseInterface):
def __init__(self, number=None, link_speed=None, members=None, **interface):
super(Bond, self).__init__(**interface)
self.number = number
self.link_speed = link_speed
self.members = members or []
## Instruction:
Support deprecated use of the interface property of Bond.
## Code After:
import warnings
from netman.core.objects.interface import BaseInterface
class Bond(BaseInterface):
def __init__(self, number=None, link_speed=None, members=None, **interface):
super(Bond, self).__init__(**interface)
self.number = number
self.link_speed = link_speed
self.members = members or []
@property
def interface(self):
warnings.warn('Deprecated: Use directly the members of Bond instead.',
category=DeprecationWarning)
return self
|
ab06e871a6c820845da2d4d60bb6b1874350cf16 | circuits/web/__init__.py | circuits/web/__init__.py |
from loggers import Logger
from core import Controller
from sessions import Sessions
from events import Request, Response
from servers import BaseServer, Server
from errors import HTTPError, Forbidden, NotFound, Redirect
from dispatchers import Static, Dispatcher, VirtualHosts, XMLRPC
try:
from dispatchers import JSONRPC
except ImportError:
pass
|
from utils import url
from loggers import Logger
from sessions import Sessions
from core import expose, Controller
from events import Request, Response
from servers import BaseServer, Server
from errors import HTTPError, Forbidden, NotFound, Redirect
from dispatchers import Static, Dispatcher, VirtualHosts, XMLRPC
try:
from dispatchers import JSONRPC
except ImportError:
pass
| Add url and expose to this namesapce | circuits.web: Add url and expose to this namesapce
| Python | mit | eriol/circuits,eriol/circuits,nizox/circuits,treemo/circuits,treemo/circuits,eriol/circuits,treemo/circuits |
+ from utils import url
from loggers import Logger
- from core import Controller
from sessions import Sessions
+ from core import expose, Controller
from events import Request, Response
from servers import BaseServer, Server
from errors import HTTPError, Forbidden, NotFound, Redirect
from dispatchers import Static, Dispatcher, VirtualHosts, XMLRPC
try:
from dispatchers import JSONRPC
except ImportError:
pass
| Add url and expose to this namesapce | ## Code Before:
from loggers import Logger
from core import Controller
from sessions import Sessions
from events import Request, Response
from servers import BaseServer, Server
from errors import HTTPError, Forbidden, NotFound, Redirect
from dispatchers import Static, Dispatcher, VirtualHosts, XMLRPC
try:
from dispatchers import JSONRPC
except ImportError:
pass
## Instruction:
Add url and expose to this namesapce
## Code After:
from utils import url
from loggers import Logger
from sessions import Sessions
from core import expose, Controller
from events import Request, Response
from servers import BaseServer, Server
from errors import HTTPError, Forbidden, NotFound, Redirect
from dispatchers import Static, Dispatcher, VirtualHosts, XMLRPC
try:
from dispatchers import JSONRPC
except ImportError:
pass
|
3ddad0538430499182c583a0a7f877884038c0a5 | Lib/test/test_symtable.py | Lib/test/test_symtable.py | from test.test_support import vereq, TestFailed
import symtable
symbols = symtable.symtable("def f(x): return x", "?", "exec")
## XXX
## Test disabled because symtable module needs to be rewritten for new compiler
##vereq(symbols[0].name, "global")
##vereq(len([ste for ste in symbols.values() if ste.name == "f"]), 1)
### Bug tickler: SyntaxError file name correct whether error raised
### while parsing or building symbol table.
##def checkfilename(brokencode):
## try:
## _symtable.symtable(brokencode, "spam", "exec")
## except SyntaxError, e:
## vereq(e.filename, "spam")
## else:
## raise TestFailed("no SyntaxError for %r" % (brokencode,))
##checkfilename("def f(x): foo)(") # parse-time
##checkfilename("def f(x): global x") # symtable-build-time
| from test import test_support
import symtable
import unittest
## XXX
## Test disabled because symtable module needs to be rewritten for new compiler
##vereq(symbols[0].name, "global")
##vereq(len([ste for ste in symbols.values() if ste.name == "f"]), 1)
### Bug tickler: SyntaxError file name correct whether error raised
### while parsing or building symbol table.
##def checkfilename(brokencode):
## try:
## _symtable.symtable(brokencode, "spam", "exec")
## except SyntaxError, e:
## vereq(e.filename, "spam")
## else:
## raise TestFailed("no SyntaxError for %r" % (brokencode,))
##checkfilename("def f(x): foo)(") # parse-time
##checkfilename("def f(x): global x") # symtable-build-time
class SymtableTest(unittest.TestCase):
def test_invalid_args(self):
self.assertRaises(TypeError, symtable.symtable, "42")
self.assertRaises(ValueError, symtable.symtable, "42", "?", "")
def test_eval(self):
symbols = symtable.symtable("42", "?", "eval")
def test_single(self):
symbols = symtable.symtable("42", "?", "single")
def test_exec(self):
symbols = symtable.symtable("def f(x): return x", "?", "exec")
def test_main():
test_support.run_unittest(SymtableTest)
if __name__ == '__main__':
test_main()
| Use unittest and make sure a few other cases don't crash | Use unittest and make sure a few other cases don't crash
| Python | mit | sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator | - from test.test_support import vereq, TestFailed
+ from test import test_support
import symtable
+ import unittest
- symbols = symtable.symtable("def f(x): return x", "?", "exec")
## XXX
## Test disabled because symtable module needs to be rewritten for new compiler
##vereq(symbols[0].name, "global")
##vereq(len([ste for ste in symbols.values() if ste.name == "f"]), 1)
### Bug tickler: SyntaxError file name correct whether error raised
### while parsing or building symbol table.
##def checkfilename(brokencode):
## try:
## _symtable.symtable(brokencode, "spam", "exec")
## except SyntaxError, e:
## vereq(e.filename, "spam")
## else:
## raise TestFailed("no SyntaxError for %r" % (brokencode,))
##checkfilename("def f(x): foo)(") # parse-time
##checkfilename("def f(x): global x") # symtable-build-time
+ class SymtableTest(unittest.TestCase):
+ def test_invalid_args(self):
+ self.assertRaises(TypeError, symtable.symtable, "42")
+ self.assertRaises(ValueError, symtable.symtable, "42", "?", "")
+
+ def test_eval(self):
+ symbols = symtable.symtable("42", "?", "eval")
+
+ def test_single(self):
+ symbols = symtable.symtable("42", "?", "single")
+
+ def test_exec(self):
+ symbols = symtable.symtable("def f(x): return x", "?", "exec")
+
+
+ def test_main():
+ test_support.run_unittest(SymtableTest)
+
+ if __name__ == '__main__':
+ test_main()
+ | Use unittest and make sure a few other cases don't crash | ## Code Before:
from test.test_support import vereq, TestFailed
import symtable
symbols = symtable.symtable("def f(x): return x", "?", "exec")
## XXX
## Test disabled because symtable module needs to be rewritten for new compiler
##vereq(symbols[0].name, "global")
##vereq(len([ste for ste in symbols.values() if ste.name == "f"]), 1)
### Bug tickler: SyntaxError file name correct whether error raised
### while parsing or building symbol table.
##def checkfilename(brokencode):
## try:
## _symtable.symtable(brokencode, "spam", "exec")
## except SyntaxError, e:
## vereq(e.filename, "spam")
## else:
## raise TestFailed("no SyntaxError for %r" % (brokencode,))
##checkfilename("def f(x): foo)(") # parse-time
##checkfilename("def f(x): global x") # symtable-build-time
## Instruction:
Use unittest and make sure a few other cases don't crash
## Code After:
from test import test_support
import symtable
import unittest
## XXX
## Test disabled because symtable module needs to be rewritten for new compiler
##vereq(symbols[0].name, "global")
##vereq(len([ste for ste in symbols.values() if ste.name == "f"]), 1)
### Bug tickler: SyntaxError file name correct whether error raised
### while parsing or building symbol table.
##def checkfilename(brokencode):
## try:
## _symtable.symtable(brokencode, "spam", "exec")
## except SyntaxError, e:
## vereq(e.filename, "spam")
## else:
## raise TestFailed("no SyntaxError for %r" % (brokencode,))
##checkfilename("def f(x): foo)(") # parse-time
##checkfilename("def f(x): global x") # symtable-build-time
class SymtableTest(unittest.TestCase):
def test_invalid_args(self):
self.assertRaises(TypeError, symtable.symtable, "42")
self.assertRaises(ValueError, symtable.symtable, "42", "?", "")
def test_eval(self):
symbols = symtable.symtable("42", "?", "eval")
def test_single(self):
symbols = symtable.symtable("42", "?", "single")
def test_exec(self):
symbols = symtable.symtable("def f(x): return x", "?", "exec")
def test_main():
test_support.run_unittest(SymtableTest)
if __name__ == '__main__':
test_main()
|
f559001d2c46fade2d9b62f9cb7a3f8053e8b80f | OMDB_api_scrape.py | OMDB_api_scrape.py |
import json, requests, sys, os
URL_BASE = 'http://www.omdbapi.com/?'
if len(sys.argv) > 1:
# Get address from command line.
mTitle = '+'.join(sys.argv[1:-1])
mYear = sys.argv[-1]
print(mTitle)
print(mYear)
else:
print("Usage: OMDB_api_scrape.py <Movie Title> <Year>")
sys.exit(1)
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
try:
response = requests.get(url)
response.raise_for_status()
except requests.exceptions.RequestException as err:
print(err)
sys.exit(1)
theJSON = json.loads(response.text)
# Save the JSON file
with open((os.path.join(os.getcwd(), (mTitle + '_' + mYear + '.json'))), 'w') as outfile:
json.dump(theJSON, outfile)
|
import requests, sys, os
import lxml.etree
URL_BASE = 'http://www.omdbapi.com/?'
if len(sys.argv) > 1:
# Get address from command line.
mTitle = '+'.join(sys.argv[1:-1])
mYear = sys.argv[-1]
print(mTitle)
print(mYear)
else:
print("Usage: OMDB_api_scrape.py <Movie Title> <Year>")
sys.exit(1)
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=xml'
# Try to get the url
try:
response = requests.get(url)
response.raise_for_status()
except requests.exceptions.RequestException as err:
print(err)
sys.exit(1)
# Save the XML file
with open((os.path.join(os.getcwd(), (mTitle + '_' + mYear + '.xml'))), 'wb') as outfile:
outfile.write(response.text)
| Convert OMDB scrapper to grab xml | Convert OMDB scrapper to grab xml
| Python | mit | samcheck/PyMedia,samcheck/PyMedia,samcheck/PyMedia |
- import json, requests, sys, os
+ import requests, sys, os
+ import lxml.etree
URL_BASE = 'http://www.omdbapi.com/?'
if len(sys.argv) > 1:
# Get address from command line.
mTitle = '+'.join(sys.argv[1:-1])
mYear = sys.argv[-1]
print(mTitle)
print(mYear)
else:
print("Usage: OMDB_api_scrape.py <Movie Title> <Year>")
sys.exit(1)
# Craft the URL
- url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
+ url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=xml'
# Try to get the url
try:
response = requests.get(url)
response.raise_for_status()
except requests.exceptions.RequestException as err:
print(err)
sys.exit(1)
- theJSON = json.loads(response.text)
+ # Save the XML file
+ with open((os.path.join(os.getcwd(), (mTitle + '_' + mYear + '.xml'))), 'wb') as outfile:
+ outfile.write(response.text)
- # Save the JSON file
- with open((os.path.join(os.getcwd(), (mTitle + '_' + mYear + '.json'))), 'w') as outfile:
- json.dump(theJSON, outfile)
- | Convert OMDB scrapper to grab xml | ## Code Before:
import json, requests, sys, os
URL_BASE = 'http://www.omdbapi.com/?'
if len(sys.argv) > 1:
# Get address from command line.
mTitle = '+'.join(sys.argv[1:-1])
mYear = sys.argv[-1]
print(mTitle)
print(mYear)
else:
print("Usage: OMDB_api_scrape.py <Movie Title> <Year>")
sys.exit(1)
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
try:
response = requests.get(url)
response.raise_for_status()
except requests.exceptions.RequestException as err:
print(err)
sys.exit(1)
theJSON = json.loads(response.text)
# Save the JSON file
with open((os.path.join(os.getcwd(), (mTitle + '_' + mYear + '.json'))), 'w') as outfile:
json.dump(theJSON, outfile)
## Instruction:
Convert OMDB scrapper to grab xml
## Code After:
import requests, sys, os
import lxml.etree
URL_BASE = 'http://www.omdbapi.com/?'
if len(sys.argv) > 1:
# Get address from command line.
mTitle = '+'.join(sys.argv[1:-1])
mYear = sys.argv[-1]
print(mTitle)
print(mYear)
else:
print("Usage: OMDB_api_scrape.py <Movie Title> <Year>")
sys.exit(1)
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=xml'
# Try to get the url
try:
response = requests.get(url)
response.raise_for_status()
except requests.exceptions.RequestException as err:
print(err)
sys.exit(1)
# Save the XML file
with open((os.path.join(os.getcwd(), (mTitle + '_' + mYear + '.xml'))), 'wb') as outfile:
outfile.write(response.text)
|
8ac20f5bec2d94e71b92e48568d18fd74be4e5d4 | fabfile.py | fabfile.py | from fabric.api import task, local
@task
def up():
"""
Deploy new release to pypi. Using twine util
"""
local("rm -rf dist")
local("rm -rf pgup.egg-info")
local("python ./setup.py sdist")
local("twine upload dist/{}".format(local("ls dist", capture=True).strip()))
@task
def docs():
"""
Deploy documentation to pythonhosted. Using sphinx
"""
local("rm -rf build/html")
local("python ./setup.py build_sphinx")
local("python ./setup.py upload_sphinx")
| from fabric.api import task, local, execute
@task
def up():
"""
Deploy new release to pypi. Using twine util
"""
local("rm -rf dist")
local("rm -rf pgup.egg-info")
local("python ./setup.py sdist")
local("twine upload dist/{}".format(local("ls dist", capture=True).strip()))
execute(syncMezzo)
@task
def docs():
"""
Deploy documentation to pythonhosted. Using sphinx
"""
local("rm -rf build/html")
local("python ./setup.py build_sphinx")
local("python ./setup.py upload_sphinx")
@task
def syncMezzo():
"""
Copy current module version to mezzo project
"""
local("rm -rf /opt/mezzo/pgup")
local("mkdir /opt/mezzo/pgup")
local("cp -R etc pgup /opt/mezzo/pgup")
local("cp LICENSE MANIFEST.in README.md setup.py /opt/mezzo/pgup")
| Add fab task for syncing pgup with mezzo | Add fab task for syncing pgup with mezzo
| Python | mit | stepan-perlov/pgup | - from fabric.api import task, local
+ from fabric.api import task, local, execute
@task
def up():
"""
Deploy new release to pypi. Using twine util
"""
local("rm -rf dist")
local("rm -rf pgup.egg-info")
local("python ./setup.py sdist")
local("twine upload dist/{}".format(local("ls dist", capture=True).strip()))
+ execute(syncMezzo)
@task
def docs():
"""
Deploy documentation to pythonhosted. Using sphinx
"""
local("rm -rf build/html")
local("python ./setup.py build_sphinx")
local("python ./setup.py upload_sphinx")
+ @task
+ def syncMezzo():
+ """
+ Copy current module version to mezzo project
+ """
+ local("rm -rf /opt/mezzo/pgup")
+ local("mkdir /opt/mezzo/pgup")
+ local("cp -R etc pgup /opt/mezzo/pgup")
+ local("cp LICENSE MANIFEST.in README.md setup.py /opt/mezzo/pgup")
+ | Add fab task for syncing pgup with mezzo | ## Code Before:
from fabric.api import task, local
@task
def up():
"""
Deploy new release to pypi. Using twine util
"""
local("rm -rf dist")
local("rm -rf pgup.egg-info")
local("python ./setup.py sdist")
local("twine upload dist/{}".format(local("ls dist", capture=True).strip()))
@task
def docs():
"""
Deploy documentation to pythonhosted. Using sphinx
"""
local("rm -rf build/html")
local("python ./setup.py build_sphinx")
local("python ./setup.py upload_sphinx")
## Instruction:
Add fab task for syncing pgup with mezzo
## Code After:
from fabric.api import task, local, execute
@task
def up():
"""
Deploy new release to pypi. Using twine util
"""
local("rm -rf dist")
local("rm -rf pgup.egg-info")
local("python ./setup.py sdist")
local("twine upload dist/{}".format(local("ls dist", capture=True).strip()))
execute(syncMezzo)
@task
def docs():
"""
Deploy documentation to pythonhosted. Using sphinx
"""
local("rm -rf build/html")
local("python ./setup.py build_sphinx")
local("python ./setup.py upload_sphinx")
@task
def syncMezzo():
"""
Copy current module version to mezzo project
"""
local("rm -rf /opt/mezzo/pgup")
local("mkdir /opt/mezzo/pgup")
local("cp -R etc pgup /opt/mezzo/pgup")
local("cp LICENSE MANIFEST.in README.md setup.py /opt/mezzo/pgup")
|
c3c26c15895a192fba22482306182950c0ccd57c | python/simple-linked-list/simple_linked_list.py | python/simple-linked-list/simple_linked_list.py | class Node(object):
def __init__(self, value):
pass
def value(self):
pass
def next(self):
pass
class LinkedList(object):
def __init__(self, values=[]):
self.size = 0
for value in values:
self.push(value)
def __len__(self):
return self.size
def head(self):
pass
def push(self, value):
node = Node(value)
if self.head is None:
self.head = node
self.size += 1
def pop(self):
pass
def reversed(self):
pass
class EmptyListException(Exception):
pass
| class Node(object):
def __init__(self, value):
pass
def value(self):
pass
def next(self):
pass
class LinkedList(object):
def __init__(self, values=[]):
self._size = 0
self._head = None
for value in values:
self.push(value)
def __len__(self):
return self._size
def head(self):
if self._head is None:
raise EmptyListException("Head is empty")
def push(self, value):
node = Node(value)
if self._head is None:
self._head = node
self._size += 1
def pop(self):
pass
def reversed(self):
pass
class EmptyListException(Exception):
def __init__(self, message):
self.message = message
| Mark instance variables private with leading _ | Mark instance variables private with leading _
| Python | mit | rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism,rootulp/exercism | class Node(object):
def __init__(self, value):
pass
def value(self):
pass
def next(self):
pass
class LinkedList(object):
def __init__(self, values=[]):
- self.size = 0
+ self._size = 0
+ self._head = None
+
for value in values:
self.push(value)
def __len__(self):
- return self.size
+ return self._size
def head(self):
- pass
+ if self._head is None:
+ raise EmptyListException("Head is empty")
def push(self, value):
node = Node(value)
- if self.head is None:
+ if self._head is None:
- self.head = node
+ self._head = node
- self.size += 1
+ self._size += 1
def pop(self):
pass
def reversed(self):
pass
class EmptyListException(Exception):
- pass
+ def __init__(self, message):
+ self.message = message
| Mark instance variables private with leading _ | ## Code Before:
class Node(object):
def __init__(self, value):
pass
def value(self):
pass
def next(self):
pass
class LinkedList(object):
def __init__(self, values=[]):
self.size = 0
for value in values:
self.push(value)
def __len__(self):
return self.size
def head(self):
pass
def push(self, value):
node = Node(value)
if self.head is None:
self.head = node
self.size += 1
def pop(self):
pass
def reversed(self):
pass
class EmptyListException(Exception):
pass
## Instruction:
Mark instance variables private with leading _
## Code After:
class Node(object):
def __init__(self, value):
pass
def value(self):
pass
def next(self):
pass
class LinkedList(object):
def __init__(self, values=[]):
self._size = 0
self._head = None
for value in values:
self.push(value)
def __len__(self):
return self._size
def head(self):
if self._head is None:
raise EmptyListException("Head is empty")
def push(self, value):
node = Node(value)
if self._head is None:
self._head = node
self._size += 1
def pop(self):
pass
def reversed(self):
pass
class EmptyListException(Exception):
def __init__(self, message):
self.message = message
|
fa609681c2732e655cde9075182af918983ccc1f | photutils/utils/_misc.py | photutils/utils/_misc.py | from datetime import datetime, timezone
def _get_version_info():
"""
Return a dictionary of the installed version numbers for photutils
and its dependencies.
Returns
-------
result : dict
A dictionary containing the version numbers for photutils and
its dependencies.
"""
versions = {}
packages = ('photutils', 'astropy', 'numpy', 'scipy', 'skimage')
for package in packages:
try:
pkg = __import__(package)
version = pkg.__version__
except ImportError:
version = None
versions[package] = version
return versions
def _get_date(utc=False):
"""
Return a string of the current date/time.
Parameters
----------
utz : bool, optional
Whether to use the UTZ timezone instead of the local timezone.
Returns
-------
result : str
The current date/time.
"""
if not utc:
now = datetime.now().astimezone()
else:
now = datetime.now(timezone.utc)
return now.strftime('%Y-%m-%d %H:%M:%S %Z')
def _get_meta(utc=False):
"""
Return a metadata dictionary with the package versions and current
date/time.
"""
return {'date': _get_date(utc=utc),
'version': _get_version_info()}
| from datetime import datetime, timezone
import sys
def _get_version_info():
"""
Return a dictionary of the installed version numbers for photutils
and its dependencies.
Returns
-------
result : dict
A dictionary containing the version numbers for photutils and
its dependencies.
"""
versions = {'Python': sys.version.split()[0]}
packages = ('photutils', 'astropy', 'numpy', 'scipy', 'skimage',
'sklearn', 'matplotlib', 'gwcs', 'bottleneck')
for package in packages:
try:
pkg = __import__(package)
version = pkg.__version__
except ImportError:
version = None
versions[package] = version
return versions
def _get_date(utc=False):
"""
Return a string of the current date/time.
Parameters
----------
utz : bool, optional
Whether to use the UTZ timezone instead of the local timezone.
Returns
-------
result : str
The current date/time.
"""
if not utc:
now = datetime.now().astimezone()
else:
now = datetime.now(timezone.utc)
return now.strftime('%Y-%m-%d %H:%M:%S %Z')
def _get_meta(utc=False):
"""
Return a metadata dictionary with the package versions and current
date/time.
"""
return {'date': _get_date(utc=utc),
'version': _get_version_info()}
| Add all optional dependencies to version info dict | Add all optional dependencies to version info dict
| Python | bsd-3-clause | larrybradley/photutils,astropy/photutils | from datetime import datetime, timezone
+ import sys
def _get_version_info():
"""
Return a dictionary of the installed version numbers for photutils
and its dependencies.
Returns
-------
result : dict
A dictionary containing the version numbers for photutils and
its dependencies.
"""
- versions = {}
+ versions = {'Python': sys.version.split()[0]}
+
- packages = ('photutils', 'astropy', 'numpy', 'scipy', 'skimage')
+ packages = ('photutils', 'astropy', 'numpy', 'scipy', 'skimage',
+ 'sklearn', 'matplotlib', 'gwcs', 'bottleneck')
for package in packages:
try:
pkg = __import__(package)
version = pkg.__version__
except ImportError:
version = None
versions[package] = version
return versions
def _get_date(utc=False):
"""
Return a string of the current date/time.
Parameters
----------
utz : bool, optional
Whether to use the UTZ timezone instead of the local timezone.
Returns
-------
result : str
The current date/time.
"""
if not utc:
now = datetime.now().astimezone()
else:
now = datetime.now(timezone.utc)
return now.strftime('%Y-%m-%d %H:%M:%S %Z')
def _get_meta(utc=False):
"""
Return a metadata dictionary with the package versions and current
date/time.
"""
return {'date': _get_date(utc=utc),
'version': _get_version_info()}
| Add all optional dependencies to version info dict | ## Code Before:
from datetime import datetime, timezone
def _get_version_info():
"""
Return a dictionary of the installed version numbers for photutils
and its dependencies.
Returns
-------
result : dict
A dictionary containing the version numbers for photutils and
its dependencies.
"""
versions = {}
packages = ('photutils', 'astropy', 'numpy', 'scipy', 'skimage')
for package in packages:
try:
pkg = __import__(package)
version = pkg.__version__
except ImportError:
version = None
versions[package] = version
return versions
def _get_date(utc=False):
"""
Return a string of the current date/time.
Parameters
----------
utz : bool, optional
Whether to use the UTZ timezone instead of the local timezone.
Returns
-------
result : str
The current date/time.
"""
if not utc:
now = datetime.now().astimezone()
else:
now = datetime.now(timezone.utc)
return now.strftime('%Y-%m-%d %H:%M:%S %Z')
def _get_meta(utc=False):
"""
Return a metadata dictionary with the package versions and current
date/time.
"""
return {'date': _get_date(utc=utc),
'version': _get_version_info()}
## Instruction:
Add all optional dependencies to version info dict
## Code After:
from datetime import datetime, timezone
import sys
def _get_version_info():
"""
Return a dictionary of the installed version numbers for photutils
and its dependencies.
Returns
-------
result : dict
A dictionary containing the version numbers for photutils and
its dependencies.
"""
versions = {'Python': sys.version.split()[0]}
packages = ('photutils', 'astropy', 'numpy', 'scipy', 'skimage',
'sklearn', 'matplotlib', 'gwcs', 'bottleneck')
for package in packages:
try:
pkg = __import__(package)
version = pkg.__version__
except ImportError:
version = None
versions[package] = version
return versions
def _get_date(utc=False):
"""
Return a string of the current date/time.
Parameters
----------
utz : bool, optional
Whether to use the UTZ timezone instead of the local timezone.
Returns
-------
result : str
The current date/time.
"""
if not utc:
now = datetime.now().astimezone()
else:
now = datetime.now(timezone.utc)
return now.strftime('%Y-%m-%d %H:%M:%S %Z')
def _get_meta(utc=False):
"""
Return a metadata dictionary with the package versions and current
date/time.
"""
return {'date': _get_date(utc=utc),
'version': _get_version_info()}
|
a346eb2b19f3a0b72dc6565e9d0c4fabf3c5784a | migrations/versions/0014_add_template_version.py | migrations/versions/0014_add_template_version.py |
# revision identifiers, used by Alembic.
revision = '0014_add_template_version'
down_revision = '0013_add_loadtest_client'
from alembic import op
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
def upgrade():
op.add_column('jobs', sa.Column('template_version', sa.Integer(), nullable=True))
op.get_bind()
op.execute('update jobs set template_version = (select version from templates where id = template_id)')
op.add_column('notifications', sa.Column('template_version', sa.Integer(), nullable=True))
op.execute('update notifications set template_version = (select version from templates where id = template_id)')
op.alter_column('jobs', 'template_version', nullable=False)
op.alter_column('notifications', 'template_version', nullable=False)
def downgrade():
op.drop_column('notifications', 'template_version')
op.drop_column('jobs', 'template_version')
|
# revision identifiers, used by Alembic.
revision = '0014_add_template_version'
down_revision = '0013_add_loadtest_client'
from alembic import op
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
def upgrade():
op.add_column('jobs', sa.Column('template_version', sa.Integer(), nullable=True))
op.get_bind()
op.execute('update jobs set template_version = (select version from templates where id = template_id)')
op.add_column('notifications', sa.Column('template_version', sa.Integer(), nullable=True))
op.execute('update notifications set template_version = (select version from templates where id = template_id)')
op.alter_column('jobs', 'template_version', nullable=False)
op.alter_column('notifications', 'template_version', nullable=False)
# fix template_history where created_by_id is not set.
query = "update templates_history set created_by_id = " \
" (select created_by_id from templates " \
" where templates.id = templates_history.id " \
" and templates.version = templates_history.version) " \
"where templates_history.created_by_id is null"
op.execute(query)
def downgrade():
op.drop_column('notifications', 'template_version')
op.drop_column('jobs', 'template_version')
| Add a update query to fix templates_history where the created_by_id is missing. | Add a update query to fix templates_history where the created_by_id is missing.
| Python | mit | alphagov/notifications-api,alphagov/notifications-api |
# revision identifiers, used by Alembic.
revision = '0014_add_template_version'
down_revision = '0013_add_loadtest_client'
from alembic import op
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
def upgrade():
op.add_column('jobs', sa.Column('template_version', sa.Integer(), nullable=True))
op.get_bind()
op.execute('update jobs set template_version = (select version from templates where id = template_id)')
op.add_column('notifications', sa.Column('template_version', sa.Integer(), nullable=True))
op.execute('update notifications set template_version = (select version from templates where id = template_id)')
op.alter_column('jobs', 'template_version', nullable=False)
op.alter_column('notifications', 'template_version', nullable=False)
+ # fix template_history where created_by_id is not set.
+ query = "update templates_history set created_by_id = " \
+ " (select created_by_id from templates " \
+ " where templates.id = templates_history.id " \
+ " and templates.version = templates_history.version) " \
+ "where templates_history.created_by_id is null"
+ op.execute(query)
+
def downgrade():
op.drop_column('notifications', 'template_version')
op.drop_column('jobs', 'template_version')
| Add a update query to fix templates_history where the created_by_id is missing. | ## Code Before:
# revision identifiers, used by Alembic.
revision = '0014_add_template_version'
down_revision = '0013_add_loadtest_client'
from alembic import op
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
def upgrade():
op.add_column('jobs', sa.Column('template_version', sa.Integer(), nullable=True))
op.get_bind()
op.execute('update jobs set template_version = (select version from templates where id = template_id)')
op.add_column('notifications', sa.Column('template_version', sa.Integer(), nullable=True))
op.execute('update notifications set template_version = (select version from templates where id = template_id)')
op.alter_column('jobs', 'template_version', nullable=False)
op.alter_column('notifications', 'template_version', nullable=False)
def downgrade():
op.drop_column('notifications', 'template_version')
op.drop_column('jobs', 'template_version')
## Instruction:
Add a update query to fix templates_history where the created_by_id is missing.
## Code After:
# revision identifiers, used by Alembic.
revision = '0014_add_template_version'
down_revision = '0013_add_loadtest_client'
from alembic import op
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
def upgrade():
op.add_column('jobs', sa.Column('template_version', sa.Integer(), nullable=True))
op.get_bind()
op.execute('update jobs set template_version = (select version from templates where id = template_id)')
op.add_column('notifications', sa.Column('template_version', sa.Integer(), nullable=True))
op.execute('update notifications set template_version = (select version from templates where id = template_id)')
op.alter_column('jobs', 'template_version', nullable=False)
op.alter_column('notifications', 'template_version', nullable=False)
# fix template_history where created_by_id is not set.
query = "update templates_history set created_by_id = " \
" (select created_by_id from templates " \
" where templates.id = templates_history.id " \
" and templates.version = templates_history.version) " \
"where templates_history.created_by_id is null"
op.execute(query)
def downgrade():
op.drop_column('notifications', 'template_version')
op.drop_column('jobs', 'template_version')
|
df000e724ce1f307a478fcf4790404183df13610 | _setup_database.py | _setup_database.py |
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
if __name__ == '__main__':
# migrating teams from json file to database
migrate_teams(simulation=True)
# creating divisions from division configuration file
create_divisions(simulation=True)
|
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
if __name__ == '__main__':
# migrating teams from json file to database
migrate_teams(simulation=True)
# creating divisions from division configuration file
create_divisions(simulation=True)
# migrating players from json file to database
migrate_players(simulation=True)
| Include player data migration in setup | Include player data migration in setup
| Python | mit | leaffan/pynhldb |
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
+ from setup.create_players import migrate_players
if __name__ == '__main__':
# migrating teams from json file to database
migrate_teams(simulation=True)
# creating divisions from division configuration file
create_divisions(simulation=True)
+ # migrating players from json file to database
+ migrate_players(simulation=True)
| Include player data migration in setup | ## Code Before:
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
if __name__ == '__main__':
# migrating teams from json file to database
migrate_teams(simulation=True)
# creating divisions from division configuration file
create_divisions(simulation=True)
## Instruction:
Include player data migration in setup
## Code After:
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
if __name__ == '__main__':
# migrating teams from json file to database
migrate_teams(simulation=True)
# creating divisions from division configuration file
create_divisions(simulation=True)
# migrating players from json file to database
migrate_players(simulation=True)
|
c0d9a94899af84e8a075cfc7cfb054a934470e3a | static_precompiler/tests/test_templatetags.py | static_precompiler/tests/test_templatetags.py | import django.template
import django.template.loader
import pretend
def test_compile_filter(monkeypatch):
compile_static = pretend.call_recorder(lambda source_path: "compiled")
monkeypatch.setattr("static_precompiler.utils.compile_static", compile_static)
template = django.template.loader.get_template_from_string("""{% load compile_static %}{{ "source"|compile }}""")
assert template.render(django.template.Context({})) == "compiled"
monkeypatch.setattr("static_precompiler.settings.PREPEND_STATIC_URL", True)
assert template.render(django.template.Context({})) == "/static/compiled"
assert compile_static.calls == [pretend.call("source"), pretend.call("source")]
def test_inlinecompile_tag(monkeypatch):
compiler = pretend.stub(compile_source=pretend.call_recorder(lambda *args: "compiled"))
get_compiler_by_name = pretend.call_recorder(lambda *args: compiler)
monkeypatch.setattr("static_precompiler.utils.get_compiler_by_name", get_compiler_by_name)
template = django.template.loader.get_template_from_string(
"{% load compile_static %}{% inlinecompile compiler='sass' %}source{% endinlinecompile %}"
)
assert template.render(django.template.Context({})) == "compiled"
assert get_compiler_by_name.calls == [pretend.call("sass")]
assert compiler.compile_source.calls == [pretend.call("source")]
| import django.template
import pretend
def test_compile_filter(monkeypatch):
compile_static = pretend.call_recorder(lambda source_path: "compiled")
monkeypatch.setattr("static_precompiler.utils.compile_static", compile_static)
template = django.template.Template("""{% load compile_static %}{{ "source"|compile }}""")
assert template.render(django.template.Context({})) == "compiled"
monkeypatch.setattr("static_precompiler.settings.PREPEND_STATIC_URL", True)
assert template.render(django.template.Context({})) == "/static/compiled"
assert compile_static.calls == [pretend.call("source"), pretend.call("source")]
def test_inlinecompile_tag(monkeypatch):
compiler = pretend.stub(compile_source=pretend.call_recorder(lambda *args: "compiled"))
get_compiler_by_name = pretend.call_recorder(lambda *args: compiler)
monkeypatch.setattr("static_precompiler.utils.get_compiler_by_name", get_compiler_by_name)
template = django.template.Template(
"{% load compile_static %}{% inlinecompile compiler='sass' %}source{% endinlinecompile %}"
)
assert template.render(django.template.Context({})) == "compiled"
assert get_compiler_by_name.calls == [pretend.call("sass")]
assert compiler.compile_source.calls == [pretend.call("source")]
| Fix tests for Django 1.8 | Fix tests for Django 1.8
| Python | mit | liumengjun/django-static-precompiler,liumengjun/django-static-precompiler,liumengjun/django-static-precompiler,liumengjun/django-static-precompiler,jaheba/django-static-precompiler,liumengjun/django-static-precompiler,jaheba/django-static-precompiler,jaheba/django-static-precompiler,jaheba/django-static-precompiler | import django.template
- import django.template.loader
import pretend
def test_compile_filter(monkeypatch):
compile_static = pretend.call_recorder(lambda source_path: "compiled")
monkeypatch.setattr("static_precompiler.utils.compile_static", compile_static)
- template = django.template.loader.get_template_from_string("""{% load compile_static %}{{ "source"|compile }}""")
+ template = django.template.Template("""{% load compile_static %}{{ "source"|compile }}""")
assert template.render(django.template.Context({})) == "compiled"
monkeypatch.setattr("static_precompiler.settings.PREPEND_STATIC_URL", True)
assert template.render(django.template.Context({})) == "/static/compiled"
assert compile_static.calls == [pretend.call("source"), pretend.call("source")]
def test_inlinecompile_tag(monkeypatch):
compiler = pretend.stub(compile_source=pretend.call_recorder(lambda *args: "compiled"))
get_compiler_by_name = pretend.call_recorder(lambda *args: compiler)
monkeypatch.setattr("static_precompiler.utils.get_compiler_by_name", get_compiler_by_name)
- template = django.template.loader.get_template_from_string(
+ template = django.template.Template(
"{% load compile_static %}{% inlinecompile compiler='sass' %}source{% endinlinecompile %}"
)
assert template.render(django.template.Context({})) == "compiled"
assert get_compiler_by_name.calls == [pretend.call("sass")]
assert compiler.compile_source.calls == [pretend.call("source")]
| Fix tests for Django 1.8 | ## Code Before:
import django.template
import django.template.loader
import pretend
def test_compile_filter(monkeypatch):
compile_static = pretend.call_recorder(lambda source_path: "compiled")
monkeypatch.setattr("static_precompiler.utils.compile_static", compile_static)
template = django.template.loader.get_template_from_string("""{% load compile_static %}{{ "source"|compile }}""")
assert template.render(django.template.Context({})) == "compiled"
monkeypatch.setattr("static_precompiler.settings.PREPEND_STATIC_URL", True)
assert template.render(django.template.Context({})) == "/static/compiled"
assert compile_static.calls == [pretend.call("source"), pretend.call("source")]
def test_inlinecompile_tag(monkeypatch):
compiler = pretend.stub(compile_source=pretend.call_recorder(lambda *args: "compiled"))
get_compiler_by_name = pretend.call_recorder(lambda *args: compiler)
monkeypatch.setattr("static_precompiler.utils.get_compiler_by_name", get_compiler_by_name)
template = django.template.loader.get_template_from_string(
"{% load compile_static %}{% inlinecompile compiler='sass' %}source{% endinlinecompile %}"
)
assert template.render(django.template.Context({})) == "compiled"
assert get_compiler_by_name.calls == [pretend.call("sass")]
assert compiler.compile_source.calls == [pretend.call("source")]
## Instruction:
Fix tests for Django 1.8
## Code After:
import django.template
import pretend
def test_compile_filter(monkeypatch):
compile_static = pretend.call_recorder(lambda source_path: "compiled")
monkeypatch.setattr("static_precompiler.utils.compile_static", compile_static)
template = django.template.Template("""{% load compile_static %}{{ "source"|compile }}""")
assert template.render(django.template.Context({})) == "compiled"
monkeypatch.setattr("static_precompiler.settings.PREPEND_STATIC_URL", True)
assert template.render(django.template.Context({})) == "/static/compiled"
assert compile_static.calls == [pretend.call("source"), pretend.call("source")]
def test_inlinecompile_tag(monkeypatch):
compiler = pretend.stub(compile_source=pretend.call_recorder(lambda *args: "compiled"))
get_compiler_by_name = pretend.call_recorder(lambda *args: compiler)
monkeypatch.setattr("static_precompiler.utils.get_compiler_by_name", get_compiler_by_name)
template = django.template.Template(
"{% load compile_static %}{% inlinecompile compiler='sass' %}source{% endinlinecompile %}"
)
assert template.render(django.template.Context({})) == "compiled"
assert get_compiler_by_name.calls == [pretend.call("sass")]
assert compiler.compile_source.calls == [pretend.call("source")]
|
eaa4de2ecbcf29c9e56ebf2fa69099055e469fbc | tests/test_conversion.py | tests/test_conversion.py | from asciisciit import conversions as conv
import numpy as np
def test_lookup_method_equivalency():
img = np.random.randint(0, 255, (300,300), dtype=np.uint8)
pil_ascii = conv.apply_lut_pil(img)
np_ascii = conv.apply_lut_numpy(img)
assert(pil_ascii == np_ascii)
pil_ascii = conv.apply_lut_pil(img, "binary")
np_ascii = conv.apply_lut_numpy(img, "binary")
assert(pil_ascii == np_ascii) | import itertools
from asciisciit import conversions as conv
import numpy as np
import pytest
@pytest.mark.parametrize("invert,equalize,lut,lookup_func",
itertools.product((True, False),
(True, False),
("simple", "binary"),
(None, conv.apply_lut_pil)))
def test_pil_to_ascii(invert, equalize, lut, lookup_func):
img = np.random.randint(0, 255, (480, 640), dtype=np.uint8)
h, w = img.shape
expected_len = int(h*0.5*conv.ASPECTCORRECTIONFACTOR)*(int(w*0.5)+1)+1
img = conv.numpy_to_pil(img)
text = conv.pil_to_ascii(img, 0.5, invert, equalize, lut, lookup_func)
assert(len(text) == expected_len)
@pytest.mark.parametrize("invert,equalize,lut",
itertools.product((True, False),
(True, False),
("simple", "binary")))
def test_numpy_to_ascii(invert, equalize, lut):
img = np.random.randint(0, 255, (480, 640), dtype=np.uint8)
h, w = img.shape
expected_len = int(h*0.5*conv.ASPECTCORRECTIONFACTOR)*(int(w*0.5)+1)+1
text = conv.numpy_to_ascii(img, 0.5, invert, equalize, lut)
assert(len(text) == expected_len)
def test_lookup_method_equivalency():
img = np.random.randint(0, 255, (300,300), dtype=np.uint8)
pil_ascii = conv.apply_lut_pil(img)
np_ascii = conv.apply_lut_numpy(img)
assert(pil_ascii == np_ascii)
pil_ascii = conv.apply_lut_pil(img, "binary")
np_ascii = conv.apply_lut_numpy(img, "binary")
assert(pil_ascii == np_ascii)
| Add tests to minimally exercise basic conversion functionality | Add tests to minimally exercise basic conversion functionality
| Python | mit | derricw/asciisciit | + import itertools
from asciisciit import conversions as conv
import numpy as np
+ import pytest
+
+
+ @pytest.mark.parametrize("invert,equalize,lut,lookup_func",
+ itertools.product((True, False),
+ (True, False),
+ ("simple", "binary"),
+ (None, conv.apply_lut_pil)))
+ def test_pil_to_ascii(invert, equalize, lut, lookup_func):
+ img = np.random.randint(0, 255, (480, 640), dtype=np.uint8)
+ h, w = img.shape
+ expected_len = int(h*0.5*conv.ASPECTCORRECTIONFACTOR)*(int(w*0.5)+1)+1
+ img = conv.numpy_to_pil(img)
+ text = conv.pil_to_ascii(img, 0.5, invert, equalize, lut, lookup_func)
+ assert(len(text) == expected_len)
+
+
+ @pytest.mark.parametrize("invert,equalize,lut",
+ itertools.product((True, False),
+ (True, False),
+ ("simple", "binary")))
+ def test_numpy_to_ascii(invert, equalize, lut):
+ img = np.random.randint(0, 255, (480, 640), dtype=np.uint8)
+ h, w = img.shape
+ expected_len = int(h*0.5*conv.ASPECTCORRECTIONFACTOR)*(int(w*0.5)+1)+1
+ text = conv.numpy_to_ascii(img, 0.5, invert, equalize, lut)
+ assert(len(text) == expected_len)
def test_lookup_method_equivalency():
img = np.random.randint(0, 255, (300,300), dtype=np.uint8)
pil_ascii = conv.apply_lut_pil(img)
np_ascii = conv.apply_lut_numpy(img)
assert(pil_ascii == np_ascii)
pil_ascii = conv.apply_lut_pil(img, "binary")
np_ascii = conv.apply_lut_numpy(img, "binary")
assert(pil_ascii == np_ascii)
+ | Add tests to minimally exercise basic conversion functionality | ## Code Before:
from asciisciit import conversions as conv
import numpy as np
def test_lookup_method_equivalency():
img = np.random.randint(0, 255, (300,300), dtype=np.uint8)
pil_ascii = conv.apply_lut_pil(img)
np_ascii = conv.apply_lut_numpy(img)
assert(pil_ascii == np_ascii)
pil_ascii = conv.apply_lut_pil(img, "binary")
np_ascii = conv.apply_lut_numpy(img, "binary")
assert(pil_ascii == np_ascii)
## Instruction:
Add tests to minimally exercise basic conversion functionality
## Code After:
import itertools
from asciisciit import conversions as conv
import numpy as np
import pytest
@pytest.mark.parametrize("invert,equalize,lut,lookup_func",
itertools.product((True, False),
(True, False),
("simple", "binary"),
(None, conv.apply_lut_pil)))
def test_pil_to_ascii(invert, equalize, lut, lookup_func):
img = np.random.randint(0, 255, (480, 640), dtype=np.uint8)
h, w = img.shape
expected_len = int(h*0.5*conv.ASPECTCORRECTIONFACTOR)*(int(w*0.5)+1)+1
img = conv.numpy_to_pil(img)
text = conv.pil_to_ascii(img, 0.5, invert, equalize, lut, lookup_func)
assert(len(text) == expected_len)
@pytest.mark.parametrize("invert,equalize,lut",
itertools.product((True, False),
(True, False),
("simple", "binary")))
def test_numpy_to_ascii(invert, equalize, lut):
img = np.random.randint(0, 255, (480, 640), dtype=np.uint8)
h, w = img.shape
expected_len = int(h*0.5*conv.ASPECTCORRECTIONFACTOR)*(int(w*0.5)+1)+1
text = conv.numpy_to_ascii(img, 0.5, invert, equalize, lut)
assert(len(text) == expected_len)
def test_lookup_method_equivalency():
img = np.random.randint(0, 255, (300,300), dtype=np.uint8)
pil_ascii = conv.apply_lut_pil(img)
np_ascii = conv.apply_lut_numpy(img)
assert(pil_ascii == np_ascii)
pil_ascii = conv.apply_lut_pil(img, "binary")
np_ascii = conv.apply_lut_numpy(img, "binary")
assert(pil_ascii == np_ascii)
|
a0420b066e0a5064ebe0944a16348debf107a9a4 | speech.py | speech.py |
import time
from say import say
def introduction(tts):
"""Make Nao introduce itself.
Keyword arguments:
tts - Nao proxy.
"""
say("Hello world!", tts)
say("Computer Science is one of the coolest subjects to study in the\
modern world.", tts)
say("Programming is a tool used by Scientists and Engineers to create\
all kinds of interesting things.", tts)
say("For example, you can program mobile phones, laptops, cars, the\
Internet and more. Every day more things are being enhanced by\
Artificial Intelligence, like me.", tts)
say("Now I want to talk about what programming is like.", tts)
say("Programming is about solving puzzles of the real world and helping\
people deal with important problems to make the world a better\
place.", tts)
time.sleep(2)
say("Now, what topic would you like to learn about? Show me a number to\
choose the activity.", tts)
say("Number one to practice input and output.", tts)
|
import time
from say import say
def introduction(tts):
"""Make Nao introduce itself.
Keyword arguments:
tts - Nao proxy.
"""
say("Hello world!", tts)
say("I'm Leyva and I will teach you how to program in a programming\
language called python", tts)
say("Computer Science is one of the coolest subjects to study in the\
modern world.", tts)
say("Programming is a tool used by Scientists and Engineers to create\
all kinds of interesting things.", tts)
say("For example, you can program mobile phones, laptops, cars, the\
Internet and more. Every day more things are being enhanced by\
Artificial Intelligence, like me.", tts)
say("Now I want to talk about what programming is like.", tts)
say("Programming is about solving puzzles of the real world and helping\
people deal with important problems to make the world a better\
place.", tts)
time.sleep(2)
say("Now, what topic would you like to learn about? Show me a number to\
choose the activity.", tts)
say("Number one to practice input and output.", tts)
| Make Nao say its name and the programming language | Make Nao say its name and the programming language
| Python | mit | AliGhahraei/nao-classroom |
import time
from say import say
def introduction(tts):
"""Make Nao introduce itself.
Keyword arguments:
tts - Nao proxy.
"""
say("Hello world!", tts)
+ say("I'm Leyva and I will teach you how to program in a programming\
+ language called python", tts)
say("Computer Science is one of the coolest subjects to study in the\
modern world.", tts)
say("Programming is a tool used by Scientists and Engineers to create\
all kinds of interesting things.", tts)
say("For example, you can program mobile phones, laptops, cars, the\
Internet and more. Every day more things are being enhanced by\
Artificial Intelligence, like me.", tts)
say("Now I want to talk about what programming is like.", tts)
say("Programming is about solving puzzles of the real world and helping\
people deal with important problems to make the world a better\
place.", tts)
time.sleep(2)
say("Now, what topic would you like to learn about? Show me a number to\
choose the activity.", tts)
say("Number one to practice input and output.", tts)
| Make Nao say its name and the programming language | ## Code Before:
import time
from say import say
def introduction(tts):
"""Make Nao introduce itself.
Keyword arguments:
tts - Nao proxy.
"""
say("Hello world!", tts)
say("Computer Science is one of the coolest subjects to study in the\
modern world.", tts)
say("Programming is a tool used by Scientists and Engineers to create\
all kinds of interesting things.", tts)
say("For example, you can program mobile phones, laptops, cars, the\
Internet and more. Every day more things are being enhanced by\
Artificial Intelligence, like me.", tts)
say("Now I want to talk about what programming is like.", tts)
say("Programming is about solving puzzles of the real world and helping\
people deal with important problems to make the world a better\
place.", tts)
time.sleep(2)
say("Now, what topic would you like to learn about? Show me a number to\
choose the activity.", tts)
say("Number one to practice input and output.", tts)
## Instruction:
Make Nao say its name and the programming language
## Code After:
import time
from say import say
def introduction(tts):
"""Make Nao introduce itself.
Keyword arguments:
tts - Nao proxy.
"""
say("Hello world!", tts)
say("I'm Leyva and I will teach you how to program in a programming\
language called python", tts)
say("Computer Science is one of the coolest subjects to study in the\
modern world.", tts)
say("Programming is a tool used by Scientists and Engineers to create\
all kinds of interesting things.", tts)
say("For example, you can program mobile phones, laptops, cars, the\
Internet and more. Every day more things are being enhanced by\
Artificial Intelligence, like me.", tts)
say("Now I want to talk about what programming is like.", tts)
say("Programming is about solving puzzles of the real world and helping\
people deal with important problems to make the world a better\
place.", tts)
time.sleep(2)
say("Now, what topic would you like to learn about? Show me a number to\
choose the activity.", tts)
say("Number one to practice input and output.", tts)
|
40122e169e6a887caa6371a0ff3029c35ce265d5 | third_party/node/node.py | third_party/node/node.py |
from os import path as os_path
import platform
import subprocess
import sys
import os
def GetBinaryPath():
return os_path.join(
os_path.dirname(__file__), *{
'Darwin': ('mac', 'node-darwin-x64', 'bin', 'node'),
'Linux': ('linux', 'node-linux-x64', 'bin', 'node'),
'Windows': ('win', 'node.exe'),
}[platform.system()])
def RunNode(cmd_parts, output=subprocess.PIPE):
cmd = [GetBinaryPath()] + cmd_parts
process = subprocess.Popen(cmd,
cwd=os.getcwd(),
stdout=output,
stderr=output)
stdout, stderr = process.communicate()
if process.returncode != 0:
print('%s failed:\n%s\n%s' % (cmd, stdout, stderr))
exit(process.returncode)
return stdout
if __name__ == '__main__':
args = sys.argv[1:]
# Accept --output as the first argument, and then remove
# it from the args entirely if present.
if len(args) > 0 and args[0] == '--output':
output = None
args = sys.argv[2:]
else:
output = subprocess.PIPE
RunNode(args, output)
|
from os import path as os_path
import platform
import subprocess
import sys
import os
def GetBinaryPath():
return os_path.join(
os_path.dirname(__file__), *{
'Darwin': ('mac', 'node-darwin-x64', 'bin', 'node'),
'Linux': ('linux', 'node-linux-x64', 'bin', 'node'),
'Windows': ('win', 'node.exe'),
}[platform.system()])
def RunNode(cmd_parts, output=subprocess.PIPE):
cmd = [GetBinaryPath()] + cmd_parts
process = subprocess.Popen(cmd,
cwd=os.getcwd(),
stdout=output,
stderr=output,
universal_newlines=True)
stdout, stderr = process.communicate()
if process.returncode != 0:
print('%s failed:\n%s\n%s' % (cmd, stdout, stderr))
exit(process.returncode)
return stdout
if __name__ == '__main__':
args = sys.argv[1:]
# Accept --output as the first argument, and then remove
# it from the args entirely if present.
if len(args) > 0 and args[0] == '--output':
output = None
args = sys.argv[2:]
else:
output = subprocess.PIPE
RunNode(args, output)
| Fix line ending printing on Python 3 | Fix line ending printing on Python 3
To reflect the changes in
https://chromium-review.googlesource.com/c/chromium/src/+/2896248/8/third_party/node/node.py
[email protected]
Bug: none
Change-Id: I25ba29042f537bfef57fba93115be2c194649864
Reviewed-on: https://chromium-review.googlesource.com/c/devtools/devtools-frontend/+/2914883
Commit-Queue: Tim van der Lippe <[email protected]>
Commit-Queue: Jack Franklin <[email protected]>
Auto-Submit: Tim van der Lippe <[email protected]>
Reviewed-by: Jack Franklin <[email protected]>
| Python | bsd-3-clause | ChromeDevTools/devtools-frontend,ChromeDevTools/devtools-frontend,ChromeDevTools/devtools-frontend,ChromeDevTools/devtools-frontend,ChromeDevTools/devtools-frontend,ChromeDevTools/devtools-frontend |
from os import path as os_path
import platform
import subprocess
import sys
import os
def GetBinaryPath():
return os_path.join(
os_path.dirname(__file__), *{
'Darwin': ('mac', 'node-darwin-x64', 'bin', 'node'),
'Linux': ('linux', 'node-linux-x64', 'bin', 'node'),
'Windows': ('win', 'node.exe'),
}[platform.system()])
def RunNode(cmd_parts, output=subprocess.PIPE):
cmd = [GetBinaryPath()] + cmd_parts
process = subprocess.Popen(cmd,
cwd=os.getcwd(),
stdout=output,
- stderr=output)
+ stderr=output,
+ universal_newlines=True)
stdout, stderr = process.communicate()
if process.returncode != 0:
print('%s failed:\n%s\n%s' % (cmd, stdout, stderr))
exit(process.returncode)
return stdout
if __name__ == '__main__':
args = sys.argv[1:]
# Accept --output as the first argument, and then remove
# it from the args entirely if present.
if len(args) > 0 and args[0] == '--output':
output = None
args = sys.argv[2:]
else:
output = subprocess.PIPE
RunNode(args, output)
| Fix line ending printing on Python 3 | ## Code Before:
from os import path as os_path
import platform
import subprocess
import sys
import os
def GetBinaryPath():
return os_path.join(
os_path.dirname(__file__), *{
'Darwin': ('mac', 'node-darwin-x64', 'bin', 'node'),
'Linux': ('linux', 'node-linux-x64', 'bin', 'node'),
'Windows': ('win', 'node.exe'),
}[platform.system()])
def RunNode(cmd_parts, output=subprocess.PIPE):
cmd = [GetBinaryPath()] + cmd_parts
process = subprocess.Popen(cmd,
cwd=os.getcwd(),
stdout=output,
stderr=output)
stdout, stderr = process.communicate()
if process.returncode != 0:
print('%s failed:\n%s\n%s' % (cmd, stdout, stderr))
exit(process.returncode)
return stdout
if __name__ == '__main__':
args = sys.argv[1:]
# Accept --output as the first argument, and then remove
# it from the args entirely if present.
if len(args) > 0 and args[0] == '--output':
output = None
args = sys.argv[2:]
else:
output = subprocess.PIPE
RunNode(args, output)
## Instruction:
Fix line ending printing on Python 3
## Code After:
from os import path as os_path
import platform
import subprocess
import sys
import os
def GetBinaryPath():
return os_path.join(
os_path.dirname(__file__), *{
'Darwin': ('mac', 'node-darwin-x64', 'bin', 'node'),
'Linux': ('linux', 'node-linux-x64', 'bin', 'node'),
'Windows': ('win', 'node.exe'),
}[platform.system()])
def RunNode(cmd_parts, output=subprocess.PIPE):
cmd = [GetBinaryPath()] + cmd_parts
process = subprocess.Popen(cmd,
cwd=os.getcwd(),
stdout=output,
stderr=output,
universal_newlines=True)
stdout, stderr = process.communicate()
if process.returncode != 0:
print('%s failed:\n%s\n%s' % (cmd, stdout, stderr))
exit(process.returncode)
return stdout
if __name__ == '__main__':
args = sys.argv[1:]
# Accept --output as the first argument, and then remove
# it from the args entirely if present.
if len(args) > 0 and args[0] == '--output':
output = None
args = sys.argv[2:]
else:
output = subprocess.PIPE
RunNode(args, output)
|
91b0bcad27f12c29681235079960ee272275e0bd | src/main/python/dlfa/solvers/__init__.py | src/main/python/dlfa/solvers/__init__.py | from .nn_solver import NNSolver
from .lstm_solver import LSTMSolver
from .tree_lstm_solver import TreeLSTMSolver
from .memory_network import MemoryNetworkSolver
from .differentiable_search import DifferentiableSearchSolver
concrete_solvers = { # pylint: disable=invalid-name
'LSTMSolver': LSTMSolver,
'TreeLSTMSolver': TreeLSTMSolver,
'MemoryNetworkSolver': MemoryNetworkSolver,
'DifferentiableSearchSolver': DifferentiableSearchSolver,
}
| from .nn_solver import NNSolver
from .lstm_solver import LSTMSolver
from .tree_lstm_solver import TreeLSTMSolver
from .memory_network import MemoryNetworkSolver
from .differentiable_search import DifferentiableSearchSolver
from .multiple_choice_memory_network import MultipleChoiceMemoryNetworkSolver
concrete_solvers = { # pylint: disable=invalid-name
'LSTMSolver': LSTMSolver,
'TreeLSTMSolver': TreeLSTMSolver,
'MemoryNetworkSolver': MemoryNetworkSolver,
'DifferentiableSearchSolver': DifferentiableSearchSolver,
'MultipleChoiceMemoryNetworkSolver': MultipleChoiceMemoryNetworkSolver,
}
| Add MCMemoryNetwork as a usable solver | Add MCMemoryNetwork as a usable solver
| Python | apache-2.0 | allenai/deep_qa,matt-gardner/deep_qa,DeNeutoy/deep_qa,allenai/deep_qa,DeNeutoy/deep_qa,matt-gardner/deep_qa | from .nn_solver import NNSolver
from .lstm_solver import LSTMSolver
from .tree_lstm_solver import TreeLSTMSolver
from .memory_network import MemoryNetworkSolver
from .differentiable_search import DifferentiableSearchSolver
+ from .multiple_choice_memory_network import MultipleChoiceMemoryNetworkSolver
concrete_solvers = { # pylint: disable=invalid-name
'LSTMSolver': LSTMSolver,
'TreeLSTMSolver': TreeLSTMSolver,
'MemoryNetworkSolver': MemoryNetworkSolver,
'DifferentiableSearchSolver': DifferentiableSearchSolver,
+ 'MultipleChoiceMemoryNetworkSolver': MultipleChoiceMemoryNetworkSolver,
}
| Add MCMemoryNetwork as a usable solver | ## Code Before:
from .nn_solver import NNSolver
from .lstm_solver import LSTMSolver
from .tree_lstm_solver import TreeLSTMSolver
from .memory_network import MemoryNetworkSolver
from .differentiable_search import DifferentiableSearchSolver
concrete_solvers = { # pylint: disable=invalid-name
'LSTMSolver': LSTMSolver,
'TreeLSTMSolver': TreeLSTMSolver,
'MemoryNetworkSolver': MemoryNetworkSolver,
'DifferentiableSearchSolver': DifferentiableSearchSolver,
}
## Instruction:
Add MCMemoryNetwork as a usable solver
## Code After:
from .nn_solver import NNSolver
from .lstm_solver import LSTMSolver
from .tree_lstm_solver import TreeLSTMSolver
from .memory_network import MemoryNetworkSolver
from .differentiable_search import DifferentiableSearchSolver
from .multiple_choice_memory_network import MultipleChoiceMemoryNetworkSolver
concrete_solvers = { # pylint: disable=invalid-name
'LSTMSolver': LSTMSolver,
'TreeLSTMSolver': TreeLSTMSolver,
'MemoryNetworkSolver': MemoryNetworkSolver,
'DifferentiableSearchSolver': DifferentiableSearchSolver,
'MultipleChoiceMemoryNetworkSolver': MultipleChoiceMemoryNetworkSolver,
}
|
68e51fd9772aaf4bd382ffe05721e27da6bddde6 | argus/exceptions.py | argus/exceptions.py | """Various exceptions that can be raised by the CI project."""
class ArgusError(Exception):
pass
class ArgusTimeoutError(ArgusError):
pass
class ArgusCLIError(ArgusError):
pass
class ArgusPermissionDenied(ArgusError):
pass
class ArgusHeatTeardown(ArgusError):
pass
class ArgusEnvironmentError(ArgusError):
"""Base class for errors related to the argus environment."""
pass
| """Various exceptions that can be raised by the CI project."""
class ArgusError(Exception):
pass
class ArgusTimeoutError(ArgusError):
pass
class ArgusCLIError(ArgusError):
pass
class ArgusPermissionDenied(ArgusError):
pass
class ArgusHeatTeardown(ArgusError):
pass
class ArgusEnvironmentError(ArgusError):
"""Base class for errors related to the argus environment."""
pass
class ErrorArgusInvalidDecorator(ArgusError):
"""The `skip_on_os` decorator was used improperly."""
pass
| Add exception for Logic Errors | Add exception for Logic Errors
| Python | apache-2.0 | stefan-caraiman/cloudbase-init-ci,cloudbase/cloudbase-init-ci,micumatei/cloudbase-init-ci | """Various exceptions that can be raised by the CI project."""
class ArgusError(Exception):
pass
class ArgusTimeoutError(ArgusError):
pass
class ArgusCLIError(ArgusError):
pass
class ArgusPermissionDenied(ArgusError):
pass
class ArgusHeatTeardown(ArgusError):
pass
class ArgusEnvironmentError(ArgusError):
"""Base class for errors related to the argus environment."""
pass
+
+ class ErrorArgusInvalidDecorator(ArgusError):
+ """The `skip_on_os` decorator was used improperly."""
+ pass
+ | Add exception for Logic Errors | ## Code Before:
"""Various exceptions that can be raised by the CI project."""
class ArgusError(Exception):
pass
class ArgusTimeoutError(ArgusError):
pass
class ArgusCLIError(ArgusError):
pass
class ArgusPermissionDenied(ArgusError):
pass
class ArgusHeatTeardown(ArgusError):
pass
class ArgusEnvironmentError(ArgusError):
"""Base class for errors related to the argus environment."""
pass
## Instruction:
Add exception for Logic Errors
## Code After:
"""Various exceptions that can be raised by the CI project."""
class ArgusError(Exception):
pass
class ArgusTimeoutError(ArgusError):
pass
class ArgusCLIError(ArgusError):
pass
class ArgusPermissionDenied(ArgusError):
pass
class ArgusHeatTeardown(ArgusError):
pass
class ArgusEnvironmentError(ArgusError):
"""Base class for errors related to the argus environment."""
pass
class ErrorArgusInvalidDecorator(ArgusError):
"""The `skip_on_os` decorator was used improperly."""
pass
|
2eb4fcb2d75e6f93f33b1ae41098919f6d4ebc92 | neovim/__init__.py | neovim/__init__.py | from client import Client
from uv_stream import UvStream
__all__ = ['Client', 'UvStream', 'c']
| from client import Client
from uv_stream import UvStream
__all__ = ['connect']
def connect(address, port=None):
client = Client(UvStream(address, port))
client.discover_api()
return client.vim
| Add helper function for connecting with Neovim | Add helper function for connecting with Neovim
| Python | apache-2.0 | fwalch/python-client,Shougo/python-client,bfredl/python-client,brcolow/python-client,starcraftman/python-client,traverseda/python-client,justinmk/python-client,neovim/python-client,meitham/python-client,meitham/python-client,0x90sled/python-client,justinmk/python-client,zchee/python-client,traverseda/python-client,bfredl/python-client,neovim/python-client,starcraftman/python-client,0x90sled/python-client,fwalch/python-client,Shougo/python-client,zchee/python-client,timeyyy/python-client,brcolow/python-client,timeyyy/python-client | from client import Client
from uv_stream import UvStream
- __all__ = ['Client', 'UvStream', 'c']
+ __all__ = ['connect']
+ def connect(address, port=None):
+ client = Client(UvStream(address, port))
+ client.discover_api()
+ return client.vim
+ | Add helper function for connecting with Neovim | ## Code Before:
from client import Client
from uv_stream import UvStream
__all__ = ['Client', 'UvStream', 'c']
## Instruction:
Add helper function for connecting with Neovim
## Code After:
from client import Client
from uv_stream import UvStream
__all__ = ['connect']
def connect(address, port=None):
client = Client(UvStream(address, port))
client.discover_api()
return client.vim
|
f71897270d9a040930bfb41801bf955ea3d0a36e | slave/skia_slave_scripts/android_run_gm.py | slave/skia_slave_scripts/android_run_gm.py |
""" Run GM on an Android device. """
from android_build_step import AndroidBuildStep
from build_step import BuildStep
from run_gm import RunGM
import sys
class AndroidRunGM(AndroidBuildStep, RunGM):
def _Run(self):
self._gm_args.append('--nopdf')
RunGM._Run(self)
if '__main__' == __name__:
sys.exit(BuildStep.RunBuildStep(AndroidRunGM))
|
""" Run GM on an Android device. """
from android_build_step import AndroidBuildStep
from build_step import BuildStep
from run_gm import RunGM
import sys
class AndroidRunGM(AndroidBuildStep, RunGM):
def __init__(self, timeout=9600, **kwargs):
super(AndroidRunGM, self).__init__(timeout=timeout, **kwargs)
def _Run(self):
self._gm_args.append('--nopdf')
RunGM._Run(self)
if '__main__' == __name__:
sys.exit(BuildStep.RunBuildStep(AndroidRunGM))
| Increase timeout for GM on Android | Increase timeout for GM on Android
Unreviewed.
(SkipBuildbotRuns)
Review URL: https://codereview.chromium.org/18845002
git-svn-id: 32fc27f4dcfb6c0385cd9719852b95fe6680452d@9906 2bbb7eff-a529-9590-31e7-b0007b416f81
| Python | bsd-3-clause | google/skia-buildbot,Tiger66639/skia-buildbot,google/skia-buildbot,Tiger66639/skia-buildbot,google/skia-buildbot,google/skia-buildbot,Tiger66639/skia-buildbot,google/skia-buildbot,Tiger66639/skia-buildbot,google/skia-buildbot,Tiger66639/skia-buildbot,google/skia-buildbot,Tiger66639/skia-buildbot,google/skia-buildbot,Tiger66639/skia-buildbot |
""" Run GM on an Android device. """
from android_build_step import AndroidBuildStep
from build_step import BuildStep
from run_gm import RunGM
import sys
class AndroidRunGM(AndroidBuildStep, RunGM):
+ def __init__(self, timeout=9600, **kwargs):
+ super(AndroidRunGM, self).__init__(timeout=timeout, **kwargs)
+
def _Run(self):
self._gm_args.append('--nopdf')
RunGM._Run(self)
if '__main__' == __name__:
sys.exit(BuildStep.RunBuildStep(AndroidRunGM))
| Increase timeout for GM on Android | ## Code Before:
""" Run GM on an Android device. """
from android_build_step import AndroidBuildStep
from build_step import BuildStep
from run_gm import RunGM
import sys
class AndroidRunGM(AndroidBuildStep, RunGM):
def _Run(self):
self._gm_args.append('--nopdf')
RunGM._Run(self)
if '__main__' == __name__:
sys.exit(BuildStep.RunBuildStep(AndroidRunGM))
## Instruction:
Increase timeout for GM on Android
## Code After:
""" Run GM on an Android device. """
from android_build_step import AndroidBuildStep
from build_step import BuildStep
from run_gm import RunGM
import sys
class AndroidRunGM(AndroidBuildStep, RunGM):
def __init__(self, timeout=9600, **kwargs):
super(AndroidRunGM, self).__init__(timeout=timeout, **kwargs)
def _Run(self):
self._gm_args.append('--nopdf')
RunGM._Run(self)
if '__main__' == __name__:
sys.exit(BuildStep.RunBuildStep(AndroidRunGM))
|
3e1033637d0bb5dd8856052ffd1667df0ccceb5d | entity.py | entity.py | from pygame.sprite import Sprite
class Entity(Sprite):
"""An object within the game.
It contains several Component objects that are used to define how it
is handled graphically and physically, as well as its behaviour.
Since it inherits from PyGame's Sprite class, it can be added to
any Groups you have created. (Components will automatically add it
to the appropriate Groups.)
This also makes discarding it from the game simple, as it will
automatically be removed from memory once it belongs to no Groups.
Attributes:
components (list): Contains all of the Component objects that
are contained in this Entity.
* Note that components will also be added as unique attributes
automatically. This will make it possible to access each
component directly, rather than having to add .components.
in-between this Entity's name and the component's name.
"""
components = []
class Component(object):
"""Part of an Entity object.
It will handle one facet of the Entity, which can be graphics,
physics, or part of its behaviour. Subclasses should define their
own attributes and methods in order to accomplish this.
Ideally, a game would have many Component subclasses to define the
many different parts that make up different game objects.
Attributes:
entity (Entity): This Component is bound to it and has access
to all of its members.
"""
pass
| from pygame.sprite import Sprite
class Entity(Sprite):
"""An object within the game.
It contains several Component objects that are used to define how it
is handled graphically and physically, as well as its behaviour.
Since it inherits from PyGame's Sprite class, it can be added to
any Groups you have created. (Components will automatically add it
to the appropriate Groups.)
This also makes discarding it from the game simple, as it will
automatically be removed from memory once it belongs to no Groups.
Attributes:
components (list): Contains all of the Component objects that
are contained in this Entity.
* Note that components will also be added as unique attributes
automatically. This will make it possible to access each
component directly, rather than having to add .components.
in-between this Entity's name and the component's name.
"""
pass
class Component(object):
"""Part of an Entity object.
It will handle one facet of the Entity, which can be graphics,
physics, or part of its behaviour. Subclasses should define their
own attributes and methods in order to accomplish this.
Ideally, a game would have many Component subclasses to define the
many different parts that make up different game objects.
Attributes:
entity (Entity): This Component is bound to it and has access
to all of its members.
"""
pass
| Remove components as class attribute | Entity: Remove components as class attribute
It is meant to be an instance attribute and should be defined in
init() instead.
| Python | unlicense | MarquisLP/gamehappy | from pygame.sprite import Sprite
class Entity(Sprite):
"""An object within the game.
It contains several Component objects that are used to define how it
is handled graphically and physically, as well as its behaviour.
Since it inherits from PyGame's Sprite class, it can be added to
any Groups you have created. (Components will automatically add it
to the appropriate Groups.)
This also makes discarding it from the game simple, as it will
automatically be removed from memory once it belongs to no Groups.
Attributes:
components (list): Contains all of the Component objects that
are contained in this Entity.
* Note that components will also be added as unique attributes
automatically. This will make it possible to access each
component directly, rather than having to add .components.
in-between this Entity's name and the component's name.
"""
- components = []
+ pass
class Component(object):
"""Part of an Entity object.
It will handle one facet of the Entity, which can be graphics,
physics, or part of its behaviour. Subclasses should define their
own attributes and methods in order to accomplish this.
Ideally, a game would have many Component subclasses to define the
many different parts that make up different game objects.
Attributes:
entity (Entity): This Component is bound to it and has access
to all of its members.
"""
pass
| Remove components as class attribute | ## Code Before:
from pygame.sprite import Sprite
class Entity(Sprite):
"""An object within the game.
It contains several Component objects that are used to define how it
is handled graphically and physically, as well as its behaviour.
Since it inherits from PyGame's Sprite class, it can be added to
any Groups you have created. (Components will automatically add it
to the appropriate Groups.)
This also makes discarding it from the game simple, as it will
automatically be removed from memory once it belongs to no Groups.
Attributes:
components (list): Contains all of the Component objects that
are contained in this Entity.
* Note that components will also be added as unique attributes
automatically. This will make it possible to access each
component directly, rather than having to add .components.
in-between this Entity's name and the component's name.
"""
components = []
class Component(object):
"""Part of an Entity object.
It will handle one facet of the Entity, which can be graphics,
physics, or part of its behaviour. Subclasses should define their
own attributes and methods in order to accomplish this.
Ideally, a game would have many Component subclasses to define the
many different parts that make up different game objects.
Attributes:
entity (Entity): This Component is bound to it and has access
to all of its members.
"""
pass
## Instruction:
Remove components as class attribute
## Code After:
from pygame.sprite import Sprite
class Entity(Sprite):
"""An object within the game.
It contains several Component objects that are used to define how it
is handled graphically and physically, as well as its behaviour.
Since it inherits from PyGame's Sprite class, it can be added to
any Groups you have created. (Components will automatically add it
to the appropriate Groups.)
This also makes discarding it from the game simple, as it will
automatically be removed from memory once it belongs to no Groups.
Attributes:
components (list): Contains all of the Component objects that
are contained in this Entity.
* Note that components will also be added as unique attributes
automatically. This will make it possible to access each
component directly, rather than having to add .components.
in-between this Entity's name and the component's name.
"""
pass
class Component(object):
"""Part of an Entity object.
It will handle one facet of the Entity, which can be graphics,
physics, or part of its behaviour. Subclasses should define their
own attributes and methods in order to accomplish this.
Ideally, a game would have many Component subclasses to define the
many different parts that make up different game objects.
Attributes:
entity (Entity): This Component is bound to it and has access
to all of its members.
"""
pass
|
368e2d6407cb021d80fe3679c65737581c3cc221 | bliski_publikator/institutions/serializers.py | bliski_publikator/institutions/serializers.py | from rest_framework import serializers
from .models import Institution
class InstitutionSerializer(serializers.HyperlinkedModelSerializer):
on_site = serializers.CharField(source='get_absolute_url', read_only=True)
class Meta:
model = Institution
fields = ('on_site',
'url',
'name',
'slug',
'user',
'email',
'region',
'regon',
'krs',
'monitorings')
extra_kwargs = {
'region': {'view_name': 'jednostkaadministracyjna-detail'}
}
| from rest_framework import serializers
from .models import Institution
class InstitutionSerializer(serializers.HyperlinkedModelSerializer):
on_site = serializers.CharField(source='get_absolute_url', read_only=True)
class Meta:
model = Institution
fields = ('on_site',
'url',
'name',
'slug',
'user',
'email',
'region',
'regon',
'krs',
'monitorings')
extra_kwargs = {
'region': {'view_name': 'jednostkaadministracyjna-detail'},
'user': {'read_only': True}
}
| Make user field read-only in InstitutionSerializer | Make user field read-only in InstitutionSerializer
| Python | mit | watchdogpolska/bliski_publikator,watchdogpolska/bliski_publikator,watchdogpolska/bliski_publikator,watchdogpolska/bliski_publikator | from rest_framework import serializers
from .models import Institution
class InstitutionSerializer(serializers.HyperlinkedModelSerializer):
on_site = serializers.CharField(source='get_absolute_url', read_only=True)
class Meta:
model = Institution
fields = ('on_site',
'url',
'name',
'slug',
'user',
'email',
'region',
'regon',
'krs',
'monitorings')
extra_kwargs = {
- 'region': {'view_name': 'jednostkaadministracyjna-detail'}
+ 'region': {'view_name': 'jednostkaadministracyjna-detail'},
+ 'user': {'read_only': True}
}
| Make user field read-only in InstitutionSerializer | ## Code Before:
from rest_framework import serializers
from .models import Institution
class InstitutionSerializer(serializers.HyperlinkedModelSerializer):
on_site = serializers.CharField(source='get_absolute_url', read_only=True)
class Meta:
model = Institution
fields = ('on_site',
'url',
'name',
'slug',
'user',
'email',
'region',
'regon',
'krs',
'monitorings')
extra_kwargs = {
'region': {'view_name': 'jednostkaadministracyjna-detail'}
}
## Instruction:
Make user field read-only in InstitutionSerializer
## Code After:
from rest_framework import serializers
from .models import Institution
class InstitutionSerializer(serializers.HyperlinkedModelSerializer):
on_site = serializers.CharField(source='get_absolute_url', read_only=True)
class Meta:
model = Institution
fields = ('on_site',
'url',
'name',
'slug',
'user',
'email',
'region',
'regon',
'krs',
'monitorings')
extra_kwargs = {
'region': {'view_name': 'jednostkaadministracyjna-detail'},
'user': {'read_only': True}
}
|
5cadfe1b0a8da0f46950ff63786bf79c42af9b90 | tutorials/forms.py | tutorials/forms.py | from django import forms
from .models import Tutorial
class TutorialForm(forms.ModelForm):
# ToDO: Set required fields??
class Meta:
model = Tutorial
fields = ('title', 'html', 'markdown') | from django import forms
from .models import Tutorial
class TutorialForm(forms.ModelForm):
# ToDO: Set required fields??
class Meta:
model = Tutorial
fields = ('category', 'title', 'markdown', 'level')
| Add new model fields to form | Add new model fields to form
| Python | agpl-3.0 | openego/oeplatform,openego/oeplatform,openego/oeplatform,openego/oeplatform | from django import forms
from .models import Tutorial
class TutorialForm(forms.ModelForm):
# ToDO: Set required fields??
class Meta:
model = Tutorial
- fields = ('title', 'html', 'markdown')
+ fields = ('category', 'title', 'markdown', 'level')
+ | Add new model fields to form | ## Code Before:
from django import forms
from .models import Tutorial
class TutorialForm(forms.ModelForm):
# ToDO: Set required fields??
class Meta:
model = Tutorial
fields = ('title', 'html', 'markdown')
## Instruction:
Add new model fields to form
## Code After:
from django import forms
from .models import Tutorial
class TutorialForm(forms.ModelForm):
# ToDO: Set required fields??
class Meta:
model = Tutorial
fields = ('category', 'title', 'markdown', 'level')
|
1abb838a1fa56af25b9c6369dff93c65e17fbc3a | manage.py | manage.py | import os
COV = None
if os.environ.get('FLASK_COVERAGE'):
import coverage
COV = coverage.coverage(branch=True, include='app/*')
COV.start()
from flask_script import Manager
from flask_migrate import Migrate, MigrateCommand
from app import app, db
from config import BASE_DIR
app.config.from_object(os.getenv('BG_CONFIG') or 'config.DevelopmentConfig')
manager = Manager(app)
migrate = Migrate(app, db)
manager.add_command('db', MigrateCommand)
@manager.command
def test(coverage=False):
"""Run the unit tests."""
if coverage and not os.environ.get('FLASK_COVERAGE'):
import sys
os.environ['FLASK_COVERAGE'] = '1'
os.execvp(sys.executable, [sys.executable] + sys.argv)
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if COV:
COV.stop()
COV.save()
print('Coverage Summary:')
COV.report()
covdir = os.path.join(BASE_DIR, 'htmlcov')
COV.html_report(directory=covdir)
print('HTML version: file://%s/index.html' % covdir)
COV.erase()
if __name__ == '__main__':
manager.run()
| import os
COV = None
if os.environ.get('FLASK_COVERAGE'):
import coverage
COV = coverage.coverage(branch=True, include='app/*')
COV.start()
from flask_script import Manager
from flask_migrate import Migrate, MigrateCommand
from app import app, db
from config import BASE_DIR
app.config.from_object(os.getenv('BG_CONFIG') or 'config.DevelopmentConfig')
manager = Manager(app)
migrate = Migrate(app, db)
manager.add_command('db', MigrateCommand)
@manager.command
def test(coverage=False):
"""Run the unit tests."""
if coverage and not os.environ.get('FLASK_COVERAGE'):
import sys
os.environ['FLASK_COVERAGE'] = '1'
os.execvp(sys.executable, [sys.executable] + sys.argv)
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if COV:
COV.stop()
COV.save()
print('Coverage Summary:')
COV.report()
covdir = os.path.join(BASE_DIR, 'htmlcov')
COV.html_report(directory=covdir)
print('HTML version: file://%s/index.html' % covdir)
if __name__ == '__main__':
manager.run()
| Remove erase of coverage file as it's needed by coveralls | Remove erase of coverage file as it's needed by coveralls
| Python | apache-2.0 | atindale/business-glossary,atindale/business-glossary | import os
COV = None
if os.environ.get('FLASK_COVERAGE'):
import coverage
COV = coverage.coverage(branch=True, include='app/*')
COV.start()
from flask_script import Manager
from flask_migrate import Migrate, MigrateCommand
from app import app, db
from config import BASE_DIR
app.config.from_object(os.getenv('BG_CONFIG') or 'config.DevelopmentConfig')
manager = Manager(app)
migrate = Migrate(app, db)
manager.add_command('db', MigrateCommand)
@manager.command
def test(coverage=False):
"""Run the unit tests."""
if coverage and not os.environ.get('FLASK_COVERAGE'):
import sys
os.environ['FLASK_COVERAGE'] = '1'
os.execvp(sys.executable, [sys.executable] + sys.argv)
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if COV:
COV.stop()
COV.save()
print('Coverage Summary:')
COV.report()
covdir = os.path.join(BASE_DIR, 'htmlcov')
COV.html_report(directory=covdir)
print('HTML version: file://%s/index.html' % covdir)
- COV.erase()
if __name__ == '__main__':
manager.run()
| Remove erase of coverage file as it's needed by coveralls | ## Code Before:
import os
COV = None
if os.environ.get('FLASK_COVERAGE'):
import coverage
COV = coverage.coverage(branch=True, include='app/*')
COV.start()
from flask_script import Manager
from flask_migrate import Migrate, MigrateCommand
from app import app, db
from config import BASE_DIR
app.config.from_object(os.getenv('BG_CONFIG') or 'config.DevelopmentConfig')
manager = Manager(app)
migrate = Migrate(app, db)
manager.add_command('db', MigrateCommand)
@manager.command
def test(coverage=False):
"""Run the unit tests."""
if coverage and not os.environ.get('FLASK_COVERAGE'):
import sys
os.environ['FLASK_COVERAGE'] = '1'
os.execvp(sys.executable, [sys.executable] + sys.argv)
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if COV:
COV.stop()
COV.save()
print('Coverage Summary:')
COV.report()
covdir = os.path.join(BASE_DIR, 'htmlcov')
COV.html_report(directory=covdir)
print('HTML version: file://%s/index.html' % covdir)
COV.erase()
if __name__ == '__main__':
manager.run()
## Instruction:
Remove erase of coverage file as it's needed by coveralls
## Code After:
import os
COV = None
if os.environ.get('FLASK_COVERAGE'):
import coverage
COV = coverage.coverage(branch=True, include='app/*')
COV.start()
from flask_script import Manager
from flask_migrate import Migrate, MigrateCommand
from app import app, db
from config import BASE_DIR
app.config.from_object(os.getenv('BG_CONFIG') or 'config.DevelopmentConfig')
manager = Manager(app)
migrate = Migrate(app, db)
manager.add_command('db', MigrateCommand)
@manager.command
def test(coverage=False):
"""Run the unit tests."""
if coverage and not os.environ.get('FLASK_COVERAGE'):
import sys
os.environ['FLASK_COVERAGE'] = '1'
os.execvp(sys.executable, [sys.executable] + sys.argv)
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if COV:
COV.stop()
COV.save()
print('Coverage Summary:')
COV.report()
covdir = os.path.join(BASE_DIR, 'htmlcov')
COV.html_report(directory=covdir)
print('HTML version: file://%s/index.html' % covdir)
if __name__ == '__main__':
manager.run()
|
1be8c268f2da618e9e9e13d55d53599d637d3a6a | abel/tests/test_tools.py | abel/tests/test_tools.py | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import os.path
import numpy as np
from numpy.testing import assert_allclose
from abel.tools import calculate_speeds
DATA_DIR = os.path.join(os.path.split(__file__)[0], 'data')
def assert_equal(x, y, message, rtol=1e-5):
assert np.allclose(x, y, rtol=1e-5), message
def test_speeds():
# This very superficial test checks that calculate_speeds is able to
# execute (no syntax errors)
n = 101
IM = np.random.randn(n, n)
calculate_speeds(IM, n)
def test_centering_function():
pass
| from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import os.path
import numpy as np
from numpy.testing import assert_allclose, assert_equal
from abel.tools import calculate_speeds, center_image
DATA_DIR = os.path.join(os.path.split(__file__)[0], 'data')
def assert_allclose_msg(x, y, message, rtol=1e-5):
assert np.allclose(x, y, rtol=1e-5), message
def test_speeds():
# This very superficial test checks that calculate_speeds is able to
# execute (no syntax errors)
n = 101
IM = np.random.randn(n, n)
calculate_speeds(IM, n)
def test_centering_function_shape():
# ni -> original shape
# n -> result of the centering function
for (ni, n) in [(20, 10), # crop image
(20, 11),
(4, 10), # pad image
(4, 11)]:
data = np.zeros((ni, ni))
res = center_image(data, (ni//2, ni//2), n)
yield assert_equal, res.shape, (n, n),\
'Centering preserves shapes for ni={}, n={}'.format(ni, n)
| Test to enforce that center_image preserves shape | Test to enforce that center_image preserves shape
| Python | mit | rth/PyAbel,huletlab/PyAbel,DhrubajyotiDas/PyAbel,PyAbel/PyAbel,stggh/PyAbel | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import os.path
import numpy as np
- from numpy.testing import assert_allclose
+ from numpy.testing import assert_allclose, assert_equal
- from abel.tools import calculate_speeds
+ from abel.tools import calculate_speeds, center_image
DATA_DIR = os.path.join(os.path.split(__file__)[0], 'data')
- def assert_equal(x, y, message, rtol=1e-5):
+ def assert_allclose_msg(x, y, message, rtol=1e-5):
assert np.allclose(x, y, rtol=1e-5), message
+
def test_speeds():
# This very superficial test checks that calculate_speeds is able to
# execute (no syntax errors)
n = 101
IM = np.random.randn(n, n)
calculate_speeds(IM, n)
- def test_centering_function():
+ def test_centering_function_shape():
- pass
+ # ni -> original shape
+ # n -> result of the centering function
+ for (ni, n) in [(20, 10), # crop image
+ (20, 11),
+ (4, 10), # pad image
+ (4, 11)]:
+ data = np.zeros((ni, ni))
+ res = center_image(data, (ni//2, ni//2), n)
+ yield assert_equal, res.shape, (n, n),\
+ 'Centering preserves shapes for ni={}, n={}'.format(ni, n)
- | Test to enforce that center_image preserves shape | ## Code Before:
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import os.path
import numpy as np
from numpy.testing import assert_allclose
from abel.tools import calculate_speeds
DATA_DIR = os.path.join(os.path.split(__file__)[0], 'data')
def assert_equal(x, y, message, rtol=1e-5):
assert np.allclose(x, y, rtol=1e-5), message
def test_speeds():
# This very superficial test checks that calculate_speeds is able to
# execute (no syntax errors)
n = 101
IM = np.random.randn(n, n)
calculate_speeds(IM, n)
def test_centering_function():
pass
## Instruction:
Test to enforce that center_image preserves shape
## Code After:
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import os.path
import numpy as np
from numpy.testing import assert_allclose, assert_equal
from abel.tools import calculate_speeds, center_image
DATA_DIR = os.path.join(os.path.split(__file__)[0], 'data')
def assert_allclose_msg(x, y, message, rtol=1e-5):
assert np.allclose(x, y, rtol=1e-5), message
def test_speeds():
# This very superficial test checks that calculate_speeds is able to
# execute (no syntax errors)
n = 101
IM = np.random.randn(n, n)
calculate_speeds(IM, n)
def test_centering_function_shape():
# ni -> original shape
# n -> result of the centering function
for (ni, n) in [(20, 10), # crop image
(20, 11),
(4, 10), # pad image
(4, 11)]:
data = np.zeros((ni, ni))
res = center_image(data, (ni//2, ni//2), n)
yield assert_equal, res.shape, (n, n),\
'Centering preserves shapes for ni={}, n={}'.format(ni, n)
|
85a873c79e3c3c7289a23da6fe3673259fac9cbd | sympy/sets/conditionset.py | sympy/sets/conditionset.py | from __future__ import print_function, division
from sympy.core.basic import Basic
from sympy.logic.boolalg import And, Or, Not, true, false
from sympy.sets.sets import (Set, Interval, Intersection, EmptySet, Union,
FiniteSet)
from sympy.core.singleton import Singleton, S
from sympy.core.sympify import _sympify
from sympy.core.decorators import deprecated
from sympy.core.function import Lambda
class ConditionSet(Set):
"""
Set of elements which satisfies a given condition.
{x | condition(x) is True for x in S}
Examples
========
>>> from sympy import Symbol, S, CondSet, Lambda, pi, Eq, sin
>>> x = Symbol('x')
>>> sin_sols = CondSet(Lambda(x, Eq(sin(x), 0)), S.Reals)
>>> 2*pi in sin_sols
True
"""
def __new__(cls, lamda, base_set):
return Basic.__new__(cls, lamda, base_set)
condition = property(lambda self: self.args[0])
base_set = property(lambda self: self.args[1])
def _is_multivariate(self):
return len(self.lamda.variables) > 1
def contains(self, other):
return And(self.condition(other), self.base_set.contains(other))
| from __future__ import print_function, division
from sympy.core.basic import Basic
from sympy.logic.boolalg import And, Or, Not, true, false
from sympy.sets.sets import (Set, Interval, Intersection, EmptySet, Union,
FiniteSet)
from sympy.core.singleton import Singleton, S
from sympy.core.sympify import _sympify
from sympy.core.decorators import deprecated
from sympy.core.function import Lambda
class ConditionSet(Set):
"""
Set of elements which satisfies a given condition.
{x | condition(x) is True for x in S}
Examples
========
>>> from sympy import Symbol, S, CondSet, Lambda, pi, Eq, sin
>>> x = Symbol('x')
>>> sin_sols = CondSet(Lambda(x, Eq(sin(x), 0)), S.Reals)
>>> 2*pi in sin_sols
True
"""
def __new__(cls, lamda, base_set):
return Basic.__new__(cls, lamda, base_set)
condition = property(lambda self: self.args[0])
base_set = property(lambda self: self.args[1])
def contains(self, other):
return And(self.condition(other), self.base_set.contains(other))
| Remove redundant _is_multivariate in ConditionSet | Remove redundant _is_multivariate in ConditionSet
Signed-off-by: Harsh Gupta <[email protected]>
| Python | bsd-3-clause | sampadsaha5/sympy,VaibhavAgarwalVA/sympy,mcdaniel67/sympy,MechCoder/sympy,farhaanbukhsh/sympy,abhiii5459/sympy,MechCoder/sympy,madan96/sympy,kaushik94/sympy,jaimahajan1997/sympy,skidzo/sympy,ahhda/sympy,abhiii5459/sympy,mafiya69/sympy,Curious72/sympy,shikil/sympy,aktech/sympy,hargup/sympy,Vishluck/sympy,oliverlee/sympy,debugger22/sympy,kaichogami/sympy,Shaswat27/sympy,postvakje/sympy,ga7g08/sympy,skidzo/sympy,kumarkrishna/sympy,emon10005/sympy,AkademieOlympia/sympy,oliverlee/sympy,atreyv/sympy,ahhda/sympy,Shaswat27/sympy,souravsingh/sympy,moble/sympy,yukoba/sympy,farhaanbukhsh/sympy,ahhda/sympy,debugger22/sympy,drufat/sympy,ChristinaZografou/sympy,chaffra/sympy,Davidjohnwilson/sympy,jbbskinny/sympy,moble/sympy,kumarkrishna/sympy,jbbskinny/sympy,Curious72/sympy,hargup/sympy,moble/sympy,mafiya69/sympy,souravsingh/sympy,rahuldan/sympy,chaffra/sympy,farhaanbukhsh/sympy,aktech/sympy,wanglongqi/sympy,Arafatk/sympy,iamutkarshtiwari/sympy,drufat/sympy,Titan-C/sympy,kaichogami/sympy,kaichogami/sympy,AkademieOlympia/sympy,jerli/sympy,rahuldan/sympy,mcdaniel67/sympy,VaibhavAgarwalVA/sympy,Shaswat27/sympy,cswiercz/sympy,iamutkarshtiwari/sympy,aktech/sympy,VaibhavAgarwalVA/sympy,madan96/sympy,kaushik94/sympy,Vishluck/sympy,drufat/sympy,jerli/sympy,pandeyadarsh/sympy,saurabhjn76/sympy,abhiii5459/sympy,Vishluck/sympy,sahmed95/sympy,wanglongqi/sympy,saurabhjn76/sympy,chaffra/sympy,Designist/sympy,ChristinaZografou/sympy,madan96/sympy,Titan-C/sympy,skidzo/sympy,sahmed95/sympy,lindsayad/sympy,Arafatk/sympy,oliverlee/sympy,kevalds51/sympy,lindsayad/sympy,shikil/sympy,pandeyadarsh/sympy,yukoba/sympy,grevutiu-gabriel/sympy,debugger22/sympy,Davidjohnwilson/sympy,cswiercz/sympy,jbbskinny/sympy,kumarkrishna/sympy,maniteja123/sympy,jaimahajan1997/sympy,mafiya69/sympy,maniteja123/sympy,rahuldan/sympy,yashsharan/sympy,hargup/sympy,yukoba/sympy,postvakje/sympy,grevutiu-gabriel/sympy,Designist/sympy,wanglongqi/sympy,iamutkarshtiwari/sympy,souravsingh/sympy,ChristinaZografou/sympy,jerli/sympy,kevalds51/sympy,yashsharan/sympy,Titan-C/sympy,atreyv/sympy,cswiercz/sympy,kevalds51/sympy,postvakje/sympy,jaimahajan1997/sympy,Davidjohnwilson/sympy,maniteja123/sympy,sampadsaha5/sympy,kaushik94/sympy,yashsharan/sympy,mcdaniel67/sympy,lindsayad/sympy,ga7g08/sympy,sahmed95/sympy,Designist/sympy,sampadsaha5/sympy,saurabhjn76/sympy,shikil/sympy,AkademieOlympia/sympy,pandeyadarsh/sympy,ga7g08/sympy,atreyv/sympy,emon10005/sympy,MechCoder/sympy,grevutiu-gabriel/sympy,Arafatk/sympy,Curious72/sympy,emon10005/sympy | from __future__ import print_function, division
from sympy.core.basic import Basic
from sympy.logic.boolalg import And, Or, Not, true, false
from sympy.sets.sets import (Set, Interval, Intersection, EmptySet, Union,
FiniteSet)
from sympy.core.singleton import Singleton, S
from sympy.core.sympify import _sympify
from sympy.core.decorators import deprecated
from sympy.core.function import Lambda
class ConditionSet(Set):
"""
Set of elements which satisfies a given condition.
{x | condition(x) is True for x in S}
Examples
========
>>> from sympy import Symbol, S, CondSet, Lambda, pi, Eq, sin
>>> x = Symbol('x')
>>> sin_sols = CondSet(Lambda(x, Eq(sin(x), 0)), S.Reals)
>>> 2*pi in sin_sols
True
"""
def __new__(cls, lamda, base_set):
return Basic.__new__(cls, lamda, base_set)
condition = property(lambda self: self.args[0])
base_set = property(lambda self: self.args[1])
- def _is_multivariate(self):
- return len(self.lamda.variables) > 1
-
def contains(self, other):
return And(self.condition(other), self.base_set.contains(other))
| Remove redundant _is_multivariate in ConditionSet | ## Code Before:
from __future__ import print_function, division
from sympy.core.basic import Basic
from sympy.logic.boolalg import And, Or, Not, true, false
from sympy.sets.sets import (Set, Interval, Intersection, EmptySet, Union,
FiniteSet)
from sympy.core.singleton import Singleton, S
from sympy.core.sympify import _sympify
from sympy.core.decorators import deprecated
from sympy.core.function import Lambda
class ConditionSet(Set):
"""
Set of elements which satisfies a given condition.
{x | condition(x) is True for x in S}
Examples
========
>>> from sympy import Symbol, S, CondSet, Lambda, pi, Eq, sin
>>> x = Symbol('x')
>>> sin_sols = CondSet(Lambda(x, Eq(sin(x), 0)), S.Reals)
>>> 2*pi in sin_sols
True
"""
def __new__(cls, lamda, base_set):
return Basic.__new__(cls, lamda, base_set)
condition = property(lambda self: self.args[0])
base_set = property(lambda self: self.args[1])
def _is_multivariate(self):
return len(self.lamda.variables) > 1
def contains(self, other):
return And(self.condition(other), self.base_set.contains(other))
## Instruction:
Remove redundant _is_multivariate in ConditionSet
## Code After:
from __future__ import print_function, division
from sympy.core.basic import Basic
from sympy.logic.boolalg import And, Or, Not, true, false
from sympy.sets.sets import (Set, Interval, Intersection, EmptySet, Union,
FiniteSet)
from sympy.core.singleton import Singleton, S
from sympy.core.sympify import _sympify
from sympy.core.decorators import deprecated
from sympy.core.function import Lambda
class ConditionSet(Set):
"""
Set of elements which satisfies a given condition.
{x | condition(x) is True for x in S}
Examples
========
>>> from sympy import Symbol, S, CondSet, Lambda, pi, Eq, sin
>>> x = Symbol('x')
>>> sin_sols = CondSet(Lambda(x, Eq(sin(x), 0)), S.Reals)
>>> 2*pi in sin_sols
True
"""
def __new__(cls, lamda, base_set):
return Basic.__new__(cls, lamda, base_set)
condition = property(lambda self: self.args[0])
base_set = property(lambda self: self.args[1])
def contains(self, other):
return And(self.condition(other), self.base_set.contains(other))
|
14c86e3c93bd5114b74c125fdb8213b22342c95c | tests/manual_cleanup.py | tests/manual_cleanup.py | from globus_cli.services.transfer import get_client as get_tc
from tests.framework.cli_testcase import default_test_config
try:
from mock import patch
except ImportError:
from unittest.mock import patch
def cleanup_bookmarks(tc):
for bm in tc.bookmark_list():
tc.delete_bookmark(bm['id'])
@patch("globus_cli.config.get_config_obj", new=default_test_config)
def main():
tc = get_tc()
cleanup_bookmarks(tc)
if __name__ == '__main__':
main()
| import click
from globus_cli.services.transfer import get_client as get_tc
from tests.framework.cli_testcase import default_test_config
try:
from mock import patch
except ImportError:
from unittest.mock import patch
def cleanup_bookmarks(tc):
for bm in tc.bookmark_list():
tc.delete_bookmark(bm['id'])
def cleanup_tasks(tc):
tasks = tc.task_list(num_results=None, filter="status:ACTIVE,INACTIVE")
for t in tasks:
tc.cancel_task(t['task_id'])
@click.command("cleanup")
@click.option("--cancel-jobs", is_flag=True)
def main(cancel_jobs):
with patch("globus_cli.config.get_config_obj", new=default_test_config):
tc = get_tc()
cleanup_bookmarks(tc)
if cancel_jobs:
cleanup_tasks(tc)
if __name__ == '__main__':
main()
| Add `--cancel-jobs` to manual cleanup script | Add `--cancel-jobs` to manual cleanup script
Add an option to this script to cancel all ACTIVE,INACTIVE tasks (i.e.
not SUCCEDED,FAILED).
While this can disrupt a run of the tests pretty badly if you run it
while the tets are running, it's pretty much the only way to "fix it" if
the tests go off the rails because of a partial or failed job, etc.
| Python | apache-2.0 | globus/globus-cli,globus/globus-cli | + import click
+
from globus_cli.services.transfer import get_client as get_tc
from tests.framework.cli_testcase import default_test_config
try:
from mock import patch
except ImportError:
from unittest.mock import patch
def cleanup_bookmarks(tc):
for bm in tc.bookmark_list():
tc.delete_bookmark(bm['id'])
+ def cleanup_tasks(tc):
+ tasks = tc.task_list(num_results=None, filter="status:ACTIVE,INACTIVE")
+ for t in tasks:
+ tc.cancel_task(t['task_id'])
+
+
+ @click.command("cleanup")
+ @click.option("--cancel-jobs", is_flag=True)
+ def main(cancel_jobs):
- @patch("globus_cli.config.get_config_obj", new=default_test_config)
+ with patch("globus_cli.config.get_config_obj", new=default_test_config):
- def main():
- tc = get_tc()
+ tc = get_tc()
- cleanup_bookmarks(tc)
+ cleanup_bookmarks(tc)
+
+ if cancel_jobs:
+ cleanup_tasks(tc)
if __name__ == '__main__':
main()
| Add `--cancel-jobs` to manual cleanup script | ## Code Before:
from globus_cli.services.transfer import get_client as get_tc
from tests.framework.cli_testcase import default_test_config
try:
from mock import patch
except ImportError:
from unittest.mock import patch
def cleanup_bookmarks(tc):
for bm in tc.bookmark_list():
tc.delete_bookmark(bm['id'])
@patch("globus_cli.config.get_config_obj", new=default_test_config)
def main():
tc = get_tc()
cleanup_bookmarks(tc)
if __name__ == '__main__':
main()
## Instruction:
Add `--cancel-jobs` to manual cleanup script
## Code After:
import click
from globus_cli.services.transfer import get_client as get_tc
from tests.framework.cli_testcase import default_test_config
try:
from mock import patch
except ImportError:
from unittest.mock import patch
def cleanup_bookmarks(tc):
for bm in tc.bookmark_list():
tc.delete_bookmark(bm['id'])
def cleanup_tasks(tc):
tasks = tc.task_list(num_results=None, filter="status:ACTIVE,INACTIVE")
for t in tasks:
tc.cancel_task(t['task_id'])
@click.command("cleanup")
@click.option("--cancel-jobs", is_flag=True)
def main(cancel_jobs):
with patch("globus_cli.config.get_config_obj", new=default_test_config):
tc = get_tc()
cleanup_bookmarks(tc)
if cancel_jobs:
cleanup_tasks(tc)
if __name__ == '__main__':
main()
|
a3d087b2d7ec42eb5afe9f0064785d25500af11b | bitforge/errors.py | bitforge/errors.py |
class BitforgeError(Exception):
def __init__(self, *args, **kwargs):
self.cause = kwargs.pop('cause', None)
self.prepare(*args, **kwargs)
self.message = self.__doc__.format(**self.__dict__)
def prepare(self):
pass
def __str__(self):
return self.message
class ObjectError(BitforgeError):
def prepare(self, object):
self.object = object
class StringError(BitforgeError):
def prepare(self, string):
self.string = repr(string)
self.length = len(string)
class NumberError(BitforgeError):
def prepare(self, number):
self.number = number
class KeyValueError(BitforgeError):
def prepare(self, key, value):
self.key = key
self.value = value
|
class BitforgeError(Exception):
def __init__(self, *args, **kwargs):
self.cause = kwargs.pop('cause', None)
self.prepare(*args, **kwargs)
message = self.__doc__.format(**self.__dict__)
super(BitforgeError, self).__init__(message)
def prepare(self):
pass
def __str__(self):
return self.message
class ObjectError(BitforgeError):
def prepare(self, object):
self.object = object
class StringError(BitforgeError):
def prepare(self, string):
self.string = repr(string)
self.length = len(string)
class NumberError(BitforgeError):
def prepare(self, number):
self.number = number
class KeyValueError(BitforgeError):
def prepare(self, key, value):
self.key = key
self.value = value
| Call super constructor for BitforgeError | Call super constructor for BitforgeError
| Python | mit | muun/bitforge,coinforge/bitforge |
class BitforgeError(Exception):
def __init__(self, *args, **kwargs):
self.cause = kwargs.pop('cause', None)
self.prepare(*args, **kwargs)
- self.message = self.__doc__.format(**self.__dict__)
+ message = self.__doc__.format(**self.__dict__)
+ super(BitforgeError, self).__init__(message)
def prepare(self):
pass
def __str__(self):
return self.message
class ObjectError(BitforgeError):
def prepare(self, object):
self.object = object
class StringError(BitforgeError):
def prepare(self, string):
self.string = repr(string)
self.length = len(string)
class NumberError(BitforgeError):
def prepare(self, number):
self.number = number
class KeyValueError(BitforgeError):
def prepare(self, key, value):
self.key = key
self.value = value
| Call super constructor for BitforgeError | ## Code Before:
class BitforgeError(Exception):
def __init__(self, *args, **kwargs):
self.cause = kwargs.pop('cause', None)
self.prepare(*args, **kwargs)
self.message = self.__doc__.format(**self.__dict__)
def prepare(self):
pass
def __str__(self):
return self.message
class ObjectError(BitforgeError):
def prepare(self, object):
self.object = object
class StringError(BitforgeError):
def prepare(self, string):
self.string = repr(string)
self.length = len(string)
class NumberError(BitforgeError):
def prepare(self, number):
self.number = number
class KeyValueError(BitforgeError):
def prepare(self, key, value):
self.key = key
self.value = value
## Instruction:
Call super constructor for BitforgeError
## Code After:
class BitforgeError(Exception):
def __init__(self, *args, **kwargs):
self.cause = kwargs.pop('cause', None)
self.prepare(*args, **kwargs)
message = self.__doc__.format(**self.__dict__)
super(BitforgeError, self).__init__(message)
def prepare(self):
pass
def __str__(self):
return self.message
class ObjectError(BitforgeError):
def prepare(self, object):
self.object = object
class StringError(BitforgeError):
def prepare(self, string):
self.string = repr(string)
self.length = len(string)
class NumberError(BitforgeError):
def prepare(self, number):
self.number = number
class KeyValueError(BitforgeError):
def prepare(self, key, value):
self.key = key
self.value = value
|
c19a64ecdbde5a387a84dec880c2ebea1013c3d6 | canopus/auth/token_factory.py | canopus/auth/token_factory.py | from datetime import datetime, timedelta
from ..schema import UserSchema
class TokenFactory(object):
def __init__(self, request, user):
self.user = user
self.request = request
def create_access_token(self):
user = self.user
if user.last_signed_in is None:
user.welcome()
user.last_signed_in = datetime.now()
token = self.request.create_jwt_token(user.id, expiration=timedelta(days=7))
user_schema = UserSchema(exclude=('enabled',))
return dict(token=token, user={})
| from datetime import datetime, timedelta
from ..schema import UserSchema
class TokenFactory(object):
def __init__(self, request, user):
self.user = user
self.request = request
def create_access_token(self):
user = self.user
user.last_signed_in = datetime.now()
token = self.request.create_jwt_token(user.id, expiration=timedelta(days=7))
user_schema = UserSchema(exclude=('enabled',))
return dict(token=token, user=user_schema.dump(user).data)
| Include user in create_access_token return value | Include user in create_access_token return value
| Python | mit | josuemontano/api-starter,josuemontano/API-platform,josuemontano/pyramid-angularjs-starter,josuemontano/API-platform,josuemontano/API-platform,josuemontano/api-starter,josuemontano/api-starter,josuemontano/pyramid-angularjs-starter,josuemontano/pyramid-angularjs-starter,josuemontano/API-platform | from datetime import datetime, timedelta
from ..schema import UserSchema
class TokenFactory(object):
def __init__(self, request, user):
self.user = user
self.request = request
def create_access_token(self):
user = self.user
- if user.last_signed_in is None:
- user.welcome()
user.last_signed_in = datetime.now()
token = self.request.create_jwt_token(user.id, expiration=timedelta(days=7))
user_schema = UserSchema(exclude=('enabled',))
- return dict(token=token, user={})
+ return dict(token=token, user=user_schema.dump(user).data)
| Include user in create_access_token return value | ## Code Before:
from datetime import datetime, timedelta
from ..schema import UserSchema
class TokenFactory(object):
def __init__(self, request, user):
self.user = user
self.request = request
def create_access_token(self):
user = self.user
if user.last_signed_in is None:
user.welcome()
user.last_signed_in = datetime.now()
token = self.request.create_jwt_token(user.id, expiration=timedelta(days=7))
user_schema = UserSchema(exclude=('enabled',))
return dict(token=token, user={})
## Instruction:
Include user in create_access_token return value
## Code After:
from datetime import datetime, timedelta
from ..schema import UserSchema
class TokenFactory(object):
def __init__(self, request, user):
self.user = user
self.request = request
def create_access_token(self):
user = self.user
user.last_signed_in = datetime.now()
token = self.request.create_jwt_token(user.id, expiration=timedelta(days=7))
user_schema = UserSchema(exclude=('enabled',))
return dict(token=token, user=user_schema.dump(user).data)
|
72f03f05adc39a3d920ee8372f4108b5c8c8671c | modules/pwnchecker.py | modules/pwnchecker.py |
import requests
import pprint
import argparse
parser = argparse.ArgumentParser(description='A tool to check if your email account has been in a breach By Jay Townsend')
parser.add_argument('-e', '--email-account', help='Email account to lookup', required=True)
args = parser.parse_args()
headers = {'User-agent': 'Have I been pwn module'}
API = 'https://haveibeenpwned.com/api/v2/breachedaccount/{0}'.format(args.email_account)
request = requests.get(API, headers=headers)
if request.status_code == 404:
print('Cannot find your account')
else:
entries = request.json()
for entry in entries:
print('Domain:', entry['Domain'])
print('DateAdded:', entry['AddedDate'])
print('BreachDate:', entry['BreachDate'])
pprint.pprint(entry['Description'])
print('IsSensitive:', entry['IsSensitive'])
print('IsVerified:', entry['IsVerified'])
print('PwnCount:', entry['PwnCount'])
|
import requests
import pprint
import argparse
parser = argparse.ArgumentParser(description='A tool to check if your email account has been in a breach By Jay Townsend')
parser.add_argument('-e', '--email-account', help='Email account to lookup', required=True)
parser.add_argument('-k', '--api-key', help='Your HIBP API key', required=True)
args = parser.parse_args()
headers = {'User-agent': 'Have I been pwn module',
f'hibp-api-key': {args.api_key}}
API = f'https://haveibeenpwned.com/api/v3/breachedaccount/{args.email_account}'
request = requests.get(API, headers=headers)
if request.status_code == 404:
print('Cannot find your account')
else:
entries = request.json()
for entry in entries:
print('Domain:', entry['Domain'])
print('DateAdded:', entry['AddedDate'])
print('BreachDate:', entry['BreachDate'])
pprint.pprint(entry['Description'])
print('IsSensitive:', entry['IsSensitive'])
print('IsVerified:', entry['IsVerified'])
print('PwnCount:', entry['PwnCount'])
| Update to the v3 api | Update to the v3 api
| Python | bsd-3-clause | L1ghtn1ng/usaf,L1ghtn1ng/usaf |
import requests
import pprint
import argparse
parser = argparse.ArgumentParser(description='A tool to check if your email account has been in a breach By Jay Townsend')
parser.add_argument('-e', '--email-account', help='Email account to lookup', required=True)
+ parser.add_argument('-k', '--api-key', help='Your HIBP API key', required=True)
args = parser.parse_args()
- headers = {'User-agent': 'Have I been pwn module'}
+ headers = {'User-agent': 'Have I been pwn module',
+ f'hibp-api-key': {args.api_key}}
- API = 'https://haveibeenpwned.com/api/v2/breachedaccount/{0}'.format(args.email_account)
+ API = f'https://haveibeenpwned.com/api/v3/breachedaccount/{args.email_account}'
request = requests.get(API, headers=headers)
if request.status_code == 404:
print('Cannot find your account')
else:
entries = request.json()
for entry in entries:
print('Domain:', entry['Domain'])
print('DateAdded:', entry['AddedDate'])
print('BreachDate:', entry['BreachDate'])
pprint.pprint(entry['Description'])
print('IsSensitive:', entry['IsSensitive'])
print('IsVerified:', entry['IsVerified'])
print('PwnCount:', entry['PwnCount'])
| Update to the v3 api | ## Code Before:
import requests
import pprint
import argparse
parser = argparse.ArgumentParser(description='A tool to check if your email account has been in a breach By Jay Townsend')
parser.add_argument('-e', '--email-account', help='Email account to lookup', required=True)
args = parser.parse_args()
headers = {'User-agent': 'Have I been pwn module'}
API = 'https://haveibeenpwned.com/api/v2/breachedaccount/{0}'.format(args.email_account)
request = requests.get(API, headers=headers)
if request.status_code == 404:
print('Cannot find your account')
else:
entries = request.json()
for entry in entries:
print('Domain:', entry['Domain'])
print('DateAdded:', entry['AddedDate'])
print('BreachDate:', entry['BreachDate'])
pprint.pprint(entry['Description'])
print('IsSensitive:', entry['IsSensitive'])
print('IsVerified:', entry['IsVerified'])
print('PwnCount:', entry['PwnCount'])
## Instruction:
Update to the v3 api
## Code After:
import requests
import pprint
import argparse
parser = argparse.ArgumentParser(description='A tool to check if your email account has been in a breach By Jay Townsend')
parser.add_argument('-e', '--email-account', help='Email account to lookup', required=True)
parser.add_argument('-k', '--api-key', help='Your HIBP API key', required=True)
args = parser.parse_args()
headers = {'User-agent': 'Have I been pwn module',
f'hibp-api-key': {args.api_key}}
API = f'https://haveibeenpwned.com/api/v3/breachedaccount/{args.email_account}'
request = requests.get(API, headers=headers)
if request.status_code == 404:
print('Cannot find your account')
else:
entries = request.json()
for entry in entries:
print('Domain:', entry['Domain'])
print('DateAdded:', entry['AddedDate'])
print('BreachDate:', entry['BreachDate'])
pprint.pprint(entry['Description'])
print('IsSensitive:', entry['IsSensitive'])
print('IsVerified:', entry['IsVerified'])
print('PwnCount:', entry['PwnCount'])
|
31a9b285a0445c895aeff02b2abbeda12bf7f3d7 | wagtail/admin/tests/pages/test_content_type_use_view.py | wagtail/admin/tests/pages/test_content_type_use_view.py | from django.test import TestCase
from django.urls import reverse
from wagtail.tests.utils import WagtailTestUtils
class TestContentTypeUse(TestCase, WagtailTestUtils):
fixtures = ['test.json']
def setUp(self):
self.user = self.login()
def test_content_type_use(self):
# Get use of event page
response = self.client.get(reverse('wagtailadmin_pages:type_use', args=('tests', 'eventpage')))
# Check response
self.assertEqual(response.status_code, 200)
self.assertTemplateUsed(response, 'wagtailadmin/pages/content_type_use.html')
self.assertContains(response, "Christmas")
| from django.test import TestCase
from django.urls import reverse
from django.utils.http import urlencode
from wagtail.tests.testapp.models import EventPage
from wagtail.tests.utils import WagtailTestUtils
class TestContentTypeUse(TestCase, WagtailTestUtils):
fixtures = ['test.json']
def setUp(self):
self.user = self.login()
self.christmas_page = EventPage.objects.get(title="Christmas")
def test_content_type_use(self):
# Get use of event page
request_url = reverse('wagtailadmin_pages:type_use', args=('tests', 'eventpage'))
response = self.client.get(request_url)
# Check response
self.assertEqual(response.status_code, 200)
self.assertTemplateUsed(response, 'wagtailadmin/pages/content_type_use.html')
self.assertContains(response, "Christmas")
# Links to 'delete' etc should include a 'next' URL parameter pointing back here
delete_url = (
reverse('wagtailadmin_pages:delete', args=(self.christmas_page.id,))
+ '?' + urlencode({'next': request_url})
)
self.assertContains(response, delete_url)
| Add test for button URLs including a 'next' parameter | Add test for button URLs including a 'next' parameter
| Python | bsd-3-clause | torchbox/wagtail,FlipperPA/wagtail,gasman/wagtail,gasman/wagtail,mixxorz/wagtail,thenewguy/wagtail,mixxorz/wagtail,torchbox/wagtail,torchbox/wagtail,thenewguy/wagtail,gasman/wagtail,rsalmaso/wagtail,mixxorz/wagtail,wagtail/wagtail,takeflight/wagtail,FlipperPA/wagtail,torchbox/wagtail,zerolab/wagtail,thenewguy/wagtail,takeflight/wagtail,zerolab/wagtail,kaedroho/wagtail,wagtail/wagtail,zerolab/wagtail,thenewguy/wagtail,gasman/wagtail,rsalmaso/wagtail,kaedroho/wagtail,kaedroho/wagtail,thenewguy/wagtail,mixxorz/wagtail,zerolab/wagtail,jnns/wagtail,takeflight/wagtail,rsalmaso/wagtail,gasman/wagtail,FlipperPA/wagtail,wagtail/wagtail,wagtail/wagtail,zerolab/wagtail,rsalmaso/wagtail,kaedroho/wagtail,FlipperPA/wagtail,jnns/wagtail,jnns/wagtail,takeflight/wagtail,kaedroho/wagtail,mixxorz/wagtail,wagtail/wagtail,rsalmaso/wagtail,jnns/wagtail | from django.test import TestCase
from django.urls import reverse
+ from django.utils.http import urlencode
+ from wagtail.tests.testapp.models import EventPage
from wagtail.tests.utils import WagtailTestUtils
class TestContentTypeUse(TestCase, WagtailTestUtils):
fixtures = ['test.json']
def setUp(self):
self.user = self.login()
+ self.christmas_page = EventPage.objects.get(title="Christmas")
def test_content_type_use(self):
# Get use of event page
- response = self.client.get(reverse('wagtailadmin_pages:type_use', args=('tests', 'eventpage')))
+ request_url = reverse('wagtailadmin_pages:type_use', args=('tests', 'eventpage'))
+ response = self.client.get(request_url)
# Check response
self.assertEqual(response.status_code, 200)
self.assertTemplateUsed(response, 'wagtailadmin/pages/content_type_use.html')
self.assertContains(response, "Christmas")
+ # Links to 'delete' etc should include a 'next' URL parameter pointing back here
+ delete_url = (
+ reverse('wagtailadmin_pages:delete', args=(self.christmas_page.id,))
+ + '?' + urlencode({'next': request_url})
+ )
+ self.assertContains(response, delete_url)
+ | Add test for button URLs including a 'next' parameter | ## Code Before:
from django.test import TestCase
from django.urls import reverse
from wagtail.tests.utils import WagtailTestUtils
class TestContentTypeUse(TestCase, WagtailTestUtils):
fixtures = ['test.json']
def setUp(self):
self.user = self.login()
def test_content_type_use(self):
# Get use of event page
response = self.client.get(reverse('wagtailadmin_pages:type_use', args=('tests', 'eventpage')))
# Check response
self.assertEqual(response.status_code, 200)
self.assertTemplateUsed(response, 'wagtailadmin/pages/content_type_use.html')
self.assertContains(response, "Christmas")
## Instruction:
Add test for button URLs including a 'next' parameter
## Code After:
from django.test import TestCase
from django.urls import reverse
from django.utils.http import urlencode
from wagtail.tests.testapp.models import EventPage
from wagtail.tests.utils import WagtailTestUtils
class TestContentTypeUse(TestCase, WagtailTestUtils):
fixtures = ['test.json']
def setUp(self):
self.user = self.login()
self.christmas_page = EventPage.objects.get(title="Christmas")
def test_content_type_use(self):
# Get use of event page
request_url = reverse('wagtailadmin_pages:type_use', args=('tests', 'eventpage'))
response = self.client.get(request_url)
# Check response
self.assertEqual(response.status_code, 200)
self.assertTemplateUsed(response, 'wagtailadmin/pages/content_type_use.html')
self.assertContains(response, "Christmas")
# Links to 'delete' etc should include a 'next' URL parameter pointing back here
delete_url = (
reverse('wagtailadmin_pages:delete', args=(self.christmas_page.id,))
+ '?' + urlencode({'next': request_url})
)
self.assertContains(response, delete_url)
|
11fd3e5c8a2c7f591dff1ba1949508d178d1e5d5 | byceps/util/irc.py | byceps/util/irc.py |
from time import sleep
from typing import List
from flask import current_app
import requests
DEFAULT_BOT_URL = 'http://127.0.0.1:12345/'
DEFAULT_ENABLED = False
DELAY_IN_SECONDS = 2
DEFAULT_TEXT_PREFIX = '[BYCEPS] '
def send_message(channels: List[str], text: str) -> None:
"""Write the text to the channels by sending it to the bot via HTTP."""
enabled = current_app.config.get('ANNOUNCE_IRC_ENABLED', DEFAULT_ENABLED)
if not enabled:
current_app.logger.warning('Announcements on IRC are disabled.')
return
text_prefix = current_app.config.get(
'ANNOUNCE_IRC_TEXT_PREFIX', DEFAULT_TEXT_PREFIX
)
text = text_prefix + text
url = current_app.config.get('IRC_BOT_URL', DEFAULT_BOT_URL)
data = {'channels': channels, 'text': text}
# Delay a bit as an attempt to avoid getting kicked from server
# because of flooding.
sleep(DELAY_IN_SECONDS)
requests.post(url, json=data) # Ignore response code for now.
|
from time import sleep
from typing import List
from flask import current_app
import requests
DEFAULT_BOT_URL = 'http://127.0.0.1:12345/'
DEFAULT_ENABLED = False
DEFAULT_DELAY_IN_SECONDS = 2
DEFAULT_TEXT_PREFIX = '[BYCEPS] '
def send_message(channels: List[str], text: str) -> None:
"""Write the text to the channels by sending it to the bot via HTTP."""
enabled = current_app.config.get('ANNOUNCE_IRC_ENABLED', DEFAULT_ENABLED)
if not enabled:
current_app.logger.warning('Announcements on IRC are disabled.')
return
text_prefix = current_app.config.get(
'ANNOUNCE_IRC_TEXT_PREFIX', DEFAULT_TEXT_PREFIX
)
text = text_prefix + text
url = current_app.config.get('IRC_BOT_URL', DEFAULT_BOT_URL)
data = {'channels': channels, 'text': text}
# Delay a bit as an attempt to avoid getting kicked from server
# because of flooding.
delay = int(
current_app.config.get('ANNOUNCE_IRC_DELAY', DEFAULT_DELAY_IN_SECONDS)
)
sleep(delay)
requests.post(url, json=data) # Ignore response code for now.
| Make IRC message delay configurable | Make IRC message delay configurable
| Python | bsd-3-clause | homeworkprod/byceps,homeworkprod/byceps,homeworkprod/byceps |
from time import sleep
from typing import List
from flask import current_app
import requests
DEFAULT_BOT_URL = 'http://127.0.0.1:12345/'
DEFAULT_ENABLED = False
-
- DELAY_IN_SECONDS = 2
+ DEFAULT_DELAY_IN_SECONDS = 2
-
DEFAULT_TEXT_PREFIX = '[BYCEPS] '
def send_message(channels: List[str], text: str) -> None:
"""Write the text to the channels by sending it to the bot via HTTP."""
enabled = current_app.config.get('ANNOUNCE_IRC_ENABLED', DEFAULT_ENABLED)
if not enabled:
current_app.logger.warning('Announcements on IRC are disabled.')
return
text_prefix = current_app.config.get(
'ANNOUNCE_IRC_TEXT_PREFIX', DEFAULT_TEXT_PREFIX
)
text = text_prefix + text
url = current_app.config.get('IRC_BOT_URL', DEFAULT_BOT_URL)
data = {'channels': channels, 'text': text}
# Delay a bit as an attempt to avoid getting kicked from server
# because of flooding.
- sleep(DELAY_IN_SECONDS)
+ delay = int(
+ current_app.config.get('ANNOUNCE_IRC_DELAY', DEFAULT_DELAY_IN_SECONDS)
+ )
+ sleep(delay)
requests.post(url, json=data) # Ignore response code for now.
| Make IRC message delay configurable | ## Code Before:
from time import sleep
from typing import List
from flask import current_app
import requests
DEFAULT_BOT_URL = 'http://127.0.0.1:12345/'
DEFAULT_ENABLED = False
DELAY_IN_SECONDS = 2
DEFAULT_TEXT_PREFIX = '[BYCEPS] '
def send_message(channels: List[str], text: str) -> None:
"""Write the text to the channels by sending it to the bot via HTTP."""
enabled = current_app.config.get('ANNOUNCE_IRC_ENABLED', DEFAULT_ENABLED)
if not enabled:
current_app.logger.warning('Announcements on IRC are disabled.')
return
text_prefix = current_app.config.get(
'ANNOUNCE_IRC_TEXT_PREFIX', DEFAULT_TEXT_PREFIX
)
text = text_prefix + text
url = current_app.config.get('IRC_BOT_URL', DEFAULT_BOT_URL)
data = {'channels': channels, 'text': text}
# Delay a bit as an attempt to avoid getting kicked from server
# because of flooding.
sleep(DELAY_IN_SECONDS)
requests.post(url, json=data) # Ignore response code for now.
## Instruction:
Make IRC message delay configurable
## Code After:
from time import sleep
from typing import List
from flask import current_app
import requests
DEFAULT_BOT_URL = 'http://127.0.0.1:12345/'
DEFAULT_ENABLED = False
DEFAULT_DELAY_IN_SECONDS = 2
DEFAULT_TEXT_PREFIX = '[BYCEPS] '
def send_message(channels: List[str], text: str) -> None:
"""Write the text to the channels by sending it to the bot via HTTP."""
enabled = current_app.config.get('ANNOUNCE_IRC_ENABLED', DEFAULT_ENABLED)
if not enabled:
current_app.logger.warning('Announcements on IRC are disabled.')
return
text_prefix = current_app.config.get(
'ANNOUNCE_IRC_TEXT_PREFIX', DEFAULT_TEXT_PREFIX
)
text = text_prefix + text
url = current_app.config.get('IRC_BOT_URL', DEFAULT_BOT_URL)
data = {'channels': channels, 'text': text}
# Delay a bit as an attempt to avoid getting kicked from server
# because of flooding.
delay = int(
current_app.config.get('ANNOUNCE_IRC_DELAY', DEFAULT_DELAY_IN_SECONDS)
)
sleep(delay)
requests.post(url, json=data) # Ignore response code for now.
|
0c42909e5649b78260d9efa4e6ff7b77c82b1934 | runtests.py | runtests.py | import sys
from os.path import abspath
from os.path import dirname
# Load Django-related settings; necessary for tests to run and for Django
# imports to work.
import local_settings
from django.test.simple import DjangoTestSuiteRunner
def runtests():
parent_dir = dirname(abspath(__file__))
sys.path.insert(0, parent_dir)
test_runner = DjangoTestSuiteRunner(
verbosity=1,
interactive=False,
failfast=False)
failures = test_runner.run_tests(['djoauth2'])
sys.exit(failures)
if __name__ == '__main__':
runtests()
| import sys
from argparse import ArgumentParser
from os.path import abspath
from os.path import dirname
# Load Django-related settings; necessary for tests to run and for Django
# imports to work.
import local_settings
# Now, imports from Django will work properly without raising errors related to
# missing or badly-configured settings.
from django.test.simple import DjangoTestSuiteRunner
def runtests(verbosity, failfast, interactive, test_labels):
# Modify the path so that our djoauth2 app is in it.
parent_dir = dirname(abspath(__file__))
sys.path.insert(0, parent_dir)
test_runner = DjangoTestSuiteRunner(
verbosity=verbosity,
interactive=interactive,
failfast=failfast)
sys.exit(test_runner.run_tests(test_labels))
if __name__ == '__main__':
# Parse any command line arguments.
parser = ArgumentParser()
parser.add_argument('--failfast',
action='store_true',
default=False,
dest='failfast')
parser.add_argument('--interactive',
action='store_true',
default=False,
dest='interactive')
parser.add_argument('--verbosity', default=1, type=int)
parser.add_argument('test_labels', nargs='*', default=('djoauth2',))
args = parser.parse_args()
# Run the tests.
runtests(args.verbosity, args.failfast, args.interactive, args.test_labels)
| Allow testing of specific apps. | Allow testing of specific apps.
| Python | mit | seler/djoauth2,seler/djoauth2,vden/djoauth2-ng,Locu/djoauth2,vden/djoauth2-ng,Locu/djoauth2 | import sys
+ from argparse import ArgumentParser
from os.path import abspath
from os.path import dirname
# Load Django-related settings; necessary for tests to run and for Django
# imports to work.
import local_settings
-
+ # Now, imports from Django will work properly without raising errors related to
+ # missing or badly-configured settings.
from django.test.simple import DjangoTestSuiteRunner
- def runtests():
+ def runtests(verbosity, failfast, interactive, test_labels):
+ # Modify the path so that our djoauth2 app is in it.
- parent_dir = dirname(abspath(__file__))
+ parent_dir = dirname(abspath(__file__))
- sys.path.insert(0, parent_dir)
+ sys.path.insert(0, parent_dir)
- test_runner = DjangoTestSuiteRunner(
+ test_runner = DjangoTestSuiteRunner(
- verbosity=1,
+ verbosity=verbosity,
- interactive=False,
+ interactive=interactive,
- failfast=False)
+ failfast=failfast)
- failures = test_runner.run_tests(['djoauth2'])
- sys.exit(failures)
+
+ sys.exit(test_runner.run_tests(test_labels))
if __name__ == '__main__':
- runtests()
+ # Parse any command line arguments.
+ parser = ArgumentParser()
+ parser.add_argument('--failfast',
+ action='store_true',
+ default=False,
+ dest='failfast')
+ parser.add_argument('--interactive',
+ action='store_true',
+ default=False,
+ dest='interactive')
+ parser.add_argument('--verbosity', default=1, type=int)
+ parser.add_argument('test_labels', nargs='*', default=('djoauth2',))
+ args = parser.parse_args()
+
+ # Run the tests.
+ runtests(args.verbosity, args.failfast, args.interactive, args.test_labels)
+ | Allow testing of specific apps. | ## Code Before:
import sys
from os.path import abspath
from os.path import dirname
# Load Django-related settings; necessary for tests to run and for Django
# imports to work.
import local_settings
from django.test.simple import DjangoTestSuiteRunner
def runtests():
parent_dir = dirname(abspath(__file__))
sys.path.insert(0, parent_dir)
test_runner = DjangoTestSuiteRunner(
verbosity=1,
interactive=False,
failfast=False)
failures = test_runner.run_tests(['djoauth2'])
sys.exit(failures)
if __name__ == '__main__':
runtests()
## Instruction:
Allow testing of specific apps.
## Code After:
import sys
from argparse import ArgumentParser
from os.path import abspath
from os.path import dirname
# Load Django-related settings; necessary for tests to run and for Django
# imports to work.
import local_settings
# Now, imports from Django will work properly without raising errors related to
# missing or badly-configured settings.
from django.test.simple import DjangoTestSuiteRunner
def runtests(verbosity, failfast, interactive, test_labels):
# Modify the path so that our djoauth2 app is in it.
parent_dir = dirname(abspath(__file__))
sys.path.insert(0, parent_dir)
test_runner = DjangoTestSuiteRunner(
verbosity=verbosity,
interactive=interactive,
failfast=failfast)
sys.exit(test_runner.run_tests(test_labels))
if __name__ == '__main__':
# Parse any command line arguments.
parser = ArgumentParser()
parser.add_argument('--failfast',
action='store_true',
default=False,
dest='failfast')
parser.add_argument('--interactive',
action='store_true',
default=False,
dest='interactive')
parser.add_argument('--verbosity', default=1, type=int)
parser.add_argument('test_labels', nargs='*', default=('djoauth2',))
args = parser.parse_args()
# Run the tests.
runtests(args.verbosity, args.failfast, args.interactive, args.test_labels)
|
d9844f5bcf6d48bde1a60d32998ccdaa87e99676 | cloud_browser/__init__.py | cloud_browser/__init__.py |
VERSION = (0, 2, 1)
__version__ = ".".join(str(v) for v in VERSION)
__version_full__ = __version__ + "".join(str(v) for v in VERSION)
|
VERSION = (0, 2, 1)
__version__ = ".".join(str(v) for v in VERSION)
__version_full__ = __version__
| Fix __version_full__ for new scheme. | Version: Fix __version_full__ for new scheme.
| Python | mit | ryan-roemer/django-cloud-browser,UrbanDaddy/django-cloud-browser,UrbanDaddy/django-cloud-browser,ryan-roemer/django-cloud-browser,ryan-roemer/django-cloud-browser |
VERSION = (0, 2, 1)
__version__ = ".".join(str(v) for v in VERSION)
- __version_full__ = __version__ + "".join(str(v) for v in VERSION)
+ __version_full__ = __version__
| Fix __version_full__ for new scheme. | ## Code Before:
VERSION = (0, 2, 1)
__version__ = ".".join(str(v) for v in VERSION)
__version_full__ = __version__ + "".join(str(v) for v in VERSION)
## Instruction:
Fix __version_full__ for new scheme.
## Code After:
VERSION = (0, 2, 1)
__version__ = ".".join(str(v) for v in VERSION)
__version_full__ = __version__
|
e9b3865e37c1f4a275c323dcbf778696a27d69bd | testapp/testapp/management.py | testapp/testapp/management.py | from bitcategory import models
from django.dispatch import receiver
from django.db.models.signals import post_syncdb
@receiver(post_syncdb, sender=models)
def load_test_categories(sender, **kwargs):
r1, c = models.Category.objects.get_or_create(name="root1")
r2, c = models.Category.objects.get_or_create(name="root2")
c11, c = models.Category.objects.get_or_create(name="cat11", parent=r1)
c12, c = models.Category.objects.get_or_create(name="cat12", parent=r1)
c111, c = models.Category.objects.get_or_create(name="cat111", parent=c11)
c112, c = models.Category.objects.get_or_create(name="cat112", parent=c11)
c113, c = models.Category.objects.get_or_create(name="cat113", parent=c11)
c21, c = models.Category.objects.get_or_create(name="cat21", parent=r2)
c22, c = models.Category.objects.get_or_create(name="cat22", parent=r2)
| from bitcategory import models
from django.dispatch import receiver
from django.db.models.signals import post_migrate
@receiver(post_migrate, sender=models)
def load_test_categories(sender, **kwargs):
r1, c = models.Category.objects.get_or_create(name="root1")
r2, c = models.Category.objects.get_or_create(name="root2")
c11, c = models.Category.objects.get_or_create(name="cat11", parent=r1)
c12, c = models.Category.objects.get_or_create(name="cat12", parent=r1)
c111, c = models.Category.objects.get_or_create(name="cat111", parent=c11)
c112, c = models.Category.objects.get_or_create(name="cat112", parent=c11)
c113, c = models.Category.objects.get_or_create(name="cat113", parent=c11)
c21, c = models.Category.objects.get_or_create(name="cat21", parent=r2)
c22, c = models.Category.objects.get_or_create(name="cat22", parent=r2)
| Update tests to work under django 1.8+ | Update tests to work under django 1.8+
| Python | bsd-3-clause | atheiste/django-bit-category,atheiste/django-bit-category,atheiste/django-bit-category | from bitcategory import models
from django.dispatch import receiver
- from django.db.models.signals import post_syncdb
+ from django.db.models.signals import post_migrate
- @receiver(post_syncdb, sender=models)
+ @receiver(post_migrate, sender=models)
def load_test_categories(sender, **kwargs):
r1, c = models.Category.objects.get_or_create(name="root1")
r2, c = models.Category.objects.get_or_create(name="root2")
c11, c = models.Category.objects.get_or_create(name="cat11", parent=r1)
c12, c = models.Category.objects.get_or_create(name="cat12", parent=r1)
c111, c = models.Category.objects.get_or_create(name="cat111", parent=c11)
c112, c = models.Category.objects.get_or_create(name="cat112", parent=c11)
c113, c = models.Category.objects.get_or_create(name="cat113", parent=c11)
c21, c = models.Category.objects.get_or_create(name="cat21", parent=r2)
c22, c = models.Category.objects.get_or_create(name="cat22", parent=r2)
| Update tests to work under django 1.8+ | ## Code Before:
from bitcategory import models
from django.dispatch import receiver
from django.db.models.signals import post_syncdb
@receiver(post_syncdb, sender=models)
def load_test_categories(sender, **kwargs):
r1, c = models.Category.objects.get_or_create(name="root1")
r2, c = models.Category.objects.get_or_create(name="root2")
c11, c = models.Category.objects.get_or_create(name="cat11", parent=r1)
c12, c = models.Category.objects.get_or_create(name="cat12", parent=r1)
c111, c = models.Category.objects.get_or_create(name="cat111", parent=c11)
c112, c = models.Category.objects.get_or_create(name="cat112", parent=c11)
c113, c = models.Category.objects.get_or_create(name="cat113", parent=c11)
c21, c = models.Category.objects.get_or_create(name="cat21", parent=r2)
c22, c = models.Category.objects.get_or_create(name="cat22", parent=r2)
## Instruction:
Update tests to work under django 1.8+
## Code After:
from bitcategory import models
from django.dispatch import receiver
from django.db.models.signals import post_migrate
@receiver(post_migrate, sender=models)
def load_test_categories(sender, **kwargs):
r1, c = models.Category.objects.get_or_create(name="root1")
r2, c = models.Category.objects.get_or_create(name="root2")
c11, c = models.Category.objects.get_or_create(name="cat11", parent=r1)
c12, c = models.Category.objects.get_or_create(name="cat12", parent=r1)
c111, c = models.Category.objects.get_or_create(name="cat111", parent=c11)
c112, c = models.Category.objects.get_or_create(name="cat112", parent=c11)
c113, c = models.Category.objects.get_or_create(name="cat113", parent=c11)
c21, c = models.Category.objects.get_or_create(name="cat21", parent=r2)
c22, c = models.Category.objects.get_or_create(name="cat22", parent=r2)
|
5749d976dee7d8a51e25842b528448a077a8f800 | report_compassion/models/ir_actions_report.py | report_compassion/models/ir_actions_report.py |
from odoo import models, api
class IrActionsReport(models.Model):
_inherit = "ir.actions.report"
@api.multi
def behaviour(self):
"""
Change behaviour to return user preference in priority.
:return: report action for printing.
"""
result = super().behaviour()
# Retrieve user default values
user = self.env.user
if user.printing_action:
default_action = user.printing_action
for key, val in result.iteritems():
result[key]["action"] = default_action
if user.printing_printer_id:
default_printer = user.printing_printer_id
for key, val in result.iteritems():
result[key]["printer"] = default_printer
return result
|
from odoo import models, api
class IrActionsReport(models.Model):
_inherit = "ir.actions.report"
@api.multi
def behaviour(self):
"""
Change behaviour to return user preference in priority.
:return: report action for printing.
"""
result = super().behaviour()
# Retrieve user default values
result.update(self._get_user_default_print_behaviour())
return result
| FIX default behaviour of printing | FIX default behaviour of printing
| Python | agpl-3.0 | CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland |
from odoo import models, api
class IrActionsReport(models.Model):
_inherit = "ir.actions.report"
@api.multi
def behaviour(self):
"""
Change behaviour to return user preference in priority.
:return: report action for printing.
"""
result = super().behaviour()
# Retrieve user default values
+ result.update(self._get_user_default_print_behaviour())
- user = self.env.user
- if user.printing_action:
- default_action = user.printing_action
- for key, val in result.iteritems():
- result[key]["action"] = default_action
- if user.printing_printer_id:
- default_printer = user.printing_printer_id
- for key, val in result.iteritems():
- result[key]["printer"] = default_printer
return result
| FIX default behaviour of printing | ## Code Before:
from odoo import models, api
class IrActionsReport(models.Model):
_inherit = "ir.actions.report"
@api.multi
def behaviour(self):
"""
Change behaviour to return user preference in priority.
:return: report action for printing.
"""
result = super().behaviour()
# Retrieve user default values
user = self.env.user
if user.printing_action:
default_action = user.printing_action
for key, val in result.iteritems():
result[key]["action"] = default_action
if user.printing_printer_id:
default_printer = user.printing_printer_id
for key, val in result.iteritems():
result[key]["printer"] = default_printer
return result
## Instruction:
FIX default behaviour of printing
## Code After:
from odoo import models, api
class IrActionsReport(models.Model):
_inherit = "ir.actions.report"
@api.multi
def behaviour(self):
"""
Change behaviour to return user preference in priority.
:return: report action for printing.
"""
result = super().behaviour()
# Retrieve user default values
result.update(self._get_user_default_print_behaviour())
return result
|
a1a8ef302ac24d56a36d0671eb692943d14c4ddf | whats_open/website/urls.py | whats_open/website/urls.py | from __future__ import (absolute_import, division, print_function,
unicode_literals)
# Django Imports
from django.conf.urls import include, url
# App Imports
from .views import CategoryViewSet, FacilityViewSet, ScheduleViewSet
# Other Imports
from rest_framework.routers import DefaultRouter
# Instiantiate our DefaultRouter
ROUTER = DefaultRouter()
# Register views to the API router
ROUTER.register(r'categories', CategoryViewSet)
ROUTER.register(r'facilities', FacilityViewSet)
ROUTER.register(r'schedules', ScheduleViewSet)
urlpatterns = [
# /api - Root API URL
url(r'^api/', include(ROUTER.urls)),
]
| from __future__ import (absolute_import, division, print_function,
unicode_literals)
# Django Imports
from django.conf.urls import include, url
from django.views.generic.base import RedirectView
# App Imports
from .views import CategoryViewSet, FacilityViewSet, ScheduleViewSet
# Other Imports
from rest_framework.routers import DefaultRouter
# Instiantiate our DefaultRouter
ROUTER = DefaultRouter()
# Register views to the API router
ROUTER.register(r'categories', CategoryViewSet)
ROUTER.register(r'facilities', FacilityViewSet)
ROUTER.register(r'schedules', ScheduleViewSet)
urlpatterns = [
# / - Default route
# We redirect to /api since this is in reality the default page for the API
url(r'^$', RedirectView.as_view(url='/api')),
# /api - Root API URL
url(r'^api/', include(ROUTER.urls)),
]
| Add a redirect for / -> /api | refactor: Add a redirect for / -> /api
- This got really annoying in development to remember to go to /api all the time
- Potentially in the future we will build a test landing page on / to test that
the API works instead of having to rely on third party tools or just manual clicks
- Until then, this will make life easier (imo)
| Python | apache-2.0 | srct/whats-open,srct/whats-open,srct/whats-open | from __future__ import (absolute_import, division, print_function,
unicode_literals)
# Django Imports
from django.conf.urls import include, url
+ from django.views.generic.base import RedirectView
# App Imports
from .views import CategoryViewSet, FacilityViewSet, ScheduleViewSet
# Other Imports
from rest_framework.routers import DefaultRouter
# Instiantiate our DefaultRouter
ROUTER = DefaultRouter()
# Register views to the API router
ROUTER.register(r'categories', CategoryViewSet)
ROUTER.register(r'facilities', FacilityViewSet)
ROUTER.register(r'schedules', ScheduleViewSet)
urlpatterns = [
+ # / - Default route
+ # We redirect to /api since this is in reality the default page for the API
+ url(r'^$', RedirectView.as_view(url='/api')),
# /api - Root API URL
url(r'^api/', include(ROUTER.urls)),
]
| Add a redirect for / -> /api | ## Code Before:
from __future__ import (absolute_import, division, print_function,
unicode_literals)
# Django Imports
from django.conf.urls import include, url
# App Imports
from .views import CategoryViewSet, FacilityViewSet, ScheduleViewSet
# Other Imports
from rest_framework.routers import DefaultRouter
# Instiantiate our DefaultRouter
ROUTER = DefaultRouter()
# Register views to the API router
ROUTER.register(r'categories', CategoryViewSet)
ROUTER.register(r'facilities', FacilityViewSet)
ROUTER.register(r'schedules', ScheduleViewSet)
urlpatterns = [
# /api - Root API URL
url(r'^api/', include(ROUTER.urls)),
]
## Instruction:
Add a redirect for / -> /api
## Code After:
from __future__ import (absolute_import, division, print_function,
unicode_literals)
# Django Imports
from django.conf.urls import include, url
from django.views.generic.base import RedirectView
# App Imports
from .views import CategoryViewSet, FacilityViewSet, ScheduleViewSet
# Other Imports
from rest_framework.routers import DefaultRouter
# Instiantiate our DefaultRouter
ROUTER = DefaultRouter()
# Register views to the API router
ROUTER.register(r'categories', CategoryViewSet)
ROUTER.register(r'facilities', FacilityViewSet)
ROUTER.register(r'schedules', ScheduleViewSet)
urlpatterns = [
# / - Default route
# We redirect to /api since this is in reality the default page for the API
url(r'^$', RedirectView.as_view(url='/api')),
# /api - Root API URL
url(r'^api/', include(ROUTER.urls)),
]
|
d3f4d43b36b8d21a3102389d53bf3f1af4e00d79 | st2common/st2common/runners/base_action.py | st2common/st2common/runners/base_action.py |
import abc
import six
from st2common.runners.utils import get_logger_for_python_runner_action
@six.add_metaclass(abc.ABCMeta)
class Action(object):
"""
Base action class other Python actions should inherit from.
"""
description = None
def __init__(self, config=None, action_service=None):
"""
:param config: Action config.
:type config: ``dict``
:param action_service: ActionService object.
:type action_service: :class:`ActionService~
"""
self.config = config or {}
self.action_service = action_service
self.logger = get_logger_for_python_runner_action(action_name=self.__class__.__name__)
@abc.abstractmethod
def run(self, **kwargs):
pass
|
import abc
import six
from st2common.runners.utils import get_logger_for_python_runner_action
__all__ = [
'Action'
]
@six.add_metaclass(abc.ABCMeta)
class Action(object):
"""
Base action class other Python actions should inherit from.
"""
description = None
def __init__(self, config=None, action_service=None):
"""
:param config: Action config.
:type config: ``dict``
:param action_service: ActionService object.
:type action_service: :class:`ActionService~
"""
self.config = config or {}
self.action_service = action_service
self.logger = get_logger_for_python_runner_action(action_name=self.__class__.__name__)
@abc.abstractmethod
def run(self, **kwargs):
pass
| Add _all__ to the module. | Add _all__ to the module.
| Python | apache-2.0 | StackStorm/st2,nzlosh/st2,Plexxi/st2,nzlosh/st2,Plexxi/st2,nzlosh/st2,Plexxi/st2,Plexxi/st2,StackStorm/st2,StackStorm/st2,nzlosh/st2,StackStorm/st2 |
import abc
import six
from st2common.runners.utils import get_logger_for_python_runner_action
+
+ __all__ = [
+ 'Action'
+ ]
@six.add_metaclass(abc.ABCMeta)
class Action(object):
"""
Base action class other Python actions should inherit from.
"""
description = None
def __init__(self, config=None, action_service=None):
"""
:param config: Action config.
:type config: ``dict``
:param action_service: ActionService object.
:type action_service: :class:`ActionService~
"""
self.config = config or {}
self.action_service = action_service
self.logger = get_logger_for_python_runner_action(action_name=self.__class__.__name__)
@abc.abstractmethod
def run(self, **kwargs):
pass
| Add _all__ to the module. | ## Code Before:
import abc
import six
from st2common.runners.utils import get_logger_for_python_runner_action
@six.add_metaclass(abc.ABCMeta)
class Action(object):
"""
Base action class other Python actions should inherit from.
"""
description = None
def __init__(self, config=None, action_service=None):
"""
:param config: Action config.
:type config: ``dict``
:param action_service: ActionService object.
:type action_service: :class:`ActionService~
"""
self.config = config or {}
self.action_service = action_service
self.logger = get_logger_for_python_runner_action(action_name=self.__class__.__name__)
@abc.abstractmethod
def run(self, **kwargs):
pass
## Instruction:
Add _all__ to the module.
## Code After:
import abc
import six
from st2common.runners.utils import get_logger_for_python_runner_action
__all__ = [
'Action'
]
@six.add_metaclass(abc.ABCMeta)
class Action(object):
"""
Base action class other Python actions should inherit from.
"""
description = None
def __init__(self, config=None, action_service=None):
"""
:param config: Action config.
:type config: ``dict``
:param action_service: ActionService object.
:type action_service: :class:`ActionService~
"""
self.config = config or {}
self.action_service = action_service
self.logger = get_logger_for_python_runner_action(action_name=self.__class__.__name__)
@abc.abstractmethod
def run(self, **kwargs):
pass
|
ee6c7caabdfcd0bddd9b92d05cddd8b6be7cbe10 | tests/functional/test_pip_runner_script.py | tests/functional/test_pip_runner_script.py | import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the environment
script.pip("uninstall", "pip", "--yes", use_module=True)
script.pip("--version", expect_error=True)
# The runner script should still invoke a usable pip
result = script.run("python", os.fspath(runner), "--version")
assert __version__ in result.stdout
| import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the environment
script.pip("uninstall", "pip", "--yes", use_module=True)
# We don't use script.pip to check here, as when testing a
# zipapp, script.pip will run pip from the zipapp.
script.run("python", "-c", "import pip", expect_error=True)
# The runner script should still invoke a usable pip
result = script.run("python", os.fspath(runner), "--version")
assert __version__ in result.stdout
| Fix test_runner_work_in_environments_with_no_pip to work under --use-zipapp | Fix test_runner_work_in_environments_with_no_pip to work under --use-zipapp
| Python | mit | pradyunsg/pip,pypa/pip,pfmoore/pip,pfmoore/pip,sbidoul/pip,pradyunsg/pip,sbidoul/pip,pypa/pip | import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the environment
script.pip("uninstall", "pip", "--yes", use_module=True)
- script.pip("--version", expect_error=True)
+ # We don't use script.pip to check here, as when testing a
+ # zipapp, script.pip will run pip from the zipapp.
+ script.run("python", "-c", "import pip", expect_error=True)
# The runner script should still invoke a usable pip
result = script.run("python", os.fspath(runner), "--version")
assert __version__ in result.stdout
| Fix test_runner_work_in_environments_with_no_pip to work under --use-zipapp | ## Code Before:
import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the environment
script.pip("uninstall", "pip", "--yes", use_module=True)
script.pip("--version", expect_error=True)
# The runner script should still invoke a usable pip
result = script.run("python", os.fspath(runner), "--version")
assert __version__ in result.stdout
## Instruction:
Fix test_runner_work_in_environments_with_no_pip to work under --use-zipapp
## Code After:
import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the environment
script.pip("uninstall", "pip", "--yes", use_module=True)
# We don't use script.pip to check here, as when testing a
# zipapp, script.pip will run pip from the zipapp.
script.run("python", "-c", "import pip", expect_error=True)
# The runner script should still invoke a usable pip
result = script.run("python", os.fspath(runner), "--version")
assert __version__ in result.stdout
|
0e835c6381374c5b00b7387057d056d679f635c4 | zproject/legacy_urls.py | zproject/legacy_urls.py | from django.conf.urls import url
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
]
| from django.urls import path
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
]
| Migrate legacy urls to use modern django pattern. | urls: Migrate legacy urls to use modern django pattern.
| Python | apache-2.0 | shubhamdhama/zulip,punchagan/zulip,kou/zulip,showell/zulip,hackerkid/zulip,timabbott/zulip,eeshangarg/zulip,kou/zulip,andersk/zulip,zulip/zulip,synicalsyntax/zulip,zulip/zulip,andersk/zulip,shubhamdhama/zulip,showell/zulip,kou/zulip,hackerkid/zulip,shubhamdhama/zulip,andersk/zulip,eeshangarg/zulip,brainwane/zulip,shubhamdhama/zulip,hackerkid/zulip,punchagan/zulip,showell/zulip,brainwane/zulip,rht/zulip,rht/zulip,shubhamdhama/zulip,hackerkid/zulip,eeshangarg/zulip,hackerkid/zulip,brainwane/zulip,brainwane/zulip,punchagan/zulip,punchagan/zulip,punchagan/zulip,synicalsyntax/zulip,timabbott/zulip,zulip/zulip,rht/zulip,eeshangarg/zulip,brainwane/zulip,punchagan/zulip,timabbott/zulip,timabbott/zulip,synicalsyntax/zulip,synicalsyntax/zulip,hackerkid/zulip,synicalsyntax/zulip,showell/zulip,timabbott/zulip,rht/zulip,kou/zulip,rht/zulip,synicalsyntax/zulip,timabbott/zulip,rht/zulip,eeshangarg/zulip,zulip/zulip,zulip/zulip,eeshangarg/zulip,zulip/zulip,punchagan/zulip,hackerkid/zulip,kou/zulip,andersk/zulip,rht/zulip,andersk/zulip,kou/zulip,shubhamdhama/zulip,andersk/zulip,showell/zulip,timabbott/zulip,synicalsyntax/zulip,eeshangarg/zulip,showell/zulip,kou/zulip,brainwane/zulip,brainwane/zulip,andersk/zulip,zulip/zulip,shubhamdhama/zulip,showell/zulip | - from django.conf.urls import url
+ from django.urls import path
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
- url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
+ path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
]
| Migrate legacy urls to use modern django pattern. | ## Code Before:
from django.conf.urls import url
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
]
## Instruction:
Migrate legacy urls to use modern django pattern.
## Code After:
from django.urls import path
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
]
|
5b9f0270aaa53a562ca65fa74769885621da4a8e | website/addons/s3/__init__.py | website/addons/s3/__init__.py | import os
from . import model
from . import routes
from . import views
MODELS = [model.AddonS3UserSettings, model.AddonS3NodeSettings, model.S3GuidFile]
USER_SETTINGS_MODEL = model.AddonS3UserSettings
NODE_SETTINGS_MODEL = model.AddonS3NodeSettings
ROUTES = [routes.settings_routes]
SHORT_NAME = 's3'
FULL_NAME = 'Amazon Simple Storage Service'
OWNERS = ['user', 'node']
ADDED_DEFAULT = []
ADDED_MANDATORY = []
VIEWS = []
CONFIGS = ['user', 'node']
CATEGORIES = ['storage']
INCLUDE_JS = {}
INCLUDE_CSS = {
'widget': [],
'page': [],
}
HAS_HGRID_FILES = True
GET_HGRID_DATA = views.hgrid.s3_hgrid_data
# 1024 ** 1024 # There really shouldnt be a limit...
MAX_FILE_SIZE = 128 # MB
HERE = os.path.dirname(os.path.abspath(__file__))
NODE_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_node_settings.mako')
USER_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_user_settings.mako')
| import os
from . import model
from . import routes
from . import views
MODELS = [model.AddonS3UserSettings, model.AddonS3NodeSettings, model.S3GuidFile]
USER_SETTINGS_MODEL = model.AddonS3UserSettings
NODE_SETTINGS_MODEL = model.AddonS3NodeSettings
ROUTES = [routes.settings_routes]
SHORT_NAME = 's3'
FULL_NAME = 'Amazon S3'
OWNERS = ['user', 'node']
ADDED_DEFAULT = []
ADDED_MANDATORY = []
VIEWS = []
CONFIGS = ['user', 'node']
CATEGORIES = ['storage']
INCLUDE_JS = {}
INCLUDE_CSS = {
'widget': [],
'page': [],
}
HAS_HGRID_FILES = True
GET_HGRID_DATA = views.hgrid.s3_hgrid_data
# 1024 ** 1024 # There really shouldnt be a limit...
MAX_FILE_SIZE = 128 # MB
HERE = os.path.dirname(os.path.abspath(__file__))
NODE_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_node_settings.mako')
USER_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_user_settings.mako')
| Change S3 full name to Amazon S3 | Change S3 full name to Amazon S3
| Python | apache-2.0 | hmoco/osf.io,jmcarp/osf.io,abought/osf.io,amyshi188/osf.io,brandonPurvis/osf.io,sloria/osf.io,barbour-em/osf.io,lyndsysimon/osf.io,GageGaskins/osf.io,brandonPurvis/osf.io,HarryRybacki/osf.io,pattisdr/osf.io,HalcyonChimera/osf.io,wearpants/osf.io,TomBaxter/osf.io,samanehsan/osf.io,zachjanicki/osf.io,mluo613/osf.io,ZobairAlijan/osf.io,ckc6cz/osf.io,MerlinZhang/osf.io,CenterForOpenScience/osf.io,jmcarp/osf.io,caneruguz/osf.io,HarryRybacki/osf.io,njantrania/osf.io,fabianvf/osf.io,danielneis/osf.io,jnayak1/osf.io,binoculars/osf.io,alexschiller/osf.io,asanfilippo7/osf.io,acshi/osf.io,leb2dg/osf.io,haoyuchen1992/osf.io,cwisecarver/osf.io,Ghalko/osf.io,jolene-esposito/osf.io,ticklemepierce/osf.io,zamattiac/osf.io,arpitar/osf.io,samchrisinger/osf.io,lyndsysimon/osf.io,hmoco/osf.io,ticklemepierce/osf.io,reinaH/osf.io,rdhyee/osf.io,binoculars/osf.io,fabianvf/osf.io,GageGaskins/osf.io,GageGaskins/osf.io,danielneis/osf.io,jolene-esposito/osf.io,kch8qx/osf.io,cldershem/osf.io,ckc6cz/osf.io,bdyetton/prettychart,revanthkolli/osf.io,brandonPurvis/osf.io,TomHeatwole/osf.io,mattclark/osf.io,SSJohns/osf.io,zachjanicki/osf.io,reinaH/osf.io,zamattiac/osf.io,cwisecarver/osf.io,mattclark/osf.io,bdyetton/prettychart,leb2dg/osf.io,HalcyonChimera/osf.io,monikagrabowska/osf.io,dplorimer/osf,billyhunt/osf.io,barbour-em/osf.io,kch8qx/osf.io,HalcyonChimera/osf.io,ckc6cz/osf.io,caseyrygt/osf.io,lamdnhan/osf.io,rdhyee/osf.io,Nesiehr/osf.io,acshi/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,abought/osf.io,CenterForOpenScience/osf.io,revanthkolli/osf.io,caseyrygt/osf.io,TomBaxter/osf.io,brandonPurvis/osf.io,acshi/osf.io,TomHeatwole/osf.io,amyshi188/osf.io,erinspace/osf.io,icereval/osf.io,jnayak1/osf.io,jmcarp/osf.io,mattclark/osf.io,doublebits/osf.io,aaxelb/osf.io,sbt9uc/osf.io,kwierman/osf.io,rdhyee/osf.io,SSJohns/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,caneruguz/osf.io,pattisdr/osf.io,caseyrygt/osf.io,petermalcolm/osf.io,DanielSBrown/osf.io,Ghalko/osf.io,cslzchen/osf.io,alexschiller/osf.io,amyshi188/osf.io,sbt9uc/osf.io,monikagrabowska/osf.io,fabianvf/osf.io,RomanZWang/osf.io,ZobairAlijan/osf.io,RomanZWang/osf.io,emetsger/osf.io,billyhunt/osf.io,emetsger/osf.io,TomHeatwole/osf.io,laurenrevere/osf.io,doublebits/osf.io,GaryKriebel/osf.io,ticklemepierce/osf.io,RomanZWang/osf.io,doublebits/osf.io,MerlinZhang/osf.io,brianjgeiger/osf.io,chrisseto/osf.io,lamdnhan/osf.io,TomHeatwole/osf.io,doublebits/osf.io,lamdnhan/osf.io,revanthkolli/osf.io,kwierman/osf.io,jeffreyliu3230/osf.io,jinluyuan/osf.io,lyndsysimon/osf.io,baylee-d/osf.io,erinspace/osf.io,barbour-em/osf.io,saradbowman/osf.io,felliott/osf.io,barbour-em/osf.io,wearpants/osf.io,cwisecarver/osf.io,hmoco/osf.io,jeffreyliu3230/osf.io,Johnetordoff/osf.io,alexschiller/osf.io,jeffreyliu3230/osf.io,sbt9uc/osf.io,aaxelb/osf.io,Nesiehr/osf.io,felliott/osf.io,njantrania/osf.io,ZobairAlijan/osf.io,danielneis/osf.io,reinaH/osf.io,icereval/osf.io,amyshi188/osf.io,KAsante95/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,haoyuchen1992/osf.io,billyhunt/osf.io,cslzchen/osf.io,samchrisinger/osf.io,HarryRybacki/osf.io,felliott/osf.io,hmoco/osf.io,ticklemepierce/osf.io,baylee-d/osf.io,fabianvf/osf.io,lyndsysimon/osf.io,brianjgeiger/osf.io,cslzchen/osf.io,binoculars/osf.io,cldershem/osf.io,chrisseto/osf.io,caneruguz/osf.io,rdhyee/osf.io,asanfilippo7/osf.io,mfraezz/osf.io,adlius/osf.io,RomanZWang/osf.io,DanielSBrown/osf.io,emetsger/osf.io,brianjgeiger/osf.io,kwierman/osf.io,pattisdr/osf.io,cosenal/osf.io,danielneis/osf.io,reinaH/osf.io,sloria/osf.io,jeffreyliu3230/osf.io,GaryKriebel/osf.io,jnayak1/osf.io,petermalcolm/osf.io,DanielSBrown/osf.io,wearpants/osf.io,emetsger/osf.io,HarryRybacki/osf.io,asanfilippo7/osf.io,zkraime/osf.io,mfraezz/osf.io,KAsante95/osf.io,arpitar/osf.io,aaxelb/osf.io,caseyrollins/osf.io,acshi/osf.io,Johnetordoff/osf.io,arpitar/osf.io,chrisseto/osf.io,mluke93/osf.io,dplorimer/osf,samanehsan/osf.io,haoyuchen1992/osf.io,KAsante95/osf.io,revanthkolli/osf.io,cosenal/osf.io,baylee-d/osf.io,acshi/osf.io,caseyrollins/osf.io,felliott/osf.io,billyhunt/osf.io,sbt9uc/osf.io,GageGaskins/osf.io,Ghalko/osf.io,jolene-esposito/osf.io,crcresearch/osf.io,CenterForOpenScience/osf.io,doublebits/osf.io,chennan47/osf.io,MerlinZhang/osf.io,mluo613/osf.io,saradbowman/osf.io,MerlinZhang/osf.io,petermalcolm/osf.io,cwisecarver/osf.io,mfraezz/osf.io,samchrisinger/osf.io,wearpants/osf.io,njantrania/osf.io,kwierman/osf.io,cosenal/osf.io,arpitar/osf.io,zkraime/osf.io,billyhunt/osf.io,abought/osf.io,jinluyuan/osf.io,njantrania/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,samanehsan/osf.io,KAsante95/osf.io,zkraime/osf.io,laurenrevere/osf.io,ckc6cz/osf.io,DanielSBrown/osf.io,Johnetordoff/osf.io,mfraezz/osf.io,jolene-esposito/osf.io,adlius/osf.io,jnayak1/osf.io,ZobairAlijan/osf.io,aaxelb/osf.io,dplorimer/osf,zachjanicki/osf.io,brianjgeiger/osf.io,RomanZWang/osf.io,bdyetton/prettychart,zkraime/osf.io,samchrisinger/osf.io,adlius/osf.io,SSJohns/osf.io,samanehsan/osf.io,chennan47/osf.io,leb2dg/osf.io,GaryKriebel/osf.io,caseyrollins/osf.io,mluke93/osf.io,crcresearch/osf.io,Ghalko/osf.io,sloria/osf.io,erinspace/osf.io,mluo613/osf.io,abought/osf.io,adlius/osf.io,cldershem/osf.io,icereval/osf.io,crcresearch/osf.io,mluke93/osf.io,cldershem/osf.io,Johnetordoff/osf.io,lamdnhan/osf.io,laurenrevere/osf.io,alexschiller/osf.io,GageGaskins/osf.io,zachjanicki/osf.io,kch8qx/osf.io,bdyetton/prettychart,petermalcolm/osf.io,kch8qx/osf.io,Nesiehr/osf.io,dplorimer/osf,mluke93/osf.io,jmcarp/osf.io,asanfilippo7/osf.io,SSJohns/osf.io,CenterForOpenScience/osf.io,brandonPurvis/osf.io,zamattiac/osf.io,kch8qx/osf.io,cslzchen/osf.io,chennan47/osf.io,GaryKriebel/osf.io,cosenal/osf.io,alexschiller/osf.io,jinluyuan/osf.io,KAsante95/osf.io,zamattiac/osf.io,mluo613/osf.io,mluo613/osf.io,caseyrygt/osf.io,chrisseto/osf.io,jinluyuan/osf.io,haoyuchen1992/osf.io | import os
from . import model
from . import routes
from . import views
MODELS = [model.AddonS3UserSettings, model.AddonS3NodeSettings, model.S3GuidFile]
USER_SETTINGS_MODEL = model.AddonS3UserSettings
NODE_SETTINGS_MODEL = model.AddonS3NodeSettings
ROUTES = [routes.settings_routes]
SHORT_NAME = 's3'
- FULL_NAME = 'Amazon Simple Storage Service'
+ FULL_NAME = 'Amazon S3'
OWNERS = ['user', 'node']
ADDED_DEFAULT = []
ADDED_MANDATORY = []
VIEWS = []
CONFIGS = ['user', 'node']
CATEGORIES = ['storage']
INCLUDE_JS = {}
INCLUDE_CSS = {
'widget': [],
'page': [],
}
HAS_HGRID_FILES = True
GET_HGRID_DATA = views.hgrid.s3_hgrid_data
# 1024 ** 1024 # There really shouldnt be a limit...
MAX_FILE_SIZE = 128 # MB
HERE = os.path.dirname(os.path.abspath(__file__))
NODE_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_node_settings.mako')
USER_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_user_settings.mako')
| Change S3 full name to Amazon S3 | ## Code Before:
import os
from . import model
from . import routes
from . import views
MODELS = [model.AddonS3UserSettings, model.AddonS3NodeSettings, model.S3GuidFile]
USER_SETTINGS_MODEL = model.AddonS3UserSettings
NODE_SETTINGS_MODEL = model.AddonS3NodeSettings
ROUTES = [routes.settings_routes]
SHORT_NAME = 's3'
FULL_NAME = 'Amazon Simple Storage Service'
OWNERS = ['user', 'node']
ADDED_DEFAULT = []
ADDED_MANDATORY = []
VIEWS = []
CONFIGS = ['user', 'node']
CATEGORIES = ['storage']
INCLUDE_JS = {}
INCLUDE_CSS = {
'widget': [],
'page': [],
}
HAS_HGRID_FILES = True
GET_HGRID_DATA = views.hgrid.s3_hgrid_data
# 1024 ** 1024 # There really shouldnt be a limit...
MAX_FILE_SIZE = 128 # MB
HERE = os.path.dirname(os.path.abspath(__file__))
NODE_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_node_settings.mako')
USER_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_user_settings.mako')
## Instruction:
Change S3 full name to Amazon S3
## Code After:
import os
from . import model
from . import routes
from . import views
MODELS = [model.AddonS3UserSettings, model.AddonS3NodeSettings, model.S3GuidFile]
USER_SETTINGS_MODEL = model.AddonS3UserSettings
NODE_SETTINGS_MODEL = model.AddonS3NodeSettings
ROUTES = [routes.settings_routes]
SHORT_NAME = 's3'
FULL_NAME = 'Amazon S3'
OWNERS = ['user', 'node']
ADDED_DEFAULT = []
ADDED_MANDATORY = []
VIEWS = []
CONFIGS = ['user', 'node']
CATEGORIES = ['storage']
INCLUDE_JS = {}
INCLUDE_CSS = {
'widget': [],
'page': [],
}
HAS_HGRID_FILES = True
GET_HGRID_DATA = views.hgrid.s3_hgrid_data
# 1024 ** 1024 # There really shouldnt be a limit...
MAX_FILE_SIZE = 128 # MB
HERE = os.path.dirname(os.path.abspath(__file__))
NODE_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_node_settings.mako')
USER_SETTINGS_TEMPLATE = os.path.join(HERE, 'templates', 's3_user_settings.mako')
|
d18ff30bbddde5049ffbe23bce19288c3c47e41b | posts/views.py | posts/views.py | from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from .models import Post
class PostListView(ListView):
model = Post
context_object_name = 'posts'
class PostDetailView(DetailView):
model = Post
context_object_name = 'post'
| from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from .models import Post
class PostListView(ListView):
model = Post
context_object_name = 'posts'
def get_queryset(self):
"""
Order posts by the day they were added, from newest, to oldest.
"""
queryset = super(PostListView, self).get_queryset()
return queryset.order_by('-added_at')
class PostDetailView(DetailView):
model = Post
context_object_name = 'post'
| Order posts from newest to oldest | posts: Order posts from newest to oldest
| Python | mit | rtrembecky/roots,tbabej/roots,rtrembecky/roots,tbabej/roots,matus-stehlik/roots,matus-stehlik/roots,matus-stehlik/glowing-batman,matus-stehlik/roots,matus-stehlik/glowing-batman,rtrembecky/roots,tbabej/roots | from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from .models import Post
class PostListView(ListView):
model = Post
context_object_name = 'posts'
+ def get_queryset(self):
+ """
+ Order posts by the day they were added, from newest, to oldest.
+ """
+
+ queryset = super(PostListView, self).get_queryset()
+ return queryset.order_by('-added_at')
+
class PostDetailView(DetailView):
model = Post
context_object_name = 'post'
| Order posts from newest to oldest | ## Code Before:
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from .models import Post
class PostListView(ListView):
model = Post
context_object_name = 'posts'
class PostDetailView(DetailView):
model = Post
context_object_name = 'post'
## Instruction:
Order posts from newest to oldest
## Code After:
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from .models import Post
class PostListView(ListView):
model = Post
context_object_name = 'posts'
def get_queryset(self):
"""
Order posts by the day they were added, from newest, to oldest.
"""
queryset = super(PostListView, self).get_queryset()
return queryset.order_by('-added_at')
class PostDetailView(DetailView):
model = Post
context_object_name = 'post'
|
b583c5fb00d1ebfa0458a6233be85d8b56173abf | python/printbag.py | python/printbag.py |
import sys
import logging
import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
if topics is None:
#topics = ['/tf', '/scan']
topics = ['/scan', '/flagbutton_pressed']
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> \n\n'
'Print contents of rosbag file.'
).format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag)
sys.exit()
| import sys
import logging
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
def print_bag(bag, topics=None):
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} [topics] <rosbag> \n\n'
'topics:\tcomma-separated list of topics\n\n'
'Print contents of rosbag file. If topics is not provided, \n'
'all topics are printed\n'
).format(__file__))
sys.exit(1)
topics = None
if len(sys.argv) == 3:
topics = [t.strip() for t in sys.argv[1].split(',')]
filename = sys.argv[2]
else:
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag, topics)
sys.exit()
| Add argument to specify bag topics | Add argument to specify bag topics
| Python | bsd-2-clause | oliverlee/antlia | -
import sys
import logging
-
- import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
- from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
- if topics is None:
- #topics = ['/tf', '/scan']
- topics = ['/scan', '/flagbutton_pressed']
-
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
- print(('Usage: {} <rosbag> \n\n'
+ print(('Usage: {} [topics] <rosbag> \n\n'
+ 'topics:\tcomma-separated list of topics\n\n'
- 'Print contents of rosbag file.'
+ 'Print contents of rosbag file. If topics is not provided, \n'
+ 'all topics are printed\n'
).format(__file__))
sys.exit(1)
- outfile = None
+ topics = None
+ if len(sys.argv) == 3:
+ topics = [t.strip() for t in sys.argv[1].split(',')]
+ filename = sys.argv[2]
+ else:
- filename = sys.argv[1]
+ filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
- print_bag(bag)
+ print_bag(bag, topics)
sys.exit()
| Add argument to specify bag topics | ## Code Before:
import sys
import logging
import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
if topics is None:
#topics = ['/tf', '/scan']
topics = ['/scan', '/flagbutton_pressed']
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> \n\n'
'Print contents of rosbag file.'
).format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag)
sys.exit()
## Instruction:
Add argument to specify bag topics
## Code After:
import sys
import logging
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
def print_bag(bag, topics=None):
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} [topics] <rosbag> \n\n'
'topics:\tcomma-separated list of topics\n\n'
'Print contents of rosbag file. If topics is not provided, \n'
'all topics are printed\n'
).format(__file__))
sys.exit(1)
topics = None
if len(sys.argv) == 3:
topics = [t.strip() for t in sys.argv[1].split(',')]
filename = sys.argv[2]
else:
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag, topics)
sys.exit()
|
ca06bf1d52cd51ccec178c98ad407bfe59f1ada1 | strobe.py | strobe.py | import RPi.GPIO as GPIO
from time import sleep
def onoff(period, pin):
"""Symmetric square wave, equal time on/off"""
half_cycle = period / 2.0
GPIO.output(pin, GPIO.HIGH)
sleep(half_cycle)
GPIO.output(pin, GPIO.LOW)
sleep(half_cycle)
def strobe(freq, dur, pin):
nflashes = freq * dur
seconds_to_sleep = 1.0 / freq
# Use Raspberry-Pi board pin numbers. In other words, 11 means pin
# number 11, not GPIO 11.
GPIO.setmode(GPIO.BOARD)
GPIO.setup(pin, GPIO.OUT) # requires root?
for i in range(nflashes):
onoff(seconds_to_sleep, pin)
GPIO.cleanup()
|
import RPi.GPIO as GPIO
from time import sleep
def onoff(ontime, offtime, pin):
GPIO.output(pin, GPIO.HIGH)
sleep(ontime)
GPIO.output(pin, GPIO.LOW)
sleep(offtime)
def strobe(freq, dur, pin):
nflashes = freq * dur
period = 1.0 / freq
# Use Raspberry-Pi board pin numbers. In other words, 11 means pin
# number 11, not GPIO 11.
GPIO.setmode(GPIO.BOARD)
GPIO.setup(pin, GPIO.OUT) # requires root?
for i in range(nflashes):
onoff(period/2.0, period/2.0, pin)
GPIO.cleanup()
| Make onoff function more versatile | Make onoff function more versatile
| Python | mit | zimolzak/Raspberry-Pi-newbie,zimolzak/Raspberry-Pi-newbie,zimolzak/Raspberry-Pi-newbie,zimolzak/Raspberry-Pi-newbie,zimolzak/Raspberry-Pi-newbie | +
import RPi.GPIO as GPIO
from time import sleep
+ def onoff(ontime, offtime, pin):
- def onoff(period, pin):
- """Symmetric square wave, equal time on/off"""
- half_cycle = period / 2.0
GPIO.output(pin, GPIO.HIGH)
- sleep(half_cycle)
+ sleep(ontime)
GPIO.output(pin, GPIO.LOW)
- sleep(half_cycle)
+ sleep(offtime)
def strobe(freq, dur, pin):
nflashes = freq * dur
- seconds_to_sleep = 1.0 / freq
+ period = 1.0 / freq
# Use Raspberry-Pi board pin numbers. In other words, 11 means pin
# number 11, not GPIO 11.
GPIO.setmode(GPIO.BOARD)
GPIO.setup(pin, GPIO.OUT) # requires root?
for i in range(nflashes):
- onoff(seconds_to_sleep, pin)
+ onoff(period/2.0, period/2.0, pin)
GPIO.cleanup()
| Make onoff function more versatile | ## Code Before:
import RPi.GPIO as GPIO
from time import sleep
def onoff(period, pin):
"""Symmetric square wave, equal time on/off"""
half_cycle = period / 2.0
GPIO.output(pin, GPIO.HIGH)
sleep(half_cycle)
GPIO.output(pin, GPIO.LOW)
sleep(half_cycle)
def strobe(freq, dur, pin):
nflashes = freq * dur
seconds_to_sleep = 1.0 / freq
# Use Raspberry-Pi board pin numbers. In other words, 11 means pin
# number 11, not GPIO 11.
GPIO.setmode(GPIO.BOARD)
GPIO.setup(pin, GPIO.OUT) # requires root?
for i in range(nflashes):
onoff(seconds_to_sleep, pin)
GPIO.cleanup()
## Instruction:
Make onoff function more versatile
## Code After:
import RPi.GPIO as GPIO
from time import sleep
def onoff(ontime, offtime, pin):
GPIO.output(pin, GPIO.HIGH)
sleep(ontime)
GPIO.output(pin, GPIO.LOW)
sleep(offtime)
def strobe(freq, dur, pin):
nflashes = freq * dur
period = 1.0 / freq
# Use Raspberry-Pi board pin numbers. In other words, 11 means pin
# number 11, not GPIO 11.
GPIO.setmode(GPIO.BOARD)
GPIO.setup(pin, GPIO.OUT) # requires root?
for i in range(nflashes):
onoff(period/2.0, period/2.0, pin)
GPIO.cleanup()
|
a08de1d3c7f7dfc72c8b3b8e9019d1b7b5ad004e | mdtraj/tests/test_load.py | mdtraj/tests/test_load.py |
from mdtraj import load
from mdtraj.testing import get_fn
def test_load_single():
"""
Just check for any raised errors coming from loading a single file.
"""
load(get_fn('frame0.pdb'))
def test_load_single_list():
"""
See if a single-element list of files is successfully loaded.
"""
load([get_fn('frame0.pdb')])
def test_load_many_list():
"""
See if a multi-element list of files is successfully loaded.
"""
traj = load(2 * [get_fn('frame0.pdb')], discard_overlapping_frames=False)
assert traj.n_frames == 2
| from mdtraj import load
from mdtraj.testing import get_fn
def test_load_single():
"""
Just check for any raised errors coming from loading a single file.
"""
load(get_fn('frame0.pdb'))
def test_load_single_list():
"""
See if a single-element list of files is successfully loaded.
"""
load([get_fn('frame0.pdb')])
def test_load_many_list():
"""
See if a multi-element list of files is successfully loaded.
"""
single = load(get_fn('frame0.pdb'))
double = load(2 * [get_fn('frame0.pdb')], discard_overlapping_frames=False)
assert 2 * single.n_frames == double.n_frames
| Fix test for loading multiple trajectories. | Fix test for loading multiple trajectories.
| Python | lgpl-2.1 | msultan/mdtraj,rmcgibbo/mdtraj,tcmoore3/mdtraj,mdtraj/mdtraj,jchodera/mdtraj,msultan/mdtraj,ctk3b/mdtraj,ctk3b/mdtraj,mdtraj/mdtraj,msultan/mdtraj,mattwthompson/mdtraj,jchodera/mdtraj,jchodera/mdtraj,leeping/mdtraj,gph82/mdtraj,rmcgibbo/mdtraj,gph82/mdtraj,jchodera/mdtraj,mdtraj/mdtraj,leeping/mdtraj,tcmoore3/mdtraj,ctk3b/mdtraj,gph82/mdtraj,leeping/mdtraj,rmcgibbo/mdtraj,mattwthompson/mdtraj,ctk3b/mdtraj,mattwthompson/mdtraj,tcmoore3/mdtraj,mattwthompson/mdtraj,ctk3b/mdtraj,tcmoore3/mdtraj,dwhswenson/mdtraj,dwhswenson/mdtraj,leeping/mdtraj,msultan/mdtraj,dwhswenson/mdtraj | -
from mdtraj import load
from mdtraj.testing import get_fn
def test_load_single():
"""
Just check for any raised errors coming from loading a single file.
"""
load(get_fn('frame0.pdb'))
def test_load_single_list():
"""
See if a single-element list of files is successfully loaded.
"""
load([get_fn('frame0.pdb')])
def test_load_many_list():
"""
See if a multi-element list of files is successfully loaded.
"""
+ single = load(get_fn('frame0.pdb'))
- traj = load(2 * [get_fn('frame0.pdb')], discard_overlapping_frames=False)
+ double = load(2 * [get_fn('frame0.pdb')], discard_overlapping_frames=False)
- assert traj.n_frames == 2
+ assert 2 * single.n_frames == double.n_frames
| Fix test for loading multiple trajectories. | ## Code Before:
from mdtraj import load
from mdtraj.testing import get_fn
def test_load_single():
"""
Just check for any raised errors coming from loading a single file.
"""
load(get_fn('frame0.pdb'))
def test_load_single_list():
"""
See if a single-element list of files is successfully loaded.
"""
load([get_fn('frame0.pdb')])
def test_load_many_list():
"""
See if a multi-element list of files is successfully loaded.
"""
traj = load(2 * [get_fn('frame0.pdb')], discard_overlapping_frames=False)
assert traj.n_frames == 2
## Instruction:
Fix test for loading multiple trajectories.
## Code After:
from mdtraj import load
from mdtraj.testing import get_fn
def test_load_single():
"""
Just check for any raised errors coming from loading a single file.
"""
load(get_fn('frame0.pdb'))
def test_load_single_list():
"""
See if a single-element list of files is successfully loaded.
"""
load([get_fn('frame0.pdb')])
def test_load_many_list():
"""
See if a multi-element list of files is successfully loaded.
"""
single = load(get_fn('frame0.pdb'))
double = load(2 * [get_fn('frame0.pdb')], discard_overlapping_frames=False)
assert 2 * single.n_frames == double.n_frames
|
b3400070d47d95bfa2eeac3a9f696b8957d88128 | conjureup/controllers/clouds/tui.py | conjureup/controllers/clouds/tui.py | from conjureup import controllers, events, juju, utils
from conjureup.app_config import app
from conjureup.consts import cloud_types
from .common import BaseCloudController
class CloudsController(BaseCloudController):
def __controller_exists(self, controller):
return juju.get_controller(controller) is not None
def finish(self):
if app.argv.model:
app.provider.model = app.argv.model
else:
app.provider.model = utils.gen_model()
return controllers.use('credentials').render()
async def _check_lxd_compat(self):
utils.info(
"Summoning {} to {}".format(app.argv.spell, app.provider.cloud))
if app.provider.cloud_type == cloud_types.LOCALHOST:
try:
app.provider._set_lxd_dir_env()
client_compatible = await app.provider.is_client_compatible()
server_compatible = await app.provider.is_server_compatible()
if client_compatible and server_compatible:
self.finish()
else:
utils.error("LXD Server or LXC client not compatible")
events.Shutdown.set(1)
except app.provider.LocalhostError:
raise
def render(self):
app.loop.create_task(self._check_lxd_compat())
_controller_class = CloudsController
| from conjureup import controllers, events, juju, utils
from conjureup.app_config import app
from conjureup.consts import cloud_types
from .common import BaseCloudController
class CloudsController(BaseCloudController):
def __controller_exists(self, controller):
return juju.get_controller(controller) is not None
def finish(self):
if app.argv.model:
app.provider.model = app.argv.model
else:
app.provider.model = utils.gen_model()
return controllers.use('credentials').render()
async def _check_lxd_compat(self):
utils.info(
"Summoning {} to {}".format(app.argv.spell, app.provider.cloud))
if app.provider.cloud_type == cloud_types.LOCALHOST:
try:
app.provider._set_lxd_dir_env()
client_compatible = await app.provider.is_client_compatible()
server_compatible = await app.provider.is_server_compatible()
if client_compatible and server_compatible:
self.finish()
else:
utils.error("LXD Server or LXC client not compatible")
events.Shutdown.set(1)
except app.provider.LocalhostError:
raise
self.finish()
def render(self):
app.loop.create_task(self._check_lxd_compat())
_controller_class = CloudsController
| Fix issue where non localhost headless clouds werent calling finish | Fix issue where non localhost headless clouds werent calling finish
Signed-off-by: Adam Stokes <49c255c1d074742f60d19fdba5e2aa5a34add567@users.noreply.github.com>
| Python | mit | conjure-up/conjure-up,Ubuntu-Solutions-Engineering/conjure,ubuntu/conjure-up,Ubuntu-Solutions-Engineering/conjure,conjure-up/conjure-up,ubuntu/conjure-up | from conjureup import controllers, events, juju, utils
from conjureup.app_config import app
from conjureup.consts import cloud_types
from .common import BaseCloudController
class CloudsController(BaseCloudController):
def __controller_exists(self, controller):
return juju.get_controller(controller) is not None
def finish(self):
if app.argv.model:
app.provider.model = app.argv.model
else:
app.provider.model = utils.gen_model()
return controllers.use('credentials').render()
async def _check_lxd_compat(self):
utils.info(
"Summoning {} to {}".format(app.argv.spell, app.provider.cloud))
if app.provider.cloud_type == cloud_types.LOCALHOST:
try:
app.provider._set_lxd_dir_env()
client_compatible = await app.provider.is_client_compatible()
server_compatible = await app.provider.is_server_compatible()
if client_compatible and server_compatible:
self.finish()
else:
utils.error("LXD Server or LXC client not compatible")
events.Shutdown.set(1)
except app.provider.LocalhostError:
raise
+ self.finish()
def render(self):
app.loop.create_task(self._check_lxd_compat())
_controller_class = CloudsController
| Fix issue where non localhost headless clouds werent calling finish | ## Code Before:
from conjureup import controllers, events, juju, utils
from conjureup.app_config import app
from conjureup.consts import cloud_types
from .common import BaseCloudController
class CloudsController(BaseCloudController):
def __controller_exists(self, controller):
return juju.get_controller(controller) is not None
def finish(self):
if app.argv.model:
app.provider.model = app.argv.model
else:
app.provider.model = utils.gen_model()
return controllers.use('credentials').render()
async def _check_lxd_compat(self):
utils.info(
"Summoning {} to {}".format(app.argv.spell, app.provider.cloud))
if app.provider.cloud_type == cloud_types.LOCALHOST:
try:
app.provider._set_lxd_dir_env()
client_compatible = await app.provider.is_client_compatible()
server_compatible = await app.provider.is_server_compatible()
if client_compatible and server_compatible:
self.finish()
else:
utils.error("LXD Server or LXC client not compatible")
events.Shutdown.set(1)
except app.provider.LocalhostError:
raise
def render(self):
app.loop.create_task(self._check_lxd_compat())
_controller_class = CloudsController
## Instruction:
Fix issue where non localhost headless clouds werent calling finish
## Code After:
from conjureup import controllers, events, juju, utils
from conjureup.app_config import app
from conjureup.consts import cloud_types
from .common import BaseCloudController
class CloudsController(BaseCloudController):
def __controller_exists(self, controller):
return juju.get_controller(controller) is not None
def finish(self):
if app.argv.model:
app.provider.model = app.argv.model
else:
app.provider.model = utils.gen_model()
return controllers.use('credentials').render()
async def _check_lxd_compat(self):
utils.info(
"Summoning {} to {}".format(app.argv.spell, app.provider.cloud))
if app.provider.cloud_type == cloud_types.LOCALHOST:
try:
app.provider._set_lxd_dir_env()
client_compatible = await app.provider.is_client_compatible()
server_compatible = await app.provider.is_server_compatible()
if client_compatible and server_compatible:
self.finish()
else:
utils.error("LXD Server or LXC client not compatible")
events.Shutdown.set(1)
except app.provider.LocalhostError:
raise
self.finish()
def render(self):
app.loop.create_task(self._check_lxd_compat())
_controller_class = CloudsController
|
b123001ea0d4fb475184727c39eafd5b46cc0964 | shopit_app/urls.py | shopit_app/urls.py | from django.conf import settings
from django.conf.urls import include, patterns, url
from rest_framework_nested import routers
from shopit_app.views import IndexView
from authentication_app.views import AccountViewSet, LoginView
router = routers.SimpleRouter()
router.register(r'accounts', AccountViewSet)
urlpatterns = patterns('',
# API endpoints
url(r'^api/v1/', include(router.urls)),
url(r'^api/v1/auth/login/$', LoginView.as_view(), name='login'),
url('^.*$', IndexView.as_view(), name='index'),
)
| from django.conf import settings
from django.conf.urls import include, patterns, url
from rest_framework_nested import routers
from shopit_app.views import IndexView
from authentication_app.views import AccountViewSet, LoginView, LogoutView
router = routers.SimpleRouter()
router.register(r'accounts', AccountViewSet)
urlpatterns = patterns('',
# API endpoints
url(r'^api/v1/', include(router.urls)),
url(r'^api/v1/auth/logout/$', LogoutView.as_view(), name='logout'),
url(r'^api/v1/auth/login/$', LoginView.as_view(), name='login'),
url('^.*$', IndexView.as_view(), name='index'),
)
| Add the endpoint for the logout. | Add the endpoint for the logout.
| Python | mit | mvpgomes/shopit-app,mvpgomes/shopit-app,mvpgomes/shopit-app,mvpgomes/shopit-app | from django.conf import settings
from django.conf.urls import include, patterns, url
from rest_framework_nested import routers
from shopit_app.views import IndexView
- from authentication_app.views import AccountViewSet, LoginView
+ from authentication_app.views import AccountViewSet, LoginView, LogoutView
router = routers.SimpleRouter()
router.register(r'accounts', AccountViewSet)
urlpatterns = patterns('',
# API endpoints
url(r'^api/v1/', include(router.urls)),
+ url(r'^api/v1/auth/logout/$', LogoutView.as_view(), name='logout'),
url(r'^api/v1/auth/login/$', LoginView.as_view(), name='login'),
url('^.*$', IndexView.as_view(), name='index'),
)
| Add the endpoint for the logout. | ## Code Before:
from django.conf import settings
from django.conf.urls import include, patterns, url
from rest_framework_nested import routers
from shopit_app.views import IndexView
from authentication_app.views import AccountViewSet, LoginView
router = routers.SimpleRouter()
router.register(r'accounts', AccountViewSet)
urlpatterns = patterns('',
# API endpoints
url(r'^api/v1/', include(router.urls)),
url(r'^api/v1/auth/login/$', LoginView.as_view(), name='login'),
url('^.*$', IndexView.as_view(), name='index'),
)
## Instruction:
Add the endpoint for the logout.
## Code After:
from django.conf import settings
from django.conf.urls import include, patterns, url
from rest_framework_nested import routers
from shopit_app.views import IndexView
from authentication_app.views import AccountViewSet, LoginView, LogoutView
router = routers.SimpleRouter()
router.register(r'accounts', AccountViewSet)
urlpatterns = patterns('',
# API endpoints
url(r'^api/v1/', include(router.urls)),
url(r'^api/v1/auth/logout/$', LogoutView.as_view(), name='logout'),
url(r'^api/v1/auth/login/$', LoginView.as_view(), name='login'),
url('^.*$', IndexView.as_view(), name='index'),
)
|
0d7921b4dcf5e3b511fdb54fc30ebc0547b14d47 | django_dzenlog/urls.py | django_dzenlog/urls.py | from django.conf.urls.defaults import *
from models import GeneralPost
from feeds import LatestPosts
post_list = {
'queryset': GeneralPost.objects.all(),
}
feeds = {
'all': LatestPosts,
}
urlpatterns = patterns('django.views.generic',
(r'^(?P<slug>[a-z0-9-]+)/$', 'list_detail.object_detail', post_list, 'dzenlog-post-details'),
(r'^$', 'list_detail.object_list', post_list, 'dzenlog-post-list'),
)
urlpatterns += patterns('django.contrib.syndication.views',
(r'^rss/(?P<url>.*)/$', 'feed', {'feed_dict': feeds}, 'dzenlog-feeds'),
)
| from django.conf.urls.defaults import *
from models import GeneralPost
from feeds import latest
post_list = {
'queryset': GeneralPost.objects.all(),
}
feeds = {
'all': latest(GeneralPost, 'dzenlog-post-list'),
}
urlpatterns = patterns('django.views.generic',
(r'^(?P<slug>[a-z0-9-]+)/$', 'list_detail.object_detail', post_list, 'dzenlog-post-details'),
(r'^$', 'list_detail.object_list', post_list, 'dzenlog-post-list'),
)
urlpatterns += patterns('django.contrib.syndication.views',
(r'^rss/(?P<url>.*)/$', 'feed', {'feed_dict': feeds}, 'dzenlog-feeds'),
)
| Use 'latest' to generate feed for GeneralPost. | Use 'latest' to generate feed for GeneralPost.
| Python | bsd-3-clause | svetlyak40wt/django-dzenlog | from django.conf.urls.defaults import *
from models import GeneralPost
- from feeds import LatestPosts
+ from feeds import latest
post_list = {
'queryset': GeneralPost.objects.all(),
}
feeds = {
- 'all': LatestPosts,
+ 'all': latest(GeneralPost, 'dzenlog-post-list'),
}
urlpatterns = patterns('django.views.generic',
(r'^(?P<slug>[a-z0-9-]+)/$', 'list_detail.object_detail', post_list, 'dzenlog-post-details'),
(r'^$', 'list_detail.object_list', post_list, 'dzenlog-post-list'),
)
urlpatterns += patterns('django.contrib.syndication.views',
(r'^rss/(?P<url>.*)/$', 'feed', {'feed_dict': feeds}, 'dzenlog-feeds'),
)
| Use 'latest' to generate feed for GeneralPost. | ## Code Before:
from django.conf.urls.defaults import *
from models import GeneralPost
from feeds import LatestPosts
post_list = {
'queryset': GeneralPost.objects.all(),
}
feeds = {
'all': LatestPosts,
}
urlpatterns = patterns('django.views.generic',
(r'^(?P<slug>[a-z0-9-]+)/$', 'list_detail.object_detail', post_list, 'dzenlog-post-details'),
(r'^$', 'list_detail.object_list', post_list, 'dzenlog-post-list'),
)
urlpatterns += patterns('django.contrib.syndication.views',
(r'^rss/(?P<url>.*)/$', 'feed', {'feed_dict': feeds}, 'dzenlog-feeds'),
)
## Instruction:
Use 'latest' to generate feed for GeneralPost.
## Code After:
from django.conf.urls.defaults import *
from models import GeneralPost
from feeds import latest
post_list = {
'queryset': GeneralPost.objects.all(),
}
feeds = {
'all': latest(GeneralPost, 'dzenlog-post-list'),
}
urlpatterns = patterns('django.views.generic',
(r'^(?P<slug>[a-z0-9-]+)/$', 'list_detail.object_detail', post_list, 'dzenlog-post-details'),
(r'^$', 'list_detail.object_list', post_list, 'dzenlog-post-list'),
)
urlpatterns += patterns('django.contrib.syndication.views',
(r'^rss/(?P<url>.*)/$', 'feed', {'feed_dict': feeds}, 'dzenlog-feeds'),
)
|
cab0f9ea3471cf88dd03da7a243ae55579b44b65 | client.py | client.py |
import RPi.GPIO as io
import requests
import sys
class Switch(object):
def __init__(self, **kwargs):
self.pin = kwargs["pin"]
io.setup(self.pin, io.IN)
@property
def is_on(self):
return io.input(self.pin)
PINS = (8, 16, 18)
switches = set()
def has_free():
global switches
return not all([s.is_on for s in switches])
def call_api(is_on):
r = requests.post("SERVER_ADDRESS",
params={"is_free": "yes" if is_on else "no"})
if __name__ == "__main__":
io.setmode(io.BOARD)
for pin in PINS:
switches.add(Switch(pin=pin))
try:
previous_state = has_free()
while True:
state = has_free()
if state is not previous_state:
call_api(state)
previous_state = state
except KeyboardInterrupt:
pass
|
import RPi.GPIO as io
import sys
class Switch(object):
def __init__(self, **kwargs):
self.pin = kwargs["pin"]
io.setup(self.pin, io.IN)
@property
def is_on(self):
return io.input(self.pin)
PINS = (8, 16, 18)
server_url = sys.argv[1]
switches = set()
def has_free():
global switches
return not all([s.is_on for s in switches])
def call_api(url, is_on):
r = requests.post(url, params={"is_free": "yes" if is_on else "no"})
if __name__ == "__main__":
io.setmode(io.BOARD)
for pin in PINS:
switches.add(Switch(pin=pin))
try:
previous_state = has_free()
while True:
state = has_free()
if state is not previous_state:
call_api(server_url, state)
previous_state = state
except KeyboardInterrupt:
pass
| Set server URL with command line argument | Set server URL with command line argument
| Python | mit | madebymany/isthetoiletfree |
import RPi.GPIO as io
- import requests
import sys
class Switch(object):
def __init__(self, **kwargs):
self.pin = kwargs["pin"]
io.setup(self.pin, io.IN)
@property
def is_on(self):
return io.input(self.pin)
PINS = (8, 16, 18)
+ server_url = sys.argv[1]
switches = set()
def has_free():
global switches
return not all([s.is_on for s in switches])
- def call_api(is_on):
+ def call_api(url, is_on):
+ r = requests.post(url, params={"is_free": "yes" if is_on else "no"})
- r = requests.post("SERVER_ADDRESS",
- params={"is_free": "yes" if is_on else "no"})
if __name__ == "__main__":
io.setmode(io.BOARD)
for pin in PINS:
switches.add(Switch(pin=pin))
try:
previous_state = has_free()
while True:
state = has_free()
if state is not previous_state:
- call_api(state)
+ call_api(server_url, state)
previous_state = state
except KeyboardInterrupt:
pass
| Set server URL with command line argument | ## Code Before:
import RPi.GPIO as io
import requests
import sys
class Switch(object):
def __init__(self, **kwargs):
self.pin = kwargs["pin"]
io.setup(self.pin, io.IN)
@property
def is_on(self):
return io.input(self.pin)
PINS = (8, 16, 18)
switches = set()
def has_free():
global switches
return not all([s.is_on for s in switches])
def call_api(is_on):
r = requests.post("SERVER_ADDRESS",
params={"is_free": "yes" if is_on else "no"})
if __name__ == "__main__":
io.setmode(io.BOARD)
for pin in PINS:
switches.add(Switch(pin=pin))
try:
previous_state = has_free()
while True:
state = has_free()
if state is not previous_state:
call_api(state)
previous_state = state
except KeyboardInterrupt:
pass
## Instruction:
Set server URL with command line argument
## Code After:
import RPi.GPIO as io
import sys
class Switch(object):
def __init__(self, **kwargs):
self.pin = kwargs["pin"]
io.setup(self.pin, io.IN)
@property
def is_on(self):
return io.input(self.pin)
PINS = (8, 16, 18)
server_url = sys.argv[1]
switches = set()
def has_free():
global switches
return not all([s.is_on for s in switches])
def call_api(url, is_on):
r = requests.post(url, params={"is_free": "yes" if is_on else "no"})
if __name__ == "__main__":
io.setmode(io.BOARD)
for pin in PINS:
switches.add(Switch(pin=pin))
try:
previous_state = has_free()
while True:
state = has_free()
if state is not previous_state:
call_api(server_url, state)
previous_state = state
except KeyboardInterrupt:
pass
|
384beaa77e2eaad642ec7f764acd09c2c3e04350 | res_company.py | res_company.py | from openerp.osv import osv, fields
from openerp.tools.translate import _
class res_company(osv.Model):
_inherit = "res.company"
_columns = {
'remittance_letter_top': fields.text(
_('Remittance Letter - top message'),
help=_('Message to write at the top of Remittance Letter '
'reports. Available variables: "$iban" for the IBAN; "$date" for '
'the payment date. HTML tags are allowed.')
),
'remittance_letter_bottom': fields.text(
_('Remittance Letter - bottom message'),
help=_('Message to write at the bottom of Remittance Letter '
'reports. HTML tags are allowed.')
),
}
| from openerp.osv import osv, fields
from openerp.tools.translate import _
class res_company(osv.Model):
_inherit = "res.company"
_columns = {
'remittance_letter_top': fields.text(
_('Remittance Letter - top message'),
help=_('Message to write at the top of Remittance Letter '
'reports. Available variables: "$iban" for the IBAN; "$date" for '
'the payment date. HTML tags are allowed.'),
translate=True),
'remittance_letter_bottom': fields.text(
_('Remittance Letter - bottom message'),
help=_('Message to write at the bottom of Remittance Letter '
'reports. HTML tags are allowed.'),
translate=True),
}
| Make Remittance Letter config messages translatable | Make Remittance Letter config messages translatable
| Python | agpl-3.0 | xcgd/account_streamline | from openerp.osv import osv, fields
from openerp.tools.translate import _
class res_company(osv.Model):
_inherit = "res.company"
_columns = {
'remittance_letter_top': fields.text(
_('Remittance Letter - top message'),
help=_('Message to write at the top of Remittance Letter '
'reports. Available variables: "$iban" for the IBAN; "$date" for '
- 'the payment date. HTML tags are allowed.')
+ 'the payment date. HTML tags are allowed.'),
- ),
+ translate=True),
'remittance_letter_bottom': fields.text(
_('Remittance Letter - bottom message'),
help=_('Message to write at the bottom of Remittance Letter '
- 'reports. HTML tags are allowed.')
+ 'reports. HTML tags are allowed.'),
- ),
+ translate=True),
}
| Make Remittance Letter config messages translatable | ## Code Before:
from openerp.osv import osv, fields
from openerp.tools.translate import _
class res_company(osv.Model):
_inherit = "res.company"
_columns = {
'remittance_letter_top': fields.text(
_('Remittance Letter - top message'),
help=_('Message to write at the top of Remittance Letter '
'reports. Available variables: "$iban" for the IBAN; "$date" for '
'the payment date. HTML tags are allowed.')
),
'remittance_letter_bottom': fields.text(
_('Remittance Letter - bottom message'),
help=_('Message to write at the bottom of Remittance Letter '
'reports. HTML tags are allowed.')
),
}
## Instruction:
Make Remittance Letter config messages translatable
## Code After:
from openerp.osv import osv, fields
from openerp.tools.translate import _
class res_company(osv.Model):
_inherit = "res.company"
_columns = {
'remittance_letter_top': fields.text(
_('Remittance Letter - top message'),
help=_('Message to write at the top of Remittance Letter '
'reports. Available variables: "$iban" for the IBAN; "$date" for '
'the payment date. HTML tags are allowed.'),
translate=True),
'remittance_letter_bottom': fields.text(
_('Remittance Letter - bottom message'),
help=_('Message to write at the bottom of Remittance Letter '
'reports. HTML tags are allowed.'),
translate=True),
}
|
813dd27a2057d2e32726ff6b43ab8ca1411303c7 | fabfile.py | fabfile.py | from fabric.api import *
from fabric.colors import *
env.colorize_errors = True
env.hosts = ['sanaprotocolbuilder.me']
env.user = 'root'
env.project_root = '/opt/sana.protocol_builder'
def prepare_deploy():
local('python sana_builder/manage.py syncdb')
local('python sana_builder/manage.py test')
local('git push')
def deploy():
with cd(env.project_root), prefix('workon sana_protocol_builder'):
print(green('Pulling latest revision...'))
run('git pull')
print(green('Installing dependencies...'))
run('pip install -qr requirements.txt')
print(green('Migrating database...'))
run('python sana_builder/manage.py syncdb')
print(green('Restarting gunicorn...'))
run('supervisorctl restart gunicorn')
| from fabric.api import *
from fabric.colors import *
env.colorize_errors = True
env.hosts = ['sanaprotocolbuilder.me']
env.user = 'root'
env.project_root = '/opt/sana.protocol_builder'
def prepare_deploy():
local('python sana_builder/manage.py syncdb')
local('python sana_builder/manage.py test')
local('git push')
def deploy():
prepare_deploy()
with cd(env.project_root), prefix('workon sana_protocol_builder'):
print(green('Pulling latest revision...'))
run('git pull')
print(green('Installing dependencies...'))
run('pip install -qr requirements.txt')
print(green('Migrating database...'))
run('python sana_builder/manage.py syncdb')
print(green('Restarting gunicorn...'))
run('supervisorctl restart gunicorn')
| Prepare for deploy in deploy script. | Prepare for deploy in deploy script.
| Python | bsd-3-clause | SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder,SanaMobile/sana.protocol_builder | from fabric.api import *
from fabric.colors import *
env.colorize_errors = True
env.hosts = ['sanaprotocolbuilder.me']
env.user = 'root'
env.project_root = '/opt/sana.protocol_builder'
def prepare_deploy():
local('python sana_builder/manage.py syncdb')
local('python sana_builder/manage.py test')
local('git push')
def deploy():
+ prepare_deploy()
+
with cd(env.project_root), prefix('workon sana_protocol_builder'):
print(green('Pulling latest revision...'))
run('git pull')
print(green('Installing dependencies...'))
run('pip install -qr requirements.txt')
print(green('Migrating database...'))
run('python sana_builder/manage.py syncdb')
print(green('Restarting gunicorn...'))
run('supervisorctl restart gunicorn')
| Prepare for deploy in deploy script. | ## Code Before:
from fabric.api import *
from fabric.colors import *
env.colorize_errors = True
env.hosts = ['sanaprotocolbuilder.me']
env.user = 'root'
env.project_root = '/opt/sana.protocol_builder'
def prepare_deploy():
local('python sana_builder/manage.py syncdb')
local('python sana_builder/manage.py test')
local('git push')
def deploy():
with cd(env.project_root), prefix('workon sana_protocol_builder'):
print(green('Pulling latest revision...'))
run('git pull')
print(green('Installing dependencies...'))
run('pip install -qr requirements.txt')
print(green('Migrating database...'))
run('python sana_builder/manage.py syncdb')
print(green('Restarting gunicorn...'))
run('supervisorctl restart gunicorn')
## Instruction:
Prepare for deploy in deploy script.
## Code After:
from fabric.api import *
from fabric.colors import *
env.colorize_errors = True
env.hosts = ['sanaprotocolbuilder.me']
env.user = 'root'
env.project_root = '/opt/sana.protocol_builder'
def prepare_deploy():
local('python sana_builder/manage.py syncdb')
local('python sana_builder/manage.py test')
local('git push')
def deploy():
prepare_deploy()
with cd(env.project_root), prefix('workon sana_protocol_builder'):
print(green('Pulling latest revision...'))
run('git pull')
print(green('Installing dependencies...'))
run('pip install -qr requirements.txt')
print(green('Migrating database...'))
run('python sana_builder/manage.py syncdb')
print(green('Restarting gunicorn...'))
run('supervisorctl restart gunicorn')
|
0ef346389b680e81ab618d4d782239640c1926f5 | tests/test_collection.py | tests/test_collection.py | import unittest
from indigo.models import Collection
from indigo.models.errors import UniqueException
from nose.tools import raises
class NodeTest(unittest.TestCase):
def test_a_create_root(self):
Collection.create(name="test_root", parent=None, path="/")
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.path == '/'
assert coll.parent is None
# Make sure this is the root collection
root = Collection.get_root_collection()
assert root.id == coll.id
def test_create_with_children(self):
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.is_root
child1 = Collection.create(name="child1", parent=str(coll.id))
child2 = Collection.create(name="child2", parent=str(coll.id))
assert child1.get_parent_collection().id == coll.id
assert child2.get_parent_collection().id == coll.id
assert child1.path == '/child1/'
assert child2.path == '/child2/'
children = coll.get_child_collections()
assert len(children) == 2
assert coll.get_child_collection_count() == 2
assert str(children[0].id) == str(child1.id)
assert str(children[1].id) == str(child2.id) | import unittest
from indigo.models import Collection
from indigo.models.errors import UniqueException
from nose.tools import raises
class NodeTest(unittest.TestCase):
def test_a_create_root(self):
Collection.create(name="test_root", parent=None, path="/")
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.path == '/'
assert coll.parent is None
# Make sure this is the root collection
root = Collection.get_root_collection()
assert root.id == coll.id
def test_create_with_children(self):
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.is_root
child1 = Collection.create(name="child1", parent=str(coll.id))
child2 = Collection.create(name="child2", parent=str(coll.id))
assert child1.get_parent_collection().id == coll.id
assert child2.get_parent_collection().id == coll.id
assert child1.path == '/child1/'
assert child2.path == '/child2/'
children = coll.get_child_collections()
assert len(children) == 2
assert coll.get_child_collection_count() == 2
| Remove unnecessary test of collection children | Remove unnecessary test of collection children
| Python | agpl-3.0 | UMD-DRASTIC/drastic | import unittest
from indigo.models import Collection
from indigo.models.errors import UniqueException
from nose.tools import raises
class NodeTest(unittest.TestCase):
def test_a_create_root(self):
Collection.create(name="test_root", parent=None, path="/")
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.path == '/'
assert coll.parent is None
# Make sure this is the root collection
root = Collection.get_root_collection()
assert root.id == coll.id
def test_create_with_children(self):
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.is_root
child1 = Collection.create(name="child1", parent=str(coll.id))
child2 = Collection.create(name="child2", parent=str(coll.id))
assert child1.get_parent_collection().id == coll.id
assert child2.get_parent_collection().id == coll.id
assert child1.path == '/child1/'
assert child2.path == '/child2/'
children = coll.get_child_collections()
assert len(children) == 2
assert coll.get_child_collection_count() == 2
+
- assert str(children[0].id) == str(child1.id)
- assert str(children[1].id) == str(child2.id) | Remove unnecessary test of collection children | ## Code Before:
import unittest
from indigo.models import Collection
from indigo.models.errors import UniqueException
from nose.tools import raises
class NodeTest(unittest.TestCase):
def test_a_create_root(self):
Collection.create(name="test_root", parent=None, path="/")
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.path == '/'
assert coll.parent is None
# Make sure this is the root collection
root = Collection.get_root_collection()
assert root.id == coll.id
def test_create_with_children(self):
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.is_root
child1 = Collection.create(name="child1", parent=str(coll.id))
child2 = Collection.create(name="child2", parent=str(coll.id))
assert child1.get_parent_collection().id == coll.id
assert child2.get_parent_collection().id == coll.id
assert child1.path == '/child1/'
assert child2.path == '/child2/'
children = coll.get_child_collections()
assert len(children) == 2
assert coll.get_child_collection_count() == 2
assert str(children[0].id) == str(child1.id)
assert str(children[1].id) == str(child2.id)
## Instruction:
Remove unnecessary test of collection children
## Code After:
import unittest
from indigo.models import Collection
from indigo.models.errors import UniqueException
from nose.tools import raises
class NodeTest(unittest.TestCase):
def test_a_create_root(self):
Collection.create(name="test_root", parent=None, path="/")
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.path == '/'
assert coll.parent is None
# Make sure this is the root collection
root = Collection.get_root_collection()
assert root.id == coll.id
def test_create_with_children(self):
coll = Collection.find("test_root")
assert coll.name == "test_root"
assert coll.is_root
child1 = Collection.create(name="child1", parent=str(coll.id))
child2 = Collection.create(name="child2", parent=str(coll.id))
assert child1.get_parent_collection().id == coll.id
assert child2.get_parent_collection().id == coll.id
assert child1.path == '/child1/'
assert child2.path == '/child2/'
children = coll.get_child_collections()
assert len(children) == 2
assert coll.get_child_collection_count() == 2
|
6858e4a2e2047c906a3b8f69b7cd7b04a0cbf666 | pivoteer/writer/censys.py | pivoteer/writer/censys.py |
from pivoteer.writer.core import CsvWriter
class CensysCsvWriter(CsvWriter):
"""
A CsvWriter implementation for IndicatorRecords with a record type of "CE" (Censys Record)
"""
def __init__(self, writer):
"""
Create a new CsvWriter for Censys Records using the given writer.
:param writer: The writer
"""
super(CensysCsvWriter, self).__init__(writer)
def create_title_rows(self, indicator, records):
yield ["Certificate Search Results"]
def create_header(self):
return ["Subject", "Issuer", "SHA256", "Validity Start", "Validity End"]
def create_rows(self, record):
info = record["info"]
records = info["records"]
for record in records:
parsed = record["parsed"]
subject = parsed["subject_dn"]
issuer = parsed["issuer_dn"]
sha256 = parsed["fingerprint_sha256"]
validity = parsed["validity"]
start = validity["start"]
end = validity["end"]
yield [subject, issuer, sha256, start, end]
|
from pivoteer.writer.core import CsvWriter
class CensysCsvWriter(CsvWriter):
"""
A CsvWriter implementation for IndicatorRecords with a record type of "CE" (Censys Record)
"""
def __init__(self, writer):
"""
Create a new CsvWriter for Censys Records using the given writer.
:param writer: The writer
"""
super(CensysCsvWriter, self).__init__(writer)
def create_title_rows(self, indicator, records):
yield ["Certificate Search Results"]
def create_header(self):
return ["Subject", "Issuer", "SHA256", "Validity Start", "Validity End"]
def create_rows(self, record):
if (record is not None and len(record) > 0):
info = record["info"]
records = info["records"]
for record in records:
parsed = record["parsed"]
subject = parsed["subject_dn"]
issuer = parsed["issuer_dn"]
sha256 = parsed["fingerprint_sha256"]
validity = parsed["validity"]
start = validity["start"]
end = validity["end"]
yield [subject, issuer, sha256, start, end]
| Resolve issues with exporting empty dataset for certificate list | Resolve issues with exporting empty dataset for certificate list
| Python | mit | gdit-cnd/RAPID,gdit-cnd/RAPID,LindaTNguyen/RAPID,gdit-cnd/RAPID,gdit-cnd/RAPID,LindaTNguyen/RAPID,gdit-cnd/RAPID,LindaTNguyen/RAPID,LindaTNguyen/RAPID,LindaTNguyen/RAPID |
from pivoteer.writer.core import CsvWriter
class CensysCsvWriter(CsvWriter):
"""
A CsvWriter implementation for IndicatorRecords with a record type of "CE" (Censys Record)
"""
def __init__(self, writer):
"""
Create a new CsvWriter for Censys Records using the given writer.
:param writer: The writer
"""
super(CensysCsvWriter, self).__init__(writer)
+
def create_title_rows(self, indicator, records):
yield ["Certificate Search Results"]
def create_header(self):
return ["Subject", "Issuer", "SHA256", "Validity Start", "Validity End"]
def create_rows(self, record):
+ if (record is not None and len(record) > 0):
- info = record["info"]
+ info = record["info"]
- records = info["records"]
+ records = info["records"]
- for record in records:
+ for record in records:
- parsed = record["parsed"]
+ parsed = record["parsed"]
- subject = parsed["subject_dn"]
+ subject = parsed["subject_dn"]
- issuer = parsed["issuer_dn"]
+ issuer = parsed["issuer_dn"]
- sha256 = parsed["fingerprint_sha256"]
+ sha256 = parsed["fingerprint_sha256"]
- validity = parsed["validity"]
+ validity = parsed["validity"]
- start = validity["start"]
+ start = validity["start"]
- end = validity["end"]
+ end = validity["end"]
- yield [subject, issuer, sha256, start, end]
+ yield [subject, issuer, sha256, start, end]
| Resolve issues with exporting empty dataset for certificate list | ## Code Before:
from pivoteer.writer.core import CsvWriter
class CensysCsvWriter(CsvWriter):
"""
A CsvWriter implementation for IndicatorRecords with a record type of "CE" (Censys Record)
"""
def __init__(self, writer):
"""
Create a new CsvWriter for Censys Records using the given writer.
:param writer: The writer
"""
super(CensysCsvWriter, self).__init__(writer)
def create_title_rows(self, indicator, records):
yield ["Certificate Search Results"]
def create_header(self):
return ["Subject", "Issuer", "SHA256", "Validity Start", "Validity End"]
def create_rows(self, record):
info = record["info"]
records = info["records"]
for record in records:
parsed = record["parsed"]
subject = parsed["subject_dn"]
issuer = parsed["issuer_dn"]
sha256 = parsed["fingerprint_sha256"]
validity = parsed["validity"]
start = validity["start"]
end = validity["end"]
yield [subject, issuer, sha256, start, end]
## Instruction:
Resolve issues with exporting empty dataset for certificate list
## Code After:
from pivoteer.writer.core import CsvWriter
class CensysCsvWriter(CsvWriter):
"""
A CsvWriter implementation for IndicatorRecords with a record type of "CE" (Censys Record)
"""
def __init__(self, writer):
"""
Create a new CsvWriter for Censys Records using the given writer.
:param writer: The writer
"""
super(CensysCsvWriter, self).__init__(writer)
def create_title_rows(self, indicator, records):
yield ["Certificate Search Results"]
def create_header(self):
return ["Subject", "Issuer", "SHA256", "Validity Start", "Validity End"]
def create_rows(self, record):
if (record is not None and len(record) > 0):
info = record["info"]
records = info["records"]
for record in records:
parsed = record["parsed"]
subject = parsed["subject_dn"]
issuer = parsed["issuer_dn"]
sha256 = parsed["fingerprint_sha256"]
validity = parsed["validity"]
start = validity["start"]
end = validity["end"]
yield [subject, issuer, sha256, start, end]
|
87d1801fefcd048f60c944c28bfc005101c5704b | dynd/tests/test_nd_groupby.py | dynd/tests/test_nd_groupby.py | import sys
import unittest
from dynd import nd, ndt
class TestGroupBy(unittest.TestCase):
def test_immutable(self):
a = nd.array([
('x', 0),
('y', 1),
('x', 2),
('x', 3),
('y', 4)],
dtype='{A: string; B: int32}').eval_immutable()
gb = nd.groupby(a, nd.fields(a, 'A'))
self.assertEqual(nd.as_py(gb.groups), [{'A': 'x'}, {'A': 'y'}])
self.assertEqual(nd.as_py(gb), [
[{'A': 'x', 'B': 0},
{'A': 'x', 'B': 2},
{'A': 'x', 'B': 3}],
[{'A': 'y', 'B': 1},
{'A': 'y', 'B': 4}]])
if __name__ == '__main__':
unittest.main()
| import sys
import unittest
from dynd import nd, ndt
class TestGroupBy(unittest.TestCase):
def test_immutable(self):
a = nd.array([
('x', 0),
('y', 1),
('x', 2),
('x', 3),
('y', 4)],
dtype='{A: string; B: int32}').eval_immutable()
gb = nd.groupby(a, nd.fields(a, 'A'))
self.assertEqual(nd.as_py(gb.groups), [{'A': 'x'}, {'A': 'y'}])
self.assertEqual(nd.as_py(gb), [
[{'A': 'x', 'B': 0},
{'A': 'x', 'B': 2},
{'A': 'x', 'B': 3}],
[{'A': 'y', 'B': 1},
{'A': 'y', 'B': 4}]])
def test_grouped_slices(self):
a = nd.asarray([[1, 2, 3], [1, 4, 5]])
gb = nd.groupby(a[:, 1:], a[:, 0])
self.assertEqual(nd.as_py(gb.groups), [1])
self.assertEqual(nd.as_py(gb), [[[2, 3], [4, 5]]])
a = nd.asarray([[1, 2, 3], [3, 1, 7], [1, 4, 5], [2, 6, 7], [3, 2, 5]])
gb = nd.groupby(a[:, 1:], a[:, 0])
self.assertEqual(nd.as_py(gb.groups), [1, 2, 3])
self.assertEqual(nd.as_py(gb), [[[2, 3], [4, 5]],
[[6, 7]],
[[1, 7], [2, 5]]])
if __name__ == '__main__':
unittest.main()
| Add some more simple nd.groupby tests | Add some more simple nd.groupby tests
| Python | bsd-2-clause | cpcloud/dynd-python,izaid/dynd-python,aterrel/dynd-python,michaelpacer/dynd-python,mwiebe/dynd-python,insertinterestingnamehere/dynd-python,aterrel/dynd-python,cpcloud/dynd-python,cpcloud/dynd-python,izaid/dynd-python,aterrel/dynd-python,mwiebe/dynd-python,michaelpacer/dynd-python,mwiebe/dynd-python,pombredanne/dynd-python,michaelpacer/dynd-python,insertinterestingnamehere/dynd-python,ContinuumIO/dynd-python,insertinterestingnamehere/dynd-python,pombredanne/dynd-python,ContinuumIO/dynd-python,izaid/dynd-python,ContinuumIO/dynd-python,cpcloud/dynd-python,izaid/dynd-python,michaelpacer/dynd-python,mwiebe/dynd-python,pombredanne/dynd-python,aterrel/dynd-python,pombredanne/dynd-python,ContinuumIO/dynd-python,insertinterestingnamehere/dynd-python | import sys
import unittest
from dynd import nd, ndt
class TestGroupBy(unittest.TestCase):
def test_immutable(self):
a = nd.array([
('x', 0),
('y', 1),
('x', 2),
('x', 3),
('y', 4)],
dtype='{A: string; B: int32}').eval_immutable()
gb = nd.groupby(a, nd.fields(a, 'A'))
self.assertEqual(nd.as_py(gb.groups), [{'A': 'x'}, {'A': 'y'}])
self.assertEqual(nd.as_py(gb), [
[{'A': 'x', 'B': 0},
{'A': 'x', 'B': 2},
{'A': 'x', 'B': 3}],
[{'A': 'y', 'B': 1},
{'A': 'y', 'B': 4}]])
+ def test_grouped_slices(self):
+ a = nd.asarray([[1, 2, 3], [1, 4, 5]])
+ gb = nd.groupby(a[:, 1:], a[:, 0])
+ self.assertEqual(nd.as_py(gb.groups), [1])
+ self.assertEqual(nd.as_py(gb), [[[2, 3], [4, 5]]])
+
+ a = nd.asarray([[1, 2, 3], [3, 1, 7], [1, 4, 5], [2, 6, 7], [3, 2, 5]])
+ gb = nd.groupby(a[:, 1:], a[:, 0])
+ self.assertEqual(nd.as_py(gb.groups), [1, 2, 3])
+ self.assertEqual(nd.as_py(gb), [[[2, 3], [4, 5]],
+ [[6, 7]],
+ [[1, 7], [2, 5]]])
+
if __name__ == '__main__':
unittest.main()
| Add some more simple nd.groupby tests | ## Code Before:
import sys
import unittest
from dynd import nd, ndt
class TestGroupBy(unittest.TestCase):
def test_immutable(self):
a = nd.array([
('x', 0),
('y', 1),
('x', 2),
('x', 3),
('y', 4)],
dtype='{A: string; B: int32}').eval_immutable()
gb = nd.groupby(a, nd.fields(a, 'A'))
self.assertEqual(nd.as_py(gb.groups), [{'A': 'x'}, {'A': 'y'}])
self.assertEqual(nd.as_py(gb), [
[{'A': 'x', 'B': 0},
{'A': 'x', 'B': 2},
{'A': 'x', 'B': 3}],
[{'A': 'y', 'B': 1},
{'A': 'y', 'B': 4}]])
if __name__ == '__main__':
unittest.main()
## Instruction:
Add some more simple nd.groupby tests
## Code After:
import sys
import unittest
from dynd import nd, ndt
class TestGroupBy(unittest.TestCase):
def test_immutable(self):
a = nd.array([
('x', 0),
('y', 1),
('x', 2),
('x', 3),
('y', 4)],
dtype='{A: string; B: int32}').eval_immutable()
gb = nd.groupby(a, nd.fields(a, 'A'))
self.assertEqual(nd.as_py(gb.groups), [{'A': 'x'}, {'A': 'y'}])
self.assertEqual(nd.as_py(gb), [
[{'A': 'x', 'B': 0},
{'A': 'x', 'B': 2},
{'A': 'x', 'B': 3}],
[{'A': 'y', 'B': 1},
{'A': 'y', 'B': 4}]])
def test_grouped_slices(self):
a = nd.asarray([[1, 2, 3], [1, 4, 5]])
gb = nd.groupby(a[:, 1:], a[:, 0])
self.assertEqual(nd.as_py(gb.groups), [1])
self.assertEqual(nd.as_py(gb), [[[2, 3], [4, 5]]])
a = nd.asarray([[1, 2, 3], [3, 1, 7], [1, 4, 5], [2, 6, 7], [3, 2, 5]])
gb = nd.groupby(a[:, 1:], a[:, 0])
self.assertEqual(nd.as_py(gb.groups), [1, 2, 3])
self.assertEqual(nd.as_py(gb), [[[2, 3], [4, 5]],
[[6, 7]],
[[1, 7], [2, 5]]])
if __name__ == '__main__':
unittest.main()
|
7b935b23e17ef873a060fdfbefbfdf232fe8b8de | git_release/release.py | git_release/release.py | import subprocess
from git_release import errors, git_helpers
def _parse_tag(tag):
major, minor = tag.split('.')
return int(major), int(minor)
def _increment_tag(tag, release_type):
major, minor = _parse_tag(tag)
if release_type == 'major':
new_major = major + 1
new_minor = 0
else:
new_major = major
new_minor = minor + 1
return '{}.{}'.format(new_major, new_minor)
def release(release_type, signed):
if not git_helpers.is_master():
raise errors.NotMasterException("Current branch is not master.\nAborting.")
tag = git_helpers.get_current_tag()
if not tag:
raise errors.NoTagException("Unable to get current tag.\nAborting.")
new_tag = _increment_tag(tag)
git_helpers.tag(signed, new_tag)
| import subprocess
from git_release import errors, git_helpers
def _parse_tag(tag):
major, minor = tag.split('.')
return int(major), int(minor)
def _increment_tag(tag, release_type):
major, minor = _parse_tag(tag)
if release_type == 'major':
new_major = major + 1
new_minor = 0
else:
new_major = major
new_minor = minor + 1
return '{}.{}'.format(new_major, new_minor)
def release(release_type, signed):
if not git_helpers.is_master():
raise errors.NotMasterException("Current branch is not master.\nAborting.")
tag = git_helpers.get_current_tag()
if not tag:
raise errors.NoTagException("Unable to get current tag.\nAborting.")
new_tag = _increment_tag(tag, release_type)
git_helpers.tag(signed, new_tag)
| Add missing argument to _increment_tag call | Add missing argument to _increment_tag call
| Python | mit | Authentise/git-release | import subprocess
from git_release import errors, git_helpers
def _parse_tag(tag):
major, minor = tag.split('.')
return int(major), int(minor)
def _increment_tag(tag, release_type):
major, minor = _parse_tag(tag)
if release_type == 'major':
new_major = major + 1
new_minor = 0
else:
new_major = major
new_minor = minor + 1
return '{}.{}'.format(new_major, new_minor)
def release(release_type, signed):
if not git_helpers.is_master():
raise errors.NotMasterException("Current branch is not master.\nAborting.")
tag = git_helpers.get_current_tag()
if not tag:
raise errors.NoTagException("Unable to get current tag.\nAborting.")
- new_tag = _increment_tag(tag)
+ new_tag = _increment_tag(tag, release_type)
git_helpers.tag(signed, new_tag)
| Add missing argument to _increment_tag call | ## Code Before:
import subprocess
from git_release import errors, git_helpers
def _parse_tag(tag):
major, minor = tag.split('.')
return int(major), int(minor)
def _increment_tag(tag, release_type):
major, minor = _parse_tag(tag)
if release_type == 'major':
new_major = major + 1
new_minor = 0
else:
new_major = major
new_minor = minor + 1
return '{}.{}'.format(new_major, new_minor)
def release(release_type, signed):
if not git_helpers.is_master():
raise errors.NotMasterException("Current branch is not master.\nAborting.")
tag = git_helpers.get_current_tag()
if not tag:
raise errors.NoTagException("Unable to get current tag.\nAborting.")
new_tag = _increment_tag(tag)
git_helpers.tag(signed, new_tag)
## Instruction:
Add missing argument to _increment_tag call
## Code After:
import subprocess
from git_release import errors, git_helpers
def _parse_tag(tag):
major, minor = tag.split('.')
return int(major), int(minor)
def _increment_tag(tag, release_type):
major, minor = _parse_tag(tag)
if release_type == 'major':
new_major = major + 1
new_minor = 0
else:
new_major = major
new_minor = minor + 1
return '{}.{}'.format(new_major, new_minor)
def release(release_type, signed):
if not git_helpers.is_master():
raise errors.NotMasterException("Current branch is not master.\nAborting.")
tag = git_helpers.get_current_tag()
if not tag:
raise errors.NoTagException("Unable to get current tag.\nAborting.")
new_tag = _increment_tag(tag, release_type)
git_helpers.tag(signed, new_tag)
|
11aab47e3c8c0d4044042aead7c01c990a152bea | tests/integration/customer/test_dispatcher.py | tests/integration/customer/test_dispatcher.py | from django.test import TestCase
from django.core import mail
from oscar.core.compat import get_user_model
from oscar.apps.customer.utils import Dispatcher
from oscar.apps.customer.models import CommunicationEventType
from oscar.test.factories import create_order
User = get_user_model()
class TestDispatcher(TestCase):
def test_sending_a_order_related_messages(self):
email = '[email protected]'
user = User.objects.create_user('testuser', email,
'somesimplepassword')
order_number = '12345'
order = create_order(number=order_number, user=user)
et = CommunicationEventType.objects.create(code="ORDER_PLACED",
name="Order Placed",
category="Order related")
messages = et.get_messages({
'order': order,
'lines': order.lines.all()
})
self.assertIn(order_number, messages['body'])
self.assertIn(order_number, messages['html'])
dispatcher = Dispatcher()
dispatcher.dispatch_order_messages(order, messages, et)
self.assertEqual(len(mail.outbox), 1)
message = mail.outbox[0]
self.assertIn(order_number, message.body)
| from django.test import TestCase
from django.core import mail
from oscar.core.compat import get_user_model
from oscar.apps.customer.utils import Dispatcher
from oscar.apps.customer.models import CommunicationEventType
from oscar.test.factories import create_order
User = get_user_model()
class TestDispatcher(TestCase):
def test_sending_a_order_related_messages(self):
email = '[email protected]'
user = User.objects.create_user('testuser', email,
'somesimplepassword')
order_number = '12345'
order = create_order(number=order_number, user=user)
et = CommunicationEventType.objects.create(code="ORDER_PLACED",
name="Order Placed",
category="Order related")
messages = et.get_messages({
'order': order,
'lines': order.lines.all()
})
self.assertIn(order_number, messages['body'])
self.assertIn(order_number, messages['html'])
dispatcher = Dispatcher()
dispatcher.dispatch_order_messages(order, messages, et)
self.assertEqual(len(mail.outbox), 1)
message = mail.outbox[0]
self.assertIn(order_number, message.body)
# test sending messages to emails without account and text body
messages.pop('body')
dispatcher.dispatch_direct_messages(email, messages)
self.assertEqual(len(mail.outbox), 2)
| Add tests for sending messages to emails without account. | Add tests for sending messages to emails without account.
| Python | bsd-3-clause | sonofatailor/django-oscar,solarissmoke/django-oscar,okfish/django-oscar,okfish/django-oscar,django-oscar/django-oscar,sasha0/django-oscar,sonofatailor/django-oscar,django-oscar/django-oscar,django-oscar/django-oscar,solarissmoke/django-oscar,django-oscar/django-oscar,okfish/django-oscar,sonofatailor/django-oscar,okfish/django-oscar,sasha0/django-oscar,solarissmoke/django-oscar,sasha0/django-oscar,sonofatailor/django-oscar,sasha0/django-oscar,solarissmoke/django-oscar | from django.test import TestCase
from django.core import mail
from oscar.core.compat import get_user_model
from oscar.apps.customer.utils import Dispatcher
from oscar.apps.customer.models import CommunicationEventType
from oscar.test.factories import create_order
User = get_user_model()
class TestDispatcher(TestCase):
def test_sending_a_order_related_messages(self):
email = '[email protected]'
user = User.objects.create_user('testuser', email,
'somesimplepassword')
order_number = '12345'
order = create_order(number=order_number, user=user)
et = CommunicationEventType.objects.create(code="ORDER_PLACED",
name="Order Placed",
category="Order related")
messages = et.get_messages({
'order': order,
'lines': order.lines.all()
})
self.assertIn(order_number, messages['body'])
self.assertIn(order_number, messages['html'])
dispatcher = Dispatcher()
dispatcher.dispatch_order_messages(order, messages, et)
self.assertEqual(len(mail.outbox), 1)
message = mail.outbox[0]
self.assertIn(order_number, message.body)
+ # test sending messages to emails without account and text body
+ messages.pop('body')
+ dispatcher.dispatch_direct_messages(email, messages)
+ self.assertEqual(len(mail.outbox), 2)
| Add tests for sending messages to emails without account. | ## Code Before:
from django.test import TestCase
from django.core import mail
from oscar.core.compat import get_user_model
from oscar.apps.customer.utils import Dispatcher
from oscar.apps.customer.models import CommunicationEventType
from oscar.test.factories import create_order
User = get_user_model()
class TestDispatcher(TestCase):
def test_sending_a_order_related_messages(self):
email = '[email protected]'
user = User.objects.create_user('testuser', email,
'somesimplepassword')
order_number = '12345'
order = create_order(number=order_number, user=user)
et = CommunicationEventType.objects.create(code="ORDER_PLACED",
name="Order Placed",
category="Order related")
messages = et.get_messages({
'order': order,
'lines': order.lines.all()
})
self.assertIn(order_number, messages['body'])
self.assertIn(order_number, messages['html'])
dispatcher = Dispatcher()
dispatcher.dispatch_order_messages(order, messages, et)
self.assertEqual(len(mail.outbox), 1)
message = mail.outbox[0]
self.assertIn(order_number, message.body)
## Instruction:
Add tests for sending messages to emails without account.
## Code After:
from django.test import TestCase
from django.core import mail
from oscar.core.compat import get_user_model
from oscar.apps.customer.utils import Dispatcher
from oscar.apps.customer.models import CommunicationEventType
from oscar.test.factories import create_order
User = get_user_model()
class TestDispatcher(TestCase):
def test_sending_a_order_related_messages(self):
email = '[email protected]'
user = User.objects.create_user('testuser', email,
'somesimplepassword')
order_number = '12345'
order = create_order(number=order_number, user=user)
et = CommunicationEventType.objects.create(code="ORDER_PLACED",
name="Order Placed",
category="Order related")
messages = et.get_messages({
'order': order,
'lines': order.lines.all()
})
self.assertIn(order_number, messages['body'])
self.assertIn(order_number, messages['html'])
dispatcher = Dispatcher()
dispatcher.dispatch_order_messages(order, messages, et)
self.assertEqual(len(mail.outbox), 1)
message = mail.outbox[0]
self.assertIn(order_number, message.body)
# test sending messages to emails without account and text body
messages.pop('body')
dispatcher.dispatch_direct_messages(email, messages)
self.assertEqual(len(mail.outbox), 2)
|
f3efb01c530db87f48d813b118f80a2ee1fd5996 | dthm4kaiako/users/apps.py | dthm4kaiako/users/apps.py | """Application configuration for the chapters application."""
from django.apps import AppConfig
class UsersAppConfig(AppConfig):
"""Configuration object for the chapters application."""
name = "users"
verbose_name = "Users"
def ready(self):
"""Import signals upon intialising application."""
import users.signals
| """Application configuration for the chapters application."""
from django.apps import AppConfig
class UsersAppConfig(AppConfig):
"""Configuration object for the chapters application."""
name = "users"
verbose_name = "Users"
def ready(self):
"""Import signals upon intialising application."""
import users.signals # noqa F401
| Exclude import from style checking | Exclude import from style checking
| Python | mit | uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers | """Application configuration for the chapters application."""
from django.apps import AppConfig
class UsersAppConfig(AppConfig):
"""Configuration object for the chapters application."""
name = "users"
verbose_name = "Users"
def ready(self):
"""Import signals upon intialising application."""
- import users.signals
+ import users.signals # noqa F401
| Exclude import from style checking | ## Code Before:
"""Application configuration for the chapters application."""
from django.apps import AppConfig
class UsersAppConfig(AppConfig):
"""Configuration object for the chapters application."""
name = "users"
verbose_name = "Users"
def ready(self):
"""Import signals upon intialising application."""
import users.signals
## Instruction:
Exclude import from style checking
## Code After:
"""Application configuration for the chapters application."""
from django.apps import AppConfig
class UsersAppConfig(AppConfig):
"""Configuration object for the chapters application."""
name = "users"
verbose_name = "Users"
def ready(self):
"""Import signals upon intialising application."""
import users.signals # noqa F401
|
6454372da6550455735cbcb3a86a966e61c134a1 | elasticsearch/__init__.py | elasticsearch/__init__.py | from __future__ import absolute_import
VERSION = (0, 4, 3)
__version__ = VERSION
__versionstr__ = '.'.join(map(str, VERSION))
from elasticsearch.client import Elasticsearch
from elasticsearch.transport import Transport
from elasticsearch.connection_pool import ConnectionPool, ConnectionSelector, \
RoundRobinSelector
from elasticsearch.serializer import JSONSerializer
from elasticsearch.connection import Connection, RequestsHttpConnection, \
Urllib3HttpConnection, MemcachedConnection
from elasticsearch.exceptions import *
| from __future__ import absolute_import
VERSION = (0, 4, 3)
__version__ = VERSION
__versionstr__ = '.'.join(map(str, VERSION))
from elasticsearch.client import Elasticsearch
from elasticsearch.transport import Transport
from elasticsearch.connection_pool import ConnectionPool, ConnectionSelector, \
RoundRobinSelector
from elasticsearch.serializer import JSONSerializer
from elasticsearch.connection import Connection, RequestsHttpConnection, \
Urllib3HttpConnection, MemcachedConnection, ThriftConnection
from elasticsearch.exceptions import *
| Allow people to import ThriftConnection from elasticsearch package itself | Allow people to import ThriftConnection from elasticsearch package itself
| Python | apache-2.0 | veatch/elasticsearch-py,chrisseto/elasticsearch-py,Garrett-R/elasticsearch-py,brunobell/elasticsearch-py,tailhook/elasticsearch-py,AlexMaskovyak/elasticsearch-py,brunobell/elasticsearch-py,mjhennig/elasticsearch-py,thomdixon/elasticsearch-py,kelp404/elasticsearch-py,gardsted/elasticsearch-py,elastic/elasticsearch-py,elastic/elasticsearch-py,konradkonrad/elasticsearch-py,liuyi1112/elasticsearch-py,prinsherbert/elasticsearch-py | from __future__ import absolute_import
VERSION = (0, 4, 3)
__version__ = VERSION
__versionstr__ = '.'.join(map(str, VERSION))
from elasticsearch.client import Elasticsearch
from elasticsearch.transport import Transport
from elasticsearch.connection_pool import ConnectionPool, ConnectionSelector, \
RoundRobinSelector
from elasticsearch.serializer import JSONSerializer
from elasticsearch.connection import Connection, RequestsHttpConnection, \
- Urllib3HttpConnection, MemcachedConnection
+ Urllib3HttpConnection, MemcachedConnection, ThriftConnection
from elasticsearch.exceptions import *
| Allow people to import ThriftConnection from elasticsearch package itself | ## Code Before:
from __future__ import absolute_import
VERSION = (0, 4, 3)
__version__ = VERSION
__versionstr__ = '.'.join(map(str, VERSION))
from elasticsearch.client import Elasticsearch
from elasticsearch.transport import Transport
from elasticsearch.connection_pool import ConnectionPool, ConnectionSelector, \
RoundRobinSelector
from elasticsearch.serializer import JSONSerializer
from elasticsearch.connection import Connection, RequestsHttpConnection, \
Urllib3HttpConnection, MemcachedConnection
from elasticsearch.exceptions import *
## Instruction:
Allow people to import ThriftConnection from elasticsearch package itself
## Code After:
from __future__ import absolute_import
VERSION = (0, 4, 3)
__version__ = VERSION
__versionstr__ = '.'.join(map(str, VERSION))
from elasticsearch.client import Elasticsearch
from elasticsearch.transport import Transport
from elasticsearch.connection_pool import ConnectionPool, ConnectionSelector, \
RoundRobinSelector
from elasticsearch.serializer import JSONSerializer
from elasticsearch.connection import Connection, RequestsHttpConnection, \
Urllib3HttpConnection, MemcachedConnection, ThriftConnection
from elasticsearch.exceptions import *
|
56446567f764625e88d8efdbfa2849e0a579d5c4 | indra/tests/test_rest_api.py | indra/tests/test_rest_api.py | import requests
from nose.plugins.attrib import attr
@attr('webservice')
def test_rest_api_responsive():
stmt_str = '{"statements": [{"sbo": "http://identifiers.org/sbo/SBO:0000526", "type": "Complex", "id": "acc6d47c-f622-41a4-8ae9-d7b0f3d24a2f", "members": [{"db_refs": {"TEXT": "MEK", "FPLX": "MEK"}, "name": "MEK"}, {"db_refs": {"TEXT": "ERK", "NCIT": "C26360", "FPLX": "ERK"}, "name": "ERK"}], "evidence": [{"text": "MEK binds ERK", "source_api": "trips"}]}]}'
url = 'http://ec2-54-88-146-250.compute-1.amazonaws.com:8080/' + \
'assemblers/cyjs'
res = requests.post(url, stmt_str)
assert res.status_code == 200
| import requests
from nose.plugins.attrib import attr
@attr('webservice')
def test_rest_api_responsive():
stmt_str = '{"statements": [{"sbo": "http://identifiers.org/sbo/SBO:0000526", "type": "Complex", "id": "acc6d47c-f622-41a4-8ae9-d7b0f3d24a2f", "members": [{"db_refs": {"TEXT": "MEK", "FPLX": "MEK"}, "name": "MEK"}, {"db_refs": {"TEXT": "ERK", "NCIT": "C26360", "FPLX": "ERK"}, "name": "ERK"}], "evidence": [{"text": "MEK binds ERK", "source_api": "trips"}]}]}'
url = 'http://indra-api-72031e2dfde08e09.elb.us-east-1.amazonaws.com:8000/' + \
'assemblers/cyjs'
res = requests.post(url, stmt_str)
assert res.status_code == 200
| Update REST API address in test | Update REST API address in test
| Python | bsd-2-clause | sorgerlab/belpy,sorgerlab/indra,sorgerlab/belpy,bgyori/indra,pvtodorov/indra,bgyori/indra,sorgerlab/indra,johnbachman/belpy,pvtodorov/indra,pvtodorov/indra,sorgerlab/indra,pvtodorov/indra,johnbachman/belpy,bgyori/indra,sorgerlab/belpy,johnbachman/belpy,johnbachman/indra,johnbachman/indra,johnbachman/indra | import requests
from nose.plugins.attrib import attr
@attr('webservice')
def test_rest_api_responsive():
stmt_str = '{"statements": [{"sbo": "http://identifiers.org/sbo/SBO:0000526", "type": "Complex", "id": "acc6d47c-f622-41a4-8ae9-d7b0f3d24a2f", "members": [{"db_refs": {"TEXT": "MEK", "FPLX": "MEK"}, "name": "MEK"}, {"db_refs": {"TEXT": "ERK", "NCIT": "C26360", "FPLX": "ERK"}, "name": "ERK"}], "evidence": [{"text": "MEK binds ERK", "source_api": "trips"}]}]}'
- url = 'http://ec2-54-88-146-250.compute-1.amazonaws.com:8080/' + \
+ url = 'http://indra-api-72031e2dfde08e09.elb.us-east-1.amazonaws.com:8000/' + \
'assemblers/cyjs'
res = requests.post(url, stmt_str)
assert res.status_code == 200
| Update REST API address in test | ## Code Before:
import requests
from nose.plugins.attrib import attr
@attr('webservice')
def test_rest_api_responsive():
stmt_str = '{"statements": [{"sbo": "http://identifiers.org/sbo/SBO:0000526", "type": "Complex", "id": "acc6d47c-f622-41a4-8ae9-d7b0f3d24a2f", "members": [{"db_refs": {"TEXT": "MEK", "FPLX": "MEK"}, "name": "MEK"}, {"db_refs": {"TEXT": "ERK", "NCIT": "C26360", "FPLX": "ERK"}, "name": "ERK"}], "evidence": [{"text": "MEK binds ERK", "source_api": "trips"}]}]}'
url = 'http://ec2-54-88-146-250.compute-1.amazonaws.com:8080/' + \
'assemblers/cyjs'
res = requests.post(url, stmt_str)
assert res.status_code == 200
## Instruction:
Update REST API address in test
## Code After:
import requests
from nose.plugins.attrib import attr
@attr('webservice')
def test_rest_api_responsive():
stmt_str = '{"statements": [{"sbo": "http://identifiers.org/sbo/SBO:0000526", "type": "Complex", "id": "acc6d47c-f622-41a4-8ae9-d7b0f3d24a2f", "members": [{"db_refs": {"TEXT": "MEK", "FPLX": "MEK"}, "name": "MEK"}, {"db_refs": {"TEXT": "ERK", "NCIT": "C26360", "FPLX": "ERK"}, "name": "ERK"}], "evidence": [{"text": "MEK binds ERK", "source_api": "trips"}]}]}'
url = 'http://indra-api-72031e2dfde08e09.elb.us-east-1.amazonaws.com:8000/' + \
'assemblers/cyjs'
res = requests.post(url, stmt_str)
assert res.status_code == 200
|
6bdbbf4d5e100856acbaba1c5fc024a9f7f78718 | tests/tools.py | tests/tools.py |
__author__ = 'mbach'
import contextlib
import shutil
import subprocess
import tempfile
@contextlib.contextmanager
def devpi_server(port=2414):
server_dir = tempfile.mkdtemp()
try:
subprocess.check_call(['devpi-server', '--start', '--serverdir={}'.format(server_dir), '--port={}'.format(port)])
try:
yield 'http://localhost:{}'.format(port)
finally:
subprocess.check_call(['devpi-server', '--stop', '--serverdir={}'.format(server_dir)])
finally:
shutil.rmtree(server_dir)
|
__author__ = 'mbach'
import contextlib
import shutil
import subprocess
import tempfile
from brandon import devpi
@contextlib.contextmanager
def devpi_server(port=2414):
server_dir = tempfile.mkdtemp()
try:
subprocess.check_call(['devpi-server', '--start', '--serverdir={}'.format(server_dir), '--port={}'.format(port)])
try:
yield 'http://localhost:{}'.format(port)
finally:
subprocess.check_call(['devpi-server', '--stop', '--serverdir={}'.format(server_dir)])
finally:
shutil.rmtree(server_dir)
@contextlib.contextmanager
def devpi_index(server_url, user, index):
"""
Creates the given user and index, and cleans it afterwards.
Yields of tuple of index-url and password. The index is created without an upstream.
"""
password = 'foo'
devpi_client = devpi.Client(server_url)
devpi_client._execute('user', '-c', user, 'password=' + password)
devpi_client._execute('login', user, '--password=' + password)
devpi_client._execute('index', '-c', 'wheels', 'bases=')
yield '{}/{}/{}'.format(server_url, user, index), password
devpi_client._execute('index', '--delete', '/{}/{}'.format(user, index))
devpi_client._execute('user', user, '--delete')
| Test tool to create temporary devpi index. | Test tool to create temporary devpi index.
| Python | bsd-3-clause | tylerdave/devpi-builder |
__author__ = 'mbach'
import contextlib
import shutil
import subprocess
import tempfile
+ from brandon import devpi
@contextlib.contextmanager
def devpi_server(port=2414):
server_dir = tempfile.mkdtemp()
try:
subprocess.check_call(['devpi-server', '--start', '--serverdir={}'.format(server_dir), '--port={}'.format(port)])
try:
yield 'http://localhost:{}'.format(port)
finally:
subprocess.check_call(['devpi-server', '--stop', '--serverdir={}'.format(server_dir)])
finally:
shutil.rmtree(server_dir)
+
+ @contextlib.contextmanager
+ def devpi_index(server_url, user, index):
+ """
+ Creates the given user and index, and cleans it afterwards.
+
+ Yields of tuple of index-url and password. The index is created without an upstream.
+ """
+ password = 'foo'
+ devpi_client = devpi.Client(server_url)
+ devpi_client._execute('user', '-c', user, 'password=' + password)
+ devpi_client._execute('login', user, '--password=' + password)
+ devpi_client._execute('index', '-c', 'wheels', 'bases=')
+
+ yield '{}/{}/{}'.format(server_url, user, index), password
+
+ devpi_client._execute('index', '--delete', '/{}/{}'.format(user, index))
+ devpi_client._execute('user', user, '--delete')
+ | Test tool to create temporary devpi index. | ## Code Before:
__author__ = 'mbach'
import contextlib
import shutil
import subprocess
import tempfile
@contextlib.contextmanager
def devpi_server(port=2414):
server_dir = tempfile.mkdtemp()
try:
subprocess.check_call(['devpi-server', '--start', '--serverdir={}'.format(server_dir), '--port={}'.format(port)])
try:
yield 'http://localhost:{}'.format(port)
finally:
subprocess.check_call(['devpi-server', '--stop', '--serverdir={}'.format(server_dir)])
finally:
shutil.rmtree(server_dir)
## Instruction:
Test tool to create temporary devpi index.
## Code After:
__author__ = 'mbach'
import contextlib
import shutil
import subprocess
import tempfile
from brandon import devpi
@contextlib.contextmanager
def devpi_server(port=2414):
server_dir = tempfile.mkdtemp()
try:
subprocess.check_call(['devpi-server', '--start', '--serverdir={}'.format(server_dir), '--port={}'.format(port)])
try:
yield 'http://localhost:{}'.format(port)
finally:
subprocess.check_call(['devpi-server', '--stop', '--serverdir={}'.format(server_dir)])
finally:
shutil.rmtree(server_dir)
@contextlib.contextmanager
def devpi_index(server_url, user, index):
"""
Creates the given user and index, and cleans it afterwards.
Yields of tuple of index-url and password. The index is created without an upstream.
"""
password = 'foo'
devpi_client = devpi.Client(server_url)
devpi_client._execute('user', '-c', user, 'password=' + password)
devpi_client._execute('login', user, '--password=' + password)
devpi_client._execute('index', '-c', 'wheels', 'bases=')
yield '{}/{}/{}'.format(server_url, user, index), password
devpi_client._execute('index', '--delete', '/{}/{}'.format(user, index))
devpi_client._execute('user', user, '--delete')
|
6d291571dca59243c0a92f9955776e1acd2e87da | falmer/content/queries.py | falmer/content/queries.py | import graphene
from django.http import Http404
from graphql import GraphQLError
from wagtail.core.models import Page
from . import types
class Query(graphene.ObjectType):
page = graphene.Field(types.Page, path=graphene.String())
all_pages = graphene.List(types.Page, path=graphene.String())
def resolve_page(self, info, **kwargs):
path = kwargs.get('path')
path = path[1:] if path.startswith('/') else path
path = path[:-1] if path.endswith('/') else path
root_page = info.context.site.root_page
try:
result = root_page.route(info.context, path.split('/'))
return result.page
except Http404:
raise GraphQLError(f'404: Page not found for {path}')
def resolve_all_pages(self, info):
return Page.objects.specific().live()
| import graphene
from django.http import Http404
from graphql import GraphQLError
from wagtail.core.models import Page
from . import types
class Query(graphene.ObjectType):
page = graphene.Field(types.Page, path=graphene.String())
all_pages = graphene.List(types.Page, path=graphene.String())
def resolve_page(self, info, **kwargs):
path = kwargs.get('path')
path = path[1:] if path.startswith('/') else path
path = path[:-1] if path.endswith('/') else path
root_page = info.context.site.root_page
try:
result = root_page.route(info.context, path.split('/'))
return result.page
except Http404:
return None
def resolve_all_pages(self, info):
return Page.objects.specific().live()
| Return empty result rather than graphql error | Return empty result rather than graphql error
| Python | mit | sussexstudent/falmer,sussexstudent/falmer,sussexstudent/falmer,sussexstudent/falmer | import graphene
from django.http import Http404
from graphql import GraphQLError
from wagtail.core.models import Page
from . import types
class Query(graphene.ObjectType):
page = graphene.Field(types.Page, path=graphene.String())
all_pages = graphene.List(types.Page, path=graphene.String())
def resolve_page(self, info, **kwargs):
path = kwargs.get('path')
path = path[1:] if path.startswith('/') else path
path = path[:-1] if path.endswith('/') else path
root_page = info.context.site.root_page
try:
result = root_page.route(info.context, path.split('/'))
return result.page
except Http404:
- raise GraphQLError(f'404: Page not found for {path}')
+ return None
def resolve_all_pages(self, info):
return Page.objects.specific().live()
| Return empty result rather than graphql error | ## Code Before:
import graphene
from django.http import Http404
from graphql import GraphQLError
from wagtail.core.models import Page
from . import types
class Query(graphene.ObjectType):
page = graphene.Field(types.Page, path=graphene.String())
all_pages = graphene.List(types.Page, path=graphene.String())
def resolve_page(self, info, **kwargs):
path = kwargs.get('path')
path = path[1:] if path.startswith('/') else path
path = path[:-1] if path.endswith('/') else path
root_page = info.context.site.root_page
try:
result = root_page.route(info.context, path.split('/'))
return result.page
except Http404:
raise GraphQLError(f'404: Page not found for {path}')
def resolve_all_pages(self, info):
return Page.objects.specific().live()
## Instruction:
Return empty result rather than graphql error
## Code After:
import graphene
from django.http import Http404
from graphql import GraphQLError
from wagtail.core.models import Page
from . import types
class Query(graphene.ObjectType):
page = graphene.Field(types.Page, path=graphene.String())
all_pages = graphene.List(types.Page, path=graphene.String())
def resolve_page(self, info, **kwargs):
path = kwargs.get('path')
path = path[1:] if path.startswith('/') else path
path = path[:-1] if path.endswith('/') else path
root_page = info.context.site.root_page
try:
result = root_page.route(info.context, path.split('/'))
return result.page
except Http404:
return None
def resolve_all_pages(self, info):
return Page.objects.specific().live()
|
d63905158f5148b07534e823d271326262369d42 | pavement.py | pavement.py | import os
import re
from paver.easy import *
from paver.setuputils import setup
def get_version():
"""
Grab the version from irclib.py.
"""
here = os.path.dirname(__file__)
irclib = os.path.join(here, 'irclib.py')
with open(irclib) as f:
content = f.read()
VERSION = eval(re.search('VERSION = (.*)', content).group(1))
VERSION = '.'.join(map(str, VERSION))
return VERSION
def read_long_description():
f = open('README')
try:
data = f.read()
finally:
f.close()
return data
setup(
name="python-irclib",
description="IRC (Internet Relay Chat) protocol client library for Python",
long_description=read_long_description(),
version=get_version(),
py_modules=["irclib", "ircbot"],
author="Joel Rosdahl",
author_email="[email protected]",
maintainer="Jason R. Coombs",
maintainer_email="[email protected]",
url="http://python-irclib.sourceforge.net",
classifiers = [
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
],
)
@task
@needs('generate_setup', 'minilib', 'distutils.command.sdist')
def sdist():
"Override sdist to make sure the setup.py gets generated"
| import os
import re
from paver.easy import *
from paver.setuputils import setup
def get_version():
"""
Grab the version from irclib.py.
"""
here = os.path.dirname(__file__)
irclib = os.path.join(here, 'irclib.py')
with open(irclib) as f:
content = f.read()
VERSION = eval(re.search('VERSION = (.*)', content).group(1))
VERSION = '.'.join(map(str, VERSION))
return VERSION
def read_long_description():
with open('README') as f:
data = f.read()
return data
setup(
name="python-irclib",
description="IRC (Internet Relay Chat) protocol client library for Python",
long_description=read_long_description(),
version=get_version(),
py_modules=["irclib", "ircbot"],
author="Joel Rosdahl",
author_email="[email protected]",
maintainer="Jason R. Coombs",
maintainer_email="[email protected]",
url="http://python-irclib.sourceforge.net",
classifiers = [
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
],
)
@task
@needs('generate_setup', 'minilib', 'distutils.command.sdist')
def sdist():
"Override sdist to make sure the setup.py gets generated"
| Use context manager to read README | Use context manager to read README
| Python | mit | jaraco/irc | import os
import re
from paver.easy import *
from paver.setuputils import setup
def get_version():
"""
Grab the version from irclib.py.
"""
here = os.path.dirname(__file__)
irclib = os.path.join(here, 'irclib.py')
with open(irclib) as f:
content = f.read()
VERSION = eval(re.search('VERSION = (.*)', content).group(1))
VERSION = '.'.join(map(str, VERSION))
return VERSION
def read_long_description():
+ with open('README') as f:
- f = open('README')
- try:
data = f.read()
- finally:
- f.close()
return data
setup(
name="python-irclib",
description="IRC (Internet Relay Chat) protocol client library for Python",
long_description=read_long_description(),
version=get_version(),
py_modules=["irclib", "ircbot"],
author="Joel Rosdahl",
author_email="[email protected]",
maintainer="Jason R. Coombs",
maintainer_email="[email protected]",
url="http://python-irclib.sourceforge.net",
classifiers = [
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
],
)
@task
@needs('generate_setup', 'minilib', 'distutils.command.sdist')
def sdist():
"Override sdist to make sure the setup.py gets generated"
| Use context manager to read README | ## Code Before:
import os
import re
from paver.easy import *
from paver.setuputils import setup
def get_version():
"""
Grab the version from irclib.py.
"""
here = os.path.dirname(__file__)
irclib = os.path.join(here, 'irclib.py')
with open(irclib) as f:
content = f.read()
VERSION = eval(re.search('VERSION = (.*)', content).group(1))
VERSION = '.'.join(map(str, VERSION))
return VERSION
def read_long_description():
f = open('README')
try:
data = f.read()
finally:
f.close()
return data
setup(
name="python-irclib",
description="IRC (Internet Relay Chat) protocol client library for Python",
long_description=read_long_description(),
version=get_version(),
py_modules=["irclib", "ircbot"],
author="Joel Rosdahl",
author_email="[email protected]",
maintainer="Jason R. Coombs",
maintainer_email="[email protected]",
url="http://python-irclib.sourceforge.net",
classifiers = [
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
],
)
@task
@needs('generate_setup', 'minilib', 'distutils.command.sdist')
def sdist():
"Override sdist to make sure the setup.py gets generated"
## Instruction:
Use context manager to read README
## Code After:
import os
import re
from paver.easy import *
from paver.setuputils import setup
def get_version():
"""
Grab the version from irclib.py.
"""
here = os.path.dirname(__file__)
irclib = os.path.join(here, 'irclib.py')
with open(irclib) as f:
content = f.read()
VERSION = eval(re.search('VERSION = (.*)', content).group(1))
VERSION = '.'.join(map(str, VERSION))
return VERSION
def read_long_description():
with open('README') as f:
data = f.read()
return data
setup(
name="python-irclib",
description="IRC (Internet Relay Chat) protocol client library for Python",
long_description=read_long_description(),
version=get_version(),
py_modules=["irclib", "ircbot"],
author="Joel Rosdahl",
author_email="[email protected]",
maintainer="Jason R. Coombs",
maintainer_email="[email protected]",
url="http://python-irclib.sourceforge.net",
classifiers = [
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
],
)
@task
@needs('generate_setup', 'minilib', 'distutils.command.sdist')
def sdist():
"Override sdist to make sure the setup.py gets generated"
|
d407f1bcd95daf4f4bd8dfe8ae3b4b9e68061cb5 | cref/sequence/fragment.py | cref/sequence/fragment.py |
def fragment(sequence, size=5):
"""
Fragment a string sequence using a sliding window given by size
:param sequence: String containing the sequence
:param size: Size of the window
:return: a fragment of the sequence with the given size
"""
for i in range(len(sequence) - size + 1):
yield sequence[i: i + size]
|
def fragment(sequence, size=5):
"""
Fragment a string sequence using a sliding window given by size
:param sequence: String containing the sequence
:param size: Size of the window
:return: a fragment of the sequence with the given size
"""
if size > 0:
for i in range(len(sequence) - size + 1):
yield sequence[i: i + size]
| Handle sliding window with size 0 | Handle sliding window with size 0
| Python | mit | mchelem/cref2,mchelem/cref2,mchelem/cref2 |
def fragment(sequence, size=5):
"""
Fragment a string sequence using a sliding window given by size
:param sequence: String containing the sequence
:param size: Size of the window
:return: a fragment of the sequence with the given size
"""
+ if size > 0:
- for i in range(len(sequence) - size + 1):
+ for i in range(len(sequence) - size + 1):
- yield sequence[i: i + size]
+ yield sequence[i: i + size]
- | Handle sliding window with size 0 | ## Code Before:
def fragment(sequence, size=5):
"""
Fragment a string sequence using a sliding window given by size
:param sequence: String containing the sequence
:param size: Size of the window
:return: a fragment of the sequence with the given size
"""
for i in range(len(sequence) - size + 1):
yield sequence[i: i + size]
## Instruction:
Handle sliding window with size 0
## Code After:
def fragment(sequence, size=5):
"""
Fragment a string sequence using a sliding window given by size
:param sequence: String containing the sequence
:param size: Size of the window
:return: a fragment of the sequence with the given size
"""
if size > 0:
for i in range(len(sequence) - size + 1):
yield sequence[i: i + size]
|
2f16eb25db856b72138f6dfb7d19e799bd460287 | tests/test_helpers.py | tests/test_helpers.py | import pytest
from os.path import basename
from helpers import utils, fixture
@pytest.mark.skipif(pytest.config.getoption("--application") is not False, reason="application passed; skipping base module tests")
class TestHelpers():
def test_wildcards1():
d = utils.get_wildcards([('"{prefix}.bam"', "medium.bam")], {})
assert d['prefix'] == "medium"
def test_wildcards2():
d = utils.get_wildcards([('"{prefix}{ext,.bam}"', "medium.bam")], {})
assert d['ext'] == ".bam"
def test_wildcards3():
d = utils.get_wildcards([('"{prefix}.bar"', "/foo/bar/medium.bar")], {})
assert d['prefix'] == 'medium'
def test_wildcards4():
d = utils.get_wildcards([('config[\'foo\'] + ".bar"', "config.yaml")], {})
assert d == {}
def test_determine_fixture():
# Non-existent filetype
ft = fixture.determine_fixture('"{prefix}.bar"')
assert ft is None
ft = fixture.determine_fixture('"{prefix}.bam"')
assert basename(ft) == "PUR.HG00731.tiny.sort.bam"
ft = fixture.determine_fixture('config[\'foo\'] + ".dict"')
assert basename(ft) == "scaffolds.dict"
| import pytest
from os.path import basename
from helpers import utils, fixture
pytestmark = pytest.mark.skipif(pytest.config.getoption("--application") is not False, reason="application passed; skipping base module tests")
def test_wildcards1():
d = utils.get_wildcards([('"{prefix}.bam"', "medium.bam")], {})
assert d['prefix'] == "medium"
def test_wildcards2():
d = utils.get_wildcards([('"{prefix}{ext,.bam}"', "medium.bam")], {})
assert d['ext'] == ".bam"
def test_wildcards3():
d = utils.get_wildcards([('"{prefix}.bar"', "/foo/bar/medium.bar")], {})
assert d['prefix'] == 'medium'
def test_wildcards4():
d = utils.get_wildcards([('config[\'foo\'] + ".bar"', "config.yaml")], {})
assert d == {}
def test_determine_fixture():
# Non-existent filetype
ft = fixture.determine_fixture('"{prefix}.bar"')
assert ft is None
ft = fixture.determine_fixture('"{prefix}.bam"')
assert basename(ft) == "PUR.HG00731.tiny.sort.bam"
ft = fixture.determine_fixture('config[\'foo\'] + ".dict"')
assert basename(ft) == "scaffolds.dict"
| Use global pytestmark to skip tests; deprecate class | Use global pytestmark to skip tests; deprecate class
| Python | mit | percyfal/snakemake-rules,percyfal/snakemake-rules,percyfal/snakemakelib-rules,percyfal/snakemakelib-rules,percyfal/snakemakelib-rules | import pytest
from os.path import basename
from helpers import utils, fixture
-
- @pytest.mark.skipif(pytest.config.getoption("--application") is not False, reason="application passed; skipping base module tests")
+ pytestmark = pytest.mark.skipif(pytest.config.getoption("--application") is not False, reason="application passed; skipping base module tests")
- class TestHelpers():
- def test_wildcards1():
- d = utils.get_wildcards([('"{prefix}.bam"', "medium.bam")], {})
- assert d['prefix'] == "medium"
-
- def test_wildcards2():
- d = utils.get_wildcards([('"{prefix}{ext,.bam}"', "medium.bam")], {})
- assert d['ext'] == ".bam"
-
- def test_wildcards3():
- d = utils.get_wildcards([('"{prefix}.bar"', "/foo/bar/medium.bar")], {})
- assert d['prefix'] == 'medium'
-
- def test_wildcards4():
- d = utils.get_wildcards([('config[\'foo\'] + ".bar"', "config.yaml")], {})
- assert d == {}
+ def test_wildcards1():
+ d = utils.get_wildcards([('"{prefix}.bam"', "medium.bam")], {})
+ assert d['prefix'] == "medium"
- def test_determine_fixture():
- # Non-existent filetype
- ft = fixture.determine_fixture('"{prefix}.bar"')
- assert ft is None
- ft = fixture.determine_fixture('"{prefix}.bam"')
- assert basename(ft) == "PUR.HG00731.tiny.sort.bam"
- ft = fixture.determine_fixture('config[\'foo\'] + ".dict"')
- assert basename(ft) == "scaffolds.dict"
+
+ def test_wildcards2():
+ d = utils.get_wildcards([('"{prefix}{ext,.bam}"', "medium.bam")], {})
+ assert d['ext'] == ".bam"
+
+
+ def test_wildcards3():
+ d = utils.get_wildcards([('"{prefix}.bar"', "/foo/bar/medium.bar")], {})
+ assert d['prefix'] == 'medium'
+
+
+ def test_wildcards4():
+ d = utils.get_wildcards([('config[\'foo\'] + ".bar"', "config.yaml")], {})
+ assert d == {}
+
+
+ def test_determine_fixture():
+ # Non-existent filetype
+ ft = fixture.determine_fixture('"{prefix}.bar"')
+ assert ft is None
+ ft = fixture.determine_fixture('"{prefix}.bam"')
+ assert basename(ft) == "PUR.HG00731.tiny.sort.bam"
+ ft = fixture.determine_fixture('config[\'foo\'] + ".dict"')
+ assert basename(ft) == "scaffolds.dict"
+ | Use global pytestmark to skip tests; deprecate class | ## Code Before:
import pytest
from os.path import basename
from helpers import utils, fixture
@pytest.mark.skipif(pytest.config.getoption("--application") is not False, reason="application passed; skipping base module tests")
class TestHelpers():
def test_wildcards1():
d = utils.get_wildcards([('"{prefix}.bam"', "medium.bam")], {})
assert d['prefix'] == "medium"
def test_wildcards2():
d = utils.get_wildcards([('"{prefix}{ext,.bam}"', "medium.bam")], {})
assert d['ext'] == ".bam"
def test_wildcards3():
d = utils.get_wildcards([('"{prefix}.bar"', "/foo/bar/medium.bar")], {})
assert d['prefix'] == 'medium'
def test_wildcards4():
d = utils.get_wildcards([('config[\'foo\'] + ".bar"', "config.yaml")], {})
assert d == {}
def test_determine_fixture():
# Non-existent filetype
ft = fixture.determine_fixture('"{prefix}.bar"')
assert ft is None
ft = fixture.determine_fixture('"{prefix}.bam"')
assert basename(ft) == "PUR.HG00731.tiny.sort.bam"
ft = fixture.determine_fixture('config[\'foo\'] + ".dict"')
assert basename(ft) == "scaffolds.dict"
## Instruction:
Use global pytestmark to skip tests; deprecate class
## Code After:
import pytest
from os.path import basename
from helpers import utils, fixture
pytestmark = pytest.mark.skipif(pytest.config.getoption("--application") is not False, reason="application passed; skipping base module tests")
def test_wildcards1():
d = utils.get_wildcards([('"{prefix}.bam"', "medium.bam")], {})
assert d['prefix'] == "medium"
def test_wildcards2():
d = utils.get_wildcards([('"{prefix}{ext,.bam}"', "medium.bam")], {})
assert d['ext'] == ".bam"
def test_wildcards3():
d = utils.get_wildcards([('"{prefix}.bar"', "/foo/bar/medium.bar")], {})
assert d['prefix'] == 'medium'
def test_wildcards4():
d = utils.get_wildcards([('config[\'foo\'] + ".bar"', "config.yaml")], {})
assert d == {}
def test_determine_fixture():
# Non-existent filetype
ft = fixture.determine_fixture('"{prefix}.bar"')
assert ft is None
ft = fixture.determine_fixture('"{prefix}.bam"')
assert basename(ft) == "PUR.HG00731.tiny.sort.bam"
ft = fixture.determine_fixture('config[\'foo\'] + ".dict"')
assert basename(ft) == "scaffolds.dict"
|
723d7410b48fd4fc42ed9afe470ba3b37381599a | noxfile.py | noxfile.py | """Development automation."""
import nox
def _install_this_editable(session, *, extras=None):
if extras is None:
extras = []
session.install("flit")
session.run(
"flit",
"install",
"-s",
"--deps=production",
"--extras",
",".join(extras),
silent=True,
)
@nox.session
def lint(session):
session.install("pre-commit")
session.run("pre-commit", "run", "--all-files", *session.posargs)
@nox.session(python=["3.6", "3.7", "3.8"])
def test(session):
_install_this_editable(session, extras=["test"])
default_args = ["--cov-report", "term", "--cov", "sphinx_autobuild"]
args = session.posargs or default_args
session.run("pytest", *args)
@nox.session
def docs(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-build", "-b", "html", "docs/", "build/docs")
| """Development automation."""
import nox
def _install_this_editable(session, *, extras=None):
if extras is None:
extras = []
session.install("flit")
session.run(
"flit",
"install",
"-s",
"--deps=production",
"--extras",
",".join(extras),
silent=True,
)
@nox.session
def lint(session):
session.install("pre-commit")
session.run("pre-commit", "run", "--all-files", *session.posargs)
@nox.session(python=["3.6", "3.7", "3.8"])
def test(session):
_install_this_editable(session, extras=["test"])
default_args = ["--cov-report", "term", "--cov", "sphinx_autobuild"]
args = session.posargs or default_args
session.run("pytest", *args)
@nox.session
def docs(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-build", "-b", "html", "docs/", "build/docs")
@nox.session(name="docs-live")
def docs_live(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-autobuild", "-b", "html", "docs/", "build/docs")
| Add docs-live to perform demo-runs | Add docs-live to perform demo-runs
| Python | mit | GaretJax/sphinx-autobuild | """Development automation."""
import nox
def _install_this_editable(session, *, extras=None):
if extras is None:
extras = []
session.install("flit")
session.run(
"flit",
"install",
"-s",
"--deps=production",
"--extras",
",".join(extras),
silent=True,
)
@nox.session
def lint(session):
session.install("pre-commit")
session.run("pre-commit", "run", "--all-files", *session.posargs)
@nox.session(python=["3.6", "3.7", "3.8"])
def test(session):
_install_this_editable(session, extras=["test"])
default_args = ["--cov-report", "term", "--cov", "sphinx_autobuild"]
args = session.posargs or default_args
session.run("pytest", *args)
@nox.session
def docs(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-build", "-b", "html", "docs/", "build/docs")
+
+ @nox.session(name="docs-live")
+ def docs_live(session):
+ _install_this_editable(session, extras=["docs"])
+ session.run("sphinx-autobuild", "-b", "html", "docs/", "build/docs")
+ | Add docs-live to perform demo-runs | ## Code Before:
"""Development automation."""
import nox
def _install_this_editable(session, *, extras=None):
if extras is None:
extras = []
session.install("flit")
session.run(
"flit",
"install",
"-s",
"--deps=production",
"--extras",
",".join(extras),
silent=True,
)
@nox.session
def lint(session):
session.install("pre-commit")
session.run("pre-commit", "run", "--all-files", *session.posargs)
@nox.session(python=["3.6", "3.7", "3.8"])
def test(session):
_install_this_editable(session, extras=["test"])
default_args = ["--cov-report", "term", "--cov", "sphinx_autobuild"]
args = session.posargs or default_args
session.run("pytest", *args)
@nox.session
def docs(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-build", "-b", "html", "docs/", "build/docs")
## Instruction:
Add docs-live to perform demo-runs
## Code After:
"""Development automation."""
import nox
def _install_this_editable(session, *, extras=None):
if extras is None:
extras = []
session.install("flit")
session.run(
"flit",
"install",
"-s",
"--deps=production",
"--extras",
",".join(extras),
silent=True,
)
@nox.session
def lint(session):
session.install("pre-commit")
session.run("pre-commit", "run", "--all-files", *session.posargs)
@nox.session(python=["3.6", "3.7", "3.8"])
def test(session):
_install_this_editable(session, extras=["test"])
default_args = ["--cov-report", "term", "--cov", "sphinx_autobuild"]
args = session.posargs or default_args
session.run("pytest", *args)
@nox.session
def docs(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-build", "-b", "html", "docs/", "build/docs")
@nox.session(name="docs-live")
def docs_live(session):
_install_this_editable(session, extras=["docs"])
session.run("sphinx-autobuild", "-b", "html", "docs/", "build/docs")
|
41209aa3e27673f003ed62a46c9bfae0c19d0bf3 | il2fb/ds/airbridge/typing.py | il2fb/ds/airbridge/typing.py |
from pathlib import Path
from typing import Callable, Optional, List, Union
from il2fb.parsers.events.events import Event
EventOrNone = Optional[Event]
EventHandler = Callable[[Event], None]
IntOrNone = Optional[int]
StringProducer = Callable[[], str]
StringHandler = Callable[[str], None]
StringOrNone = Optional[str]
StringOrNoneProducer = Callable[[], StringOrNone]
StringOrPath = Union[str, Path]
StringList = List[str]
|
from pathlib import Path
from typing import Callable, Optional, List, Union
from il2fb.commons.events import Event
EventOrNone = Optional[Event]
EventHandler = Callable[[Event], None]
IntOrNone = Optional[int]
StringProducer = Callable[[], str]
StringHandler = Callable[[str], None]
StringOrNone = Optional[str]
StringOrNoneProducer = Callable[[], StringOrNone]
StringOrPath = Union[str, Path]
StringList = List[str]
| Update import of Event class | Update import of Event class
| Python | mit | IL2HorusTeam/il2fb-ds-airbridge |
from pathlib import Path
from typing import Callable, Optional, List, Union
- from il2fb.parsers.events.events import Event
+ from il2fb.commons.events import Event
EventOrNone = Optional[Event]
EventHandler = Callable[[Event], None]
IntOrNone = Optional[int]
StringProducer = Callable[[], str]
StringHandler = Callable[[str], None]
StringOrNone = Optional[str]
StringOrNoneProducer = Callable[[], StringOrNone]
StringOrPath = Union[str, Path]
StringList = List[str]
| Update import of Event class | ## Code Before:
from pathlib import Path
from typing import Callable, Optional, List, Union
from il2fb.parsers.events.events import Event
EventOrNone = Optional[Event]
EventHandler = Callable[[Event], None]
IntOrNone = Optional[int]
StringProducer = Callable[[], str]
StringHandler = Callable[[str], None]
StringOrNone = Optional[str]
StringOrNoneProducer = Callable[[], StringOrNone]
StringOrPath = Union[str, Path]
StringList = List[str]
## Instruction:
Update import of Event class
## Code After:
from pathlib import Path
from typing import Callable, Optional, List, Union
from il2fb.commons.events import Event
EventOrNone = Optional[Event]
EventHandler = Callable[[Event], None]
IntOrNone = Optional[int]
StringProducer = Callable[[], str]
StringHandler = Callable[[str], None]
StringOrNone = Optional[str]
StringOrNoneProducer = Callable[[], StringOrNone]
StringOrPath = Union[str, Path]
StringList = List[str]
|
7ee86e9b52292a8824dfa7bab632526cbb365b51 | routes.py | routes.py |
from flask import request, redirect
import requests
cookiename = 'openAMUserCookieName'
amURL = 'https://openam.example.com/'
validTokenAPI = amURL + 'openam/identity/istokenvalid?tokenid='
loginURL = amURL + 'openam/UI/Login'
def session_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
usercookie = request.cookies.get(cookiename)
if usercookie:
amQuery = requests.get(validTokenAPI + usercookie)
if 'boolean=true' in amQuery.text:
return f(*args, **kwargs)
return redirect(loginURL)
return decorated_function
@app.route('/members_page')
@session_required
def members_page():
pass
|
from flask import request, redirect
import requests
cookiename = 'openAMUserCookieName'
amURL = 'https://openam.example.com/'
validTokenAPI = amURL + 'openam/json/sessions/{token}?_action=validate'
loginURL = amURL + 'openam/UI/Login'
def session_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
usercookie = request.cookies.get(cookiename)
if usercookie:
amQuery = requests.post(validTokenAPI.format(token=usercookie))
if amQuery.json()['valid']:
return f(*args, **kwargs)
return redirect(loginURL)
return decorated_function
@app.route('/members_page')
@session_required
def members_page():
pass
| Use new OpenAM token validation endpoint | Use new OpenAM token validation endpoint
| Python | unlicense | timhberry/openam-flask-decorator |
from flask import request, redirect
import requests
cookiename = 'openAMUserCookieName'
amURL = 'https://openam.example.com/'
- validTokenAPI = amURL + 'openam/identity/istokenvalid?tokenid='
+ validTokenAPI = amURL + 'openam/json/sessions/{token}?_action=validate'
loginURL = amURL + 'openam/UI/Login'
def session_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
usercookie = request.cookies.get(cookiename)
if usercookie:
- amQuery = requests.get(validTokenAPI + usercookie)
+ amQuery = requests.post(validTokenAPI.format(token=usercookie))
- if 'boolean=true' in amQuery.text:
+ if amQuery.json()['valid']:
return f(*args, **kwargs)
return redirect(loginURL)
return decorated_function
@app.route('/members_page')
@session_required
def members_page():
pass
| Use new OpenAM token validation endpoint | ## Code Before:
from flask import request, redirect
import requests
cookiename = 'openAMUserCookieName'
amURL = 'https://openam.example.com/'
validTokenAPI = amURL + 'openam/identity/istokenvalid?tokenid='
loginURL = amURL + 'openam/UI/Login'
def session_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
usercookie = request.cookies.get(cookiename)
if usercookie:
amQuery = requests.get(validTokenAPI + usercookie)
if 'boolean=true' in amQuery.text:
return f(*args, **kwargs)
return redirect(loginURL)
return decorated_function
@app.route('/members_page')
@session_required
def members_page():
pass
## Instruction:
Use new OpenAM token validation endpoint
## Code After:
from flask import request, redirect
import requests
cookiename = 'openAMUserCookieName'
amURL = 'https://openam.example.com/'
validTokenAPI = amURL + 'openam/json/sessions/{token}?_action=validate'
loginURL = amURL + 'openam/UI/Login'
def session_required(f):
@wraps(f)
def decorated_function(*args, **kwargs):
usercookie = request.cookies.get(cookiename)
if usercookie:
amQuery = requests.post(validTokenAPI.format(token=usercookie))
if amQuery.json()['valid']:
return f(*args, **kwargs)
return redirect(loginURL)
return decorated_function
@app.route('/members_page')
@session_required
def members_page():
pass
|
463abcce738ca1c47729cc0e465da9dc399e21dd | examples/remote_download.py | examples/remote_download.py |
from xunleipy.remote import XunLeiRemote
def remote_download(username, password, rk_username, rk_password, download_links, proxy=None, path='C:/TD/', peer=0):
remote_client = XunLeiRemote(username, password, rk_username, rk_password, proxy=proxy)
remote_client.login()
peer_list = remote_client.get_remote_peer_list()
if len(peer_list) == 0:
print 'No valid remote devices'
return
pid = peer_list[peer]['pid']
return remote_client.add_urls_to_remote(pid, path, download_links)
if __name__ == '__main__':
import sys
download_link = sys.argv[1]
with open('config.json', 'r') as f:
import json
config = json.load(f)
username = config.get('username', '')
password = config.get('password', '')
rk_username = config.get('rk_username', '')
rk_password = config.get('rk_password', '')
proxy = config.get('proxy', None)
if not username or not password:
print 'Invalid username or password!'
else:
path = config.get('path', 'C:/TDDOWNLOAD/')
print remote_download(username, password, rk_username, rk_password, [download_link], proxy)
| import sys
import os
from xunleipy.remote import XunLeiRemote
sys.path.append('/Users/gunner/workspace/xunleipy')
def remote_download(username,
password,
rk_username,
rk_password,
download_links,
proxy=None,
path='C:/TD/',
peer=0):
remote_client = XunLeiRemote(
username, password, rk_username, rk_password, proxy=proxy
)
remote_client.login()
peer_list = remote_client.get_remote_peer_list()
if len(peer_list) == 0:
print('No valid remote devices')
return
pid = peer_list[peer]['pid']
return remote_client.add_urls_to_remote(pid, path, download_links)
if __name__ == '__main__':
import sys
download_link = sys.argv[1]
with open('config.json', 'r') as f:
import json
config = json.load(f)
username = config.get('username', '')
password = config.get('password', '')
rk_username = config.get('rk_username', '')
rk_password = config.get('rk_password', '')
proxy = config.get('proxy', None)
if not username or not password:
print('Invalid username or password!')
else:
path = config.get('path', 'C:/TDDOWNLOAD/')
print(
remote_download(
username, password, rk_username,
rk_password, [download_link], proxy
)
)
| Change example style for python3 | Change example style for python3
| Python | mit | lazygunner/xunleipy | + import sys
+ import os
from xunleipy.remote import XunLeiRemote
+ sys.path.append('/Users/gunner/workspace/xunleipy')
- def remote_download(username, password, rk_username, rk_password, download_links, proxy=None, path='C:/TD/', peer=0):
+ def remote_download(username,
+ password,
+ rk_username,
+ rk_password,
+ download_links,
+ proxy=None,
+ path='C:/TD/',
+ peer=0):
+ remote_client = XunLeiRemote(
- remote_client = XunLeiRemote(username, password, rk_username, rk_password, proxy=proxy)
+ username, password, rk_username, rk_password, proxy=proxy
+ )
remote_client.login()
peer_list = remote_client.get_remote_peer_list()
if len(peer_list) == 0:
- print 'No valid remote devices'
+ print('No valid remote devices')
return
pid = peer_list[peer]['pid']
return remote_client.add_urls_to_remote(pid, path, download_links)
if __name__ == '__main__':
import sys
download_link = sys.argv[1]
with open('config.json', 'r') as f:
import json
config = json.load(f)
username = config.get('username', '')
password = config.get('password', '')
rk_username = config.get('rk_username', '')
rk_password = config.get('rk_password', '')
proxy = config.get('proxy', None)
if not username or not password:
- print 'Invalid username or password!'
+ print('Invalid username or password!')
else:
path = config.get('path', 'C:/TDDOWNLOAD/')
- print remote_download(username, password, rk_username, rk_password, [download_link], proxy)
+ print(
+ remote_download(
+ username, password, rk_username,
+ rk_password, [download_link], proxy
+ )
+ )
| Change example style for python3 | ## Code Before:
from xunleipy.remote import XunLeiRemote
def remote_download(username, password, rk_username, rk_password, download_links, proxy=None, path='C:/TD/', peer=0):
remote_client = XunLeiRemote(username, password, rk_username, rk_password, proxy=proxy)
remote_client.login()
peer_list = remote_client.get_remote_peer_list()
if len(peer_list) == 0:
print 'No valid remote devices'
return
pid = peer_list[peer]['pid']
return remote_client.add_urls_to_remote(pid, path, download_links)
if __name__ == '__main__':
import sys
download_link = sys.argv[1]
with open('config.json', 'r') as f:
import json
config = json.load(f)
username = config.get('username', '')
password = config.get('password', '')
rk_username = config.get('rk_username', '')
rk_password = config.get('rk_password', '')
proxy = config.get('proxy', None)
if not username or not password:
print 'Invalid username or password!'
else:
path = config.get('path', 'C:/TDDOWNLOAD/')
print remote_download(username, password, rk_username, rk_password, [download_link], proxy)
## Instruction:
Change example style for python3
## Code After:
import sys
import os
from xunleipy.remote import XunLeiRemote
sys.path.append('/Users/gunner/workspace/xunleipy')
def remote_download(username,
password,
rk_username,
rk_password,
download_links,
proxy=None,
path='C:/TD/',
peer=0):
remote_client = XunLeiRemote(
username, password, rk_username, rk_password, proxy=proxy
)
remote_client.login()
peer_list = remote_client.get_remote_peer_list()
if len(peer_list) == 0:
print('No valid remote devices')
return
pid = peer_list[peer]['pid']
return remote_client.add_urls_to_remote(pid, path, download_links)
if __name__ == '__main__':
import sys
download_link = sys.argv[1]
with open('config.json', 'r') as f:
import json
config = json.load(f)
username = config.get('username', '')
password = config.get('password', '')
rk_username = config.get('rk_username', '')
rk_password = config.get('rk_password', '')
proxy = config.get('proxy', None)
if not username or not password:
print('Invalid username or password!')
else:
path = config.get('path', 'C:/TDDOWNLOAD/')
print(
remote_download(
username, password, rk_username,
rk_password, [download_link], proxy
)
)
|
ab47c678b37527a7b8a970b365503b65ffccda87 | populous/cli.py | populous/cli.py | import click
from .loader import load_yaml
from .blueprint import Blueprint
from .exceptions import ValidationError, YAMLError
def get_blueprint(*files):
try:
return Blueprint.from_description(load_yaml(*files))
except (YAMLError, ValidationError) as e:
raise click.ClickException(e.message)
except Exception as e:
pass
@click.group()
@click.version_option()
def cli():
pass
@cli.command()
@click.argument('files', nargs=-1)
def predict(files):
"""
Predict how many objects will be created if the given files are used.
"""
blueprint = get_blueprint(*files)
for item in blueprint:
click.echo("{name}: {count} {by}".format(
name=item.name, count=item.total,
by="({} by {})".format(item.count.number, item.count.by)
if item.count.by else ""
))
| import click
from .loader import load_yaml
from .blueprint import Blueprint
from .exceptions import ValidationError, YAMLError
def get_blueprint(*files):
try:
return Blueprint.from_description(load_yaml(*files))
except (YAMLError, ValidationError) as e:
raise click.ClickException(e.message)
except Exception as e:
raise click.ClickException("Unexpected error during the blueprint "
"loading: {}".format(e.message))
@click.group()
@click.version_option()
def cli():
pass
@cli.command()
@click.argument('files', nargs=-1)
def predict(files):
"""
Predict how many objects will be created if the given files are used.
"""
blueprint = get_blueprint(*files)
for item in blueprint:
click.echo("{name}: {count} {by}".format(
name=item.name, count=item.total,
by="({} by {})".format(item.count.number, item.count.by)
if item.count.by else ""
))
| Handle unexpected errors properly in load_blueprint | Handle unexpected errors properly in load_blueprint
| Python | mit | novafloss/populous | import click
from .loader import load_yaml
from .blueprint import Blueprint
from .exceptions import ValidationError, YAMLError
def get_blueprint(*files):
try:
return Blueprint.from_description(load_yaml(*files))
except (YAMLError, ValidationError) as e:
raise click.ClickException(e.message)
except Exception as e:
- pass
+ raise click.ClickException("Unexpected error during the blueprint "
+ "loading: {}".format(e.message))
@click.group()
@click.version_option()
def cli():
pass
@cli.command()
@click.argument('files', nargs=-1)
def predict(files):
"""
Predict how many objects will be created if the given files are used.
"""
blueprint = get_blueprint(*files)
for item in blueprint:
click.echo("{name}: {count} {by}".format(
name=item.name, count=item.total,
by="({} by {})".format(item.count.number, item.count.by)
if item.count.by else ""
))
| Handle unexpected errors properly in load_blueprint | ## Code Before:
import click
from .loader import load_yaml
from .blueprint import Blueprint
from .exceptions import ValidationError, YAMLError
def get_blueprint(*files):
try:
return Blueprint.from_description(load_yaml(*files))
except (YAMLError, ValidationError) as e:
raise click.ClickException(e.message)
except Exception as e:
pass
@click.group()
@click.version_option()
def cli():
pass
@cli.command()
@click.argument('files', nargs=-1)
def predict(files):
"""
Predict how many objects will be created if the given files are used.
"""
blueprint = get_blueprint(*files)
for item in blueprint:
click.echo("{name}: {count} {by}".format(
name=item.name, count=item.total,
by="({} by {})".format(item.count.number, item.count.by)
if item.count.by else ""
))
## Instruction:
Handle unexpected errors properly in load_blueprint
## Code After:
import click
from .loader import load_yaml
from .blueprint import Blueprint
from .exceptions import ValidationError, YAMLError
def get_blueprint(*files):
try:
return Blueprint.from_description(load_yaml(*files))
except (YAMLError, ValidationError) as e:
raise click.ClickException(e.message)
except Exception as e:
raise click.ClickException("Unexpected error during the blueprint "
"loading: {}".format(e.message))
@click.group()
@click.version_option()
def cli():
pass
@cli.command()
@click.argument('files', nargs=-1)
def predict(files):
"""
Predict how many objects will be created if the given files are used.
"""
blueprint = get_blueprint(*files)
for item in blueprint:
click.echo("{name}: {count} {by}".format(
name=item.name, count=item.total,
by="({} by {})".format(item.count.number, item.count.by)
if item.count.by else ""
))
|
0485e6dcaf19061812d0e571890e58b85b5dea12 | lava_results_app/utils.py | lava_results_app/utils.py | import os
import yaml
import logging
from django.utils.translation import ungettext_lazy
from django.conf import settings
def help_max_length(max_length):
return ungettext_lazy(
u"Maximum length: {0} character",
u"Maximum length: {0} characters",
max_length).format(max_length)
class StreamEcho(object):
def write(self, value):
return value
def description_filename(job_id):
logger = logging.getLogger('lava_results_app')
filename = os.path.join(settings.MEDIA_ROOT, 'job-output', 'job-%s' % job_id, 'description.yaml')
if not os.path.exists(filename):
logger.error("No description.yaml for job %s" % job_id)
return None
return filename
def description_data(job_id):
logger = logging.getLogger('lava_results_app')
filename = description_filename(job_id)
if not filename:
return {}
try:
data = yaml.load(open(filename, 'r'))
except yaml.YAMLError:
logger.error("Unable to parse description for %s" % job_id)
return {}
return data
| import os
import yaml
import logging
from django.utils.translation import ungettext_lazy
from django.conf import settings
def help_max_length(max_length):
return ungettext_lazy(
u"Maximum length: {0} character",
u"Maximum length: {0} characters",
max_length).format(max_length)
class StreamEcho(object):
def write(self, value):
return value
def description_filename(job_id):
logger = logging.getLogger('lava_results_app')
filename = os.path.join(settings.MEDIA_ROOT, 'job-output', 'job-%s' % job_id, 'description.yaml')
if not os.path.exists(filename):
logger.error("No description.yaml for job %s" % job_id)
return None
return filename
def description_data(job_id):
logger = logging.getLogger('lava_results_app')
filename = description_filename(job_id)
if not filename:
return {}
try:
data = yaml.load(open(filename, 'r'))
except yaml.YAMLError:
logger.error("Unable to parse description for %s" % job_id)
return {}
if not data:
return {}
return data
| Return an empty dict if no data | Return an empty dict if no data
Avoids a HTTP500 on slow instances where the file
may be created before data is written, causing the
YAML parser to return None.
Change-Id: I13b92941f3e368839a9665fe3197c706babd9335
| Python | agpl-3.0 | Linaro/lava-server,Linaro/lava-server,Linaro/lava-server,Linaro/lava-server | import os
import yaml
import logging
from django.utils.translation import ungettext_lazy
from django.conf import settings
def help_max_length(max_length):
return ungettext_lazy(
u"Maximum length: {0} character",
u"Maximum length: {0} characters",
max_length).format(max_length)
class StreamEcho(object):
def write(self, value):
return value
def description_filename(job_id):
logger = logging.getLogger('lava_results_app')
filename = os.path.join(settings.MEDIA_ROOT, 'job-output', 'job-%s' % job_id, 'description.yaml')
if not os.path.exists(filename):
logger.error("No description.yaml for job %s" % job_id)
return None
return filename
def description_data(job_id):
logger = logging.getLogger('lava_results_app')
filename = description_filename(job_id)
if not filename:
return {}
try:
data = yaml.load(open(filename, 'r'))
except yaml.YAMLError:
logger.error("Unable to parse description for %s" % job_id)
return {}
+ if not data:
+ return {}
return data
| Return an empty dict if no data | ## Code Before:
import os
import yaml
import logging
from django.utils.translation import ungettext_lazy
from django.conf import settings
def help_max_length(max_length):
return ungettext_lazy(
u"Maximum length: {0} character",
u"Maximum length: {0} characters",
max_length).format(max_length)
class StreamEcho(object):
def write(self, value):
return value
def description_filename(job_id):
logger = logging.getLogger('lava_results_app')
filename = os.path.join(settings.MEDIA_ROOT, 'job-output', 'job-%s' % job_id, 'description.yaml')
if not os.path.exists(filename):
logger.error("No description.yaml for job %s" % job_id)
return None
return filename
def description_data(job_id):
logger = logging.getLogger('lava_results_app')
filename = description_filename(job_id)
if not filename:
return {}
try:
data = yaml.load(open(filename, 'r'))
except yaml.YAMLError:
logger.error("Unable to parse description for %s" % job_id)
return {}
return data
## Instruction:
Return an empty dict if no data
## Code After:
import os
import yaml
import logging
from django.utils.translation import ungettext_lazy
from django.conf import settings
def help_max_length(max_length):
return ungettext_lazy(
u"Maximum length: {0} character",
u"Maximum length: {0} characters",
max_length).format(max_length)
class StreamEcho(object):
def write(self, value):
return value
def description_filename(job_id):
logger = logging.getLogger('lava_results_app')
filename = os.path.join(settings.MEDIA_ROOT, 'job-output', 'job-%s' % job_id, 'description.yaml')
if not os.path.exists(filename):
logger.error("No description.yaml for job %s" % job_id)
return None
return filename
def description_data(job_id):
logger = logging.getLogger('lava_results_app')
filename = description_filename(job_id)
if not filename:
return {}
try:
data = yaml.load(open(filename, 'r'))
except yaml.YAMLError:
logger.error("Unable to parse description for %s" % job_id)
return {}
if not data:
return {}
return data
|
28c6af1381a1fc38b20ce05e85f494f3ae2beeb4 | arcutils/masquerade/templatetags/masquerade.py | arcutils/masquerade/templatetags/masquerade.py | from django import template
from .. import perms
from ..settings import get_user_attr
register = template.Library()
@register.filter
def is_masquerading(user):
info = getattr(user, get_user_attr())
return info['is_masquerading']
@register.filter
def can_masquerade(user):
return perms.can_masquerade(user)
@register.filter
def can_masquerade_as(user, masquerade_user):
return perms.can_masquerade_as(user, masquerade_user)
| from django import template
from .. import perms
from ..settings import get_user_attr, is_enabled
register = template.Library()
@register.filter
def is_masquerading(user):
if not is_enabled():
return False
info = getattr(user, get_user_attr(), None)
return info['is_masquerading']
@register.filter
def can_masquerade(user):
return perms.can_masquerade(user)
@register.filter
def can_masquerade_as(user, masquerade_user):
return perms.can_masquerade_as(user, masquerade_user)
| Make is_masquerading template tag more robust | Make is_masquerading template tag more robust
When masquerading is not enabled, immediately return False to avoid
checking for a request attribute that won't be present.
| Python | mit | PSU-OIT-ARC/django-arcutils,wylee/django-arcutils,wylee/django-arcutils,PSU-OIT-ARC/django-arcutils | from django import template
from .. import perms
- from ..settings import get_user_attr
+ from ..settings import get_user_attr, is_enabled
register = template.Library()
@register.filter
def is_masquerading(user):
+ if not is_enabled():
+ return False
- info = getattr(user, get_user_attr())
+ info = getattr(user, get_user_attr(), None)
return info['is_masquerading']
@register.filter
def can_masquerade(user):
return perms.can_masquerade(user)
@register.filter
def can_masquerade_as(user, masquerade_user):
return perms.can_masquerade_as(user, masquerade_user)
| Make is_masquerading template tag more robust | ## Code Before:
from django import template
from .. import perms
from ..settings import get_user_attr
register = template.Library()
@register.filter
def is_masquerading(user):
info = getattr(user, get_user_attr())
return info['is_masquerading']
@register.filter
def can_masquerade(user):
return perms.can_masquerade(user)
@register.filter
def can_masquerade_as(user, masquerade_user):
return perms.can_masquerade_as(user, masquerade_user)
## Instruction:
Make is_masquerading template tag more robust
## Code After:
from django import template
from .. import perms
from ..settings import get_user_attr, is_enabled
register = template.Library()
@register.filter
def is_masquerading(user):
if not is_enabled():
return False
info = getattr(user, get_user_attr(), None)
return info['is_masquerading']
@register.filter
def can_masquerade(user):
return perms.can_masquerade(user)
@register.filter
def can_masquerade_as(user, masquerade_user):
return perms.can_masquerade_as(user, masquerade_user)
|
263e517004df36938b430d8802d4fc80067fadf5 | djangoreact/urls.py | djangoreact/urls.py | from django.conf.urls import url, include
from django.contrib import admin
from rest_framework import routers
from server import views
router = routers.DefaultRouter()
router.register(r'users', views.UserViewSet)
router.register(r'groups', views.GroupViewSet)
urlpatterns = [
url(r'^$', views.index),
url(r'^api/auth/', include('rest_auth.urls')),
url(r'^api/', include(router.urls)),
url(r'^api-auth/', include('rest_framework.urls')),
url(r'^admin/', admin.site.urls),
]
| from django.conf.urls import url, include
from django.contrib import admin
from rest_framework import routers
from server import views
router = routers.DefaultRouter()
router.register(r'users', views.UserViewSet)
router.register(r'groups', views.GroupViewSet)
urlpatterns = [
url(r'^api/auth/', include('rest_auth.urls')),
url(r'^api/', include(router.urls)),
url(r'^api-auth/', include('rest_framework.urls')),
url(r'^admin/', admin.site.urls),
url(r'^', views.index),
]
| Fix to use react-router for all unmatched routes. | Fix to use react-router for all unmatched routes.
| Python | mit | willy-claes/django-react,willy-claes/django-react,willy-claes/django-react | from django.conf.urls import url, include
from django.contrib import admin
from rest_framework import routers
from server import views
router = routers.DefaultRouter()
router.register(r'users', views.UserViewSet)
router.register(r'groups', views.GroupViewSet)
urlpatterns = [
- url(r'^$', views.index),
url(r'^api/auth/', include('rest_auth.urls')),
url(r'^api/', include(router.urls)),
url(r'^api-auth/', include('rest_framework.urls')),
url(r'^admin/', admin.site.urls),
+ url(r'^', views.index),
]
| Fix to use react-router for all unmatched routes. | ## Code Before:
from django.conf.urls import url, include
from django.contrib import admin
from rest_framework import routers
from server import views
router = routers.DefaultRouter()
router.register(r'users', views.UserViewSet)
router.register(r'groups', views.GroupViewSet)
urlpatterns = [
url(r'^$', views.index),
url(r'^api/auth/', include('rest_auth.urls')),
url(r'^api/', include(router.urls)),
url(r'^api-auth/', include('rest_framework.urls')),
url(r'^admin/', admin.site.urls),
]
## Instruction:
Fix to use react-router for all unmatched routes.
## Code After:
from django.conf.urls import url, include
from django.contrib import admin
from rest_framework import routers
from server import views
router = routers.DefaultRouter()
router.register(r'users', views.UserViewSet)
router.register(r'groups', views.GroupViewSet)
urlpatterns = [
url(r'^api/auth/', include('rest_auth.urls')),
url(r'^api/', include(router.urls)),
url(r'^api-auth/', include('rest_framework.urls')),
url(r'^admin/', admin.site.urls),
url(r'^', views.index),
]
|
f83282b1747e255d35e18e9fecad1750d1564f9e | do_record/record.py | do_record/record.py | """DigitalOcean DNS Records."""
from certbot_dns_auth.printer import printer
from do_record import http
class Record(object):
"""Handle DigitalOcean DNS records."""
def __init__(self, api_key, domain, hostname):
self._number = None
self.domain = domain
self.hostname = hostname
self.api_key = api_key
def create(self, value):
"""Create this record on DigitalOcean with the supplied value."""
self._number = http.create(self, value)
return self.number
def delete(self, record_id=None):
"""Delete this record on DigitalOcean, identified by record_id."""
if record_id is None:
record_id = self.number
http.delete(self, record_id)
def printer(self):
printer(self.number)
@property
def number(self):
return self._number
@number.setter
def number(self, value):
if self.number is None:
self._number = value
else:
raise ValueError(
'Cannot externally reset a record\'s number identifier.')
| """DigitalOcean DNS Records."""
from certbot_dns_auth.printer import printer
from do_record import http
class Record(object):
"""Handle DigitalOcean DNS records."""
def __init__(self, api_key, domain, hostname):
self._number = None
self.domain = domain
self.hostname = hostname
self.api_key = api_key
def create(self, value):
"""Create this record on DigitalOcean with the supplied value."""
self._number = http.create(self, value)
return self.number
def delete(self, record_id=None):
"""Delete this record on DigitalOcean, identified by record_id."""
if record_id is None:
record_id = self.number
http.delete(self, record_id)
def printer(self):
printer(self.number)
@property
def number(self):
return self._number
@number.setter
def number(self, value):
self._number = value
| Remove Code That Doesn't Have a Test | Remove Code That Doesn't Have a Test
| Python | apache-2.0 | Jitsusama/lets-do-dns | """DigitalOcean DNS Records."""
from certbot_dns_auth.printer import printer
from do_record import http
class Record(object):
"""Handle DigitalOcean DNS records."""
def __init__(self, api_key, domain, hostname):
self._number = None
self.domain = domain
self.hostname = hostname
self.api_key = api_key
def create(self, value):
"""Create this record on DigitalOcean with the supplied value."""
self._number = http.create(self, value)
return self.number
def delete(self, record_id=None):
"""Delete this record on DigitalOcean, identified by record_id."""
if record_id is None:
record_id = self.number
http.delete(self, record_id)
def printer(self):
printer(self.number)
@property
def number(self):
return self._number
@number.setter
def number(self, value):
- if self.number is None:
- self._number = value
+ self._number = value
- else:
- raise ValueError(
- 'Cannot externally reset a record\'s number identifier.')
| Remove Code That Doesn't Have a Test | ## Code Before:
"""DigitalOcean DNS Records."""
from certbot_dns_auth.printer import printer
from do_record import http
class Record(object):
"""Handle DigitalOcean DNS records."""
def __init__(self, api_key, domain, hostname):
self._number = None
self.domain = domain
self.hostname = hostname
self.api_key = api_key
def create(self, value):
"""Create this record on DigitalOcean with the supplied value."""
self._number = http.create(self, value)
return self.number
def delete(self, record_id=None):
"""Delete this record on DigitalOcean, identified by record_id."""
if record_id is None:
record_id = self.number
http.delete(self, record_id)
def printer(self):
printer(self.number)
@property
def number(self):
return self._number
@number.setter
def number(self, value):
if self.number is None:
self._number = value
else:
raise ValueError(
'Cannot externally reset a record\'s number identifier.')
## Instruction:
Remove Code That Doesn't Have a Test
## Code After:
"""DigitalOcean DNS Records."""
from certbot_dns_auth.printer import printer
from do_record import http
class Record(object):
"""Handle DigitalOcean DNS records."""
def __init__(self, api_key, domain, hostname):
self._number = None
self.domain = domain
self.hostname = hostname
self.api_key = api_key
def create(self, value):
"""Create this record on DigitalOcean with the supplied value."""
self._number = http.create(self, value)
return self.number
def delete(self, record_id=None):
"""Delete this record on DigitalOcean, identified by record_id."""
if record_id is None:
record_id = self.number
http.delete(self, record_id)
def printer(self):
printer(self.number)
@property
def number(self):
return self._number
@number.setter
def number(self, value):
self._number = value
|
413c3e9e8a093e3f336e27a663f347f5ea9866a6 | performanceplatform/collector/ga/__init__.py | performanceplatform/collector/ga/__init__.py | from pkgutil import extend_path
__path__ = extend_path(__path__, __name__)
from performanceplatform.collector.ga.core \
import create_client, query_documents_for, send_data
from performanceplatform.collector.write import DataSet
def main(credentials, data_set_config, query, options, start_at, end_at):
client = create_client(credentials)
documents = query_documents_for(
client, query, options,
data_set_config['data-type'], start_at, end_at
)
data_set = DataSet.from_config(data_set_config)
send_data(data_set, documents)
| from pkgutil import extend_path
__path__ = extend_path(__path__, __name__)
from performanceplatform.collector.ga.core \
import create_client, query_documents_for, send_data
from performanceplatform.collector.write import DataSet
def main(credentials, data_set_config, query, options, start_at, end_at):
client = create_client(credentials)
documents = query_documents_for(
client, query, options,
options.get('dataType', data_set_config['data-type']),
start_at, end_at)
data_set = DataSet.from_config(data_set_config)
send_data(data_set, documents)
| Allow the 'dataType' field to be overriden | Allow the 'dataType' field to be overriden
The 'dataType' field in records predates data groups and data types. As
such they don't always match the new world order of data types. It's
fine to change in all cases other than Licensing which is run on
limelight, that we don't really want to touch.
| Python | mit | alphagov/performanceplatform-collector,alphagov/performanceplatform-collector,alphagov/performanceplatform-collector | from pkgutil import extend_path
__path__ = extend_path(__path__, __name__)
from performanceplatform.collector.ga.core \
import create_client, query_documents_for, send_data
from performanceplatform.collector.write import DataSet
def main(credentials, data_set_config, query, options, start_at, end_at):
client = create_client(credentials)
documents = query_documents_for(
client, query, options,
- data_set_config['data-type'], start_at, end_at
- )
+ options.get('dataType', data_set_config['data-type']),
+ start_at, end_at)
data_set = DataSet.from_config(data_set_config)
send_data(data_set, documents)
| Allow the 'dataType' field to be overriden | ## Code Before:
from pkgutil import extend_path
__path__ = extend_path(__path__, __name__)
from performanceplatform.collector.ga.core \
import create_client, query_documents_for, send_data
from performanceplatform.collector.write import DataSet
def main(credentials, data_set_config, query, options, start_at, end_at):
client = create_client(credentials)
documents = query_documents_for(
client, query, options,
data_set_config['data-type'], start_at, end_at
)
data_set = DataSet.from_config(data_set_config)
send_data(data_set, documents)
## Instruction:
Allow the 'dataType' field to be overriden
## Code After:
from pkgutil import extend_path
__path__ = extend_path(__path__, __name__)
from performanceplatform.collector.ga.core \
import create_client, query_documents_for, send_data
from performanceplatform.collector.write import DataSet
def main(credentials, data_set_config, query, options, start_at, end_at):
client = create_client(credentials)
documents = query_documents_for(
client, query, options,
options.get('dataType', data_set_config['data-type']),
start_at, end_at)
data_set = DataSet.from_config(data_set_config)
send_data(data_set, documents)
|
95542ab1b7c22a6e0160e242349c66f2cef7e390 | syntacticframes_project/syntacticframes/management/commands/check_correspondance_errors.py | syntacticframes_project/syntacticframes/management/commands/check_correspondance_errors.py | from django.core.management.base import BaseCommand
from syntacticframes.models import VerbNetClass
from parsecorrespondance import parse
from loadmapping import mapping
class Command(BaseCommand):
def handle(self, *args, **options):
for vn_class in VerbNetClass.objects.all():
try:
parse.get_ladl_list(vn_class.ladl_string)
except parse.UnknownClassException as e:
print('{:<30} {}'.format(vn_class.name, e))
try:
parse.get_lvf_list(vn_class.lvf_string)
except parse.UnknownClassException as e:
print('{:<30} {}'.format(vn_class.name, e))
| from django.core.management.base import BaseCommand
from syntacticframes.models import VerbNetFrameSet
from parsecorrespondance import parse
from loadmapping import mapping
class Command(BaseCommand):
def handle(self, *args, **options):
for frameset in VerbNetFrameSet.objects.all():
print("{}: {}/{}".format(frameset.name, frameset.ladl_string, frameset.lvf_string))
if frameset.ladl_string:
try:
parse.FrenchMapping('LADL', frameset.ladl_string).result()
except parse.UnknownClassException as e:
print('{:<30} {}'.format(frameset.name, e))
if frameset.lvf_string:
try:
parse.FrenchMapping('LVF', frameset.lvf_string)
except parse.UnknownClassException as e:
print('{:<30} {}'.format(frameset.name, e))
| Check correspondances in framesets now | Check correspondances in framesets now
| Python | mit | aymara/verbenet-editor,aymara/verbenet-editor,aymara/verbenet-editor | from django.core.management.base import BaseCommand
- from syntacticframes.models import VerbNetClass
+ from syntacticframes.models import VerbNetFrameSet
from parsecorrespondance import parse
from loadmapping import mapping
class Command(BaseCommand):
def handle(self, *args, **options):
- for vn_class in VerbNetClass.objects.all():
+ for frameset in VerbNetFrameSet.objects.all():
+ print("{}: {}/{}".format(frameset.name, frameset.ladl_string, frameset.lvf_string))
- try:
- parse.get_ladl_list(vn_class.ladl_string)
- except parse.UnknownClassException as e:
- print('{:<30} {}'.format(vn_class.name, e))
+ if frameset.ladl_string:
- try:
+ try:
- parse.get_lvf_list(vn_class.lvf_string)
+ parse.FrenchMapping('LADL', frameset.ladl_string).result()
- except parse.UnknownClassException as e:
+ except parse.UnknownClassException as e:
- print('{:<30} {}'.format(vn_class.name, e))
+ print('{:<30} {}'.format(frameset.name, e))
+ if frameset.lvf_string:
+ try:
+ parse.FrenchMapping('LVF', frameset.lvf_string)
+ except parse.UnknownClassException as e:
+ print('{:<30} {}'.format(frameset.name, e))
+ | Check correspondances in framesets now | ## Code Before:
from django.core.management.base import BaseCommand
from syntacticframes.models import VerbNetClass
from parsecorrespondance import parse
from loadmapping import mapping
class Command(BaseCommand):
def handle(self, *args, **options):
for vn_class in VerbNetClass.objects.all():
try:
parse.get_ladl_list(vn_class.ladl_string)
except parse.UnknownClassException as e:
print('{:<30} {}'.format(vn_class.name, e))
try:
parse.get_lvf_list(vn_class.lvf_string)
except parse.UnknownClassException as e:
print('{:<30} {}'.format(vn_class.name, e))
## Instruction:
Check correspondances in framesets now
## Code After:
from django.core.management.base import BaseCommand
from syntacticframes.models import VerbNetFrameSet
from parsecorrespondance import parse
from loadmapping import mapping
class Command(BaseCommand):
def handle(self, *args, **options):
for frameset in VerbNetFrameSet.objects.all():
print("{}: {}/{}".format(frameset.name, frameset.ladl_string, frameset.lvf_string))
if frameset.ladl_string:
try:
parse.FrenchMapping('LADL', frameset.ladl_string).result()
except parse.UnknownClassException as e:
print('{:<30} {}'.format(frameset.name, e))
if frameset.lvf_string:
try:
parse.FrenchMapping('LVF', frameset.lvf_string)
except parse.UnknownClassException as e:
print('{:<30} {}'.format(frameset.name, e))
|
6c54fc230e8c889a2351f20b524382a5c6e29d1c | examples/apps.py | examples/apps.py | import os
import sys
from pysuru import TsuruClient
TSURU_TARGET = os.environ.get('TSURU_TARGET', None)
TSURU_TOKEN = os.environ.get('TSURU_TOKEN', None)
if not TSURU_TARGET or not TSURU_TOKEN:
print('You must set TSURU_TARGET and TSURU_TOKEN.')
sys.exit(1)
api = TsuruClient(TSURU_TARGET, TSURU_TOKEN)
# List all apps that this token has access to
for app in api.apps:
print(app.name)
# Update one specific app
api.apps.update('my-awesome-app', {'description': 'My awesome app'})
# Get information for one app
app = App.get('my-awesome-app')
print('%s: %s' % (app.name, app.description))
# List all services instances for app
for service in app.services:
print('Service: %s' % service.name)
| import os
import sys
from pysuru import TsuruClient
TSURU_TARGET = os.environ.get('TSURU_TARGET', None)
TSURU_TOKEN = os.environ.get('TSURU_TOKEN', None)
if not TSURU_TARGET or not TSURU_TOKEN:
print('You must set TSURU_TARGET and TSURU_TOKEN env variables.')
sys.exit(1)
# Creating TsuruClient instance
tsuru = TsuruClient(TSURU_TARGET, TSURU_TOKEN)
# List all apps that this user has access to
for app in tsuru.apps.list():
print('App: {}'.format(app.name))
# Get information for one app
app = tsuru.apps.get('my-awesome-app')
print('{app.name}: {app.description}'.format(app=app))
# Update specific app
tsuru.apps.update('my-awesome-app', {'description': 'My new awesome description'})
| Update examples to match docs | Update examples to match docs
Use the interface defined in the docs in the examples scripts.
| Python | mit | rcmachado/pysuru | import os
import sys
from pysuru import TsuruClient
TSURU_TARGET = os.environ.get('TSURU_TARGET', None)
TSURU_TOKEN = os.environ.get('TSURU_TOKEN', None)
if not TSURU_TARGET or not TSURU_TOKEN:
- print('You must set TSURU_TARGET and TSURU_TOKEN.')
+ print('You must set TSURU_TARGET and TSURU_TOKEN env variables.')
sys.exit(1)
+ # Creating TsuruClient instance
- api = TsuruClient(TSURU_TARGET, TSURU_TOKEN)
+ tsuru = TsuruClient(TSURU_TARGET, TSURU_TOKEN)
- # List all apps that this token has access to
+ # List all apps that this user has access to
+ for app in tsuru.apps.list():
+ print('App: {}'.format(app.name))
- for app in api.apps:
- print(app.name)
-
- # Update one specific app
- api.apps.update('my-awesome-app', {'description': 'My awesome app'})
# Get information for one app
- app = App.get('my-awesome-app')
+ app = tsuru.apps.get('my-awesome-app')
- print('%s: %s' % (app.name, app.description))
+ print('{app.name}: {app.description}'.format(app=app))
+ # Update specific app
+ tsuru.apps.update('my-awesome-app', {'description': 'My new awesome description'})
- # List all services instances for app
- for service in app.services:
- print('Service: %s' % service.name)
| Update examples to match docs | ## Code Before:
import os
import sys
from pysuru import TsuruClient
TSURU_TARGET = os.environ.get('TSURU_TARGET', None)
TSURU_TOKEN = os.environ.get('TSURU_TOKEN', None)
if not TSURU_TARGET or not TSURU_TOKEN:
print('You must set TSURU_TARGET and TSURU_TOKEN.')
sys.exit(1)
api = TsuruClient(TSURU_TARGET, TSURU_TOKEN)
# List all apps that this token has access to
for app in api.apps:
print(app.name)
# Update one specific app
api.apps.update('my-awesome-app', {'description': 'My awesome app'})
# Get information for one app
app = App.get('my-awesome-app')
print('%s: %s' % (app.name, app.description))
# List all services instances for app
for service in app.services:
print('Service: %s' % service.name)
## Instruction:
Update examples to match docs
## Code After:
import os
import sys
from pysuru import TsuruClient
TSURU_TARGET = os.environ.get('TSURU_TARGET', None)
TSURU_TOKEN = os.environ.get('TSURU_TOKEN', None)
if not TSURU_TARGET or not TSURU_TOKEN:
print('You must set TSURU_TARGET and TSURU_TOKEN env variables.')
sys.exit(1)
# Creating TsuruClient instance
tsuru = TsuruClient(TSURU_TARGET, TSURU_TOKEN)
# List all apps that this user has access to
for app in tsuru.apps.list():
print('App: {}'.format(app.name))
# Get information for one app
app = tsuru.apps.get('my-awesome-app')
print('{app.name}: {app.description}'.format(app=app))
# Update specific app
tsuru.apps.update('my-awesome-app', {'description': 'My new awesome description'})
|
b5a8e7b6926bf7224abed6bd335d62b3f1ad1fb1 | performance_testing/command_line.py | performance_testing/command_line.py | import os
import yaml
from performance_testing.errors import ConfigFileError, ConfigKeyError
from performance_testing import web
from datetime import datetime as date
from time import time
class Tool:
def __init__(self, config='config.yml', result_directory='result'):
self.read_config(config_file=config)
self.create_result_file(directory=result_directory)
def read_config(self, config_file):
try:
config_stream = open(config_file, 'r')
config_data = yaml.load(config_stream)
config_stream.close()
self.host = config_data['host']
self.requests = config_data['requests']
self.clients = config_data['clients']
self.time = config_data['time']
self.urls = config_data['urls']
except KeyError as ex:
raise ConfigKeyError(ex.args[0])
except IOError:
raise ConfigFileError(config_file)
def create_result_file(self, directory):
datetime = date.fromtimestamp(time())
file_name = '%d-%d-%d_%d-%d-%d' % (datetime.year,
datetime.month,
datetime.day,
datetime.hour,
datetime.minute,
datetime.second)
file_path = os.path.join(directory, file_name)
if not os.path.exists(directory):
os.makedirs(directory)
open(file_path, 'a').close()
self.result_file = file_path
def start_testing(self):
pass
def run(self):
file_stream = open(self.result_file, 'w')
print('Start tests ...')
for url in self.urls:
full_url = self.host + url
file_stream.write('URL: %s\n' % url)
for i in range(0, self.requests):
file_stream.write(' %i - %.3f\n' % (i, web.request(full_url)))
print('Finished tests!')
| import os
import yaml
from performance_testing.errors import ConfigFileError, ConfigKeyError
from performance_testing import web
from performance_testing.config import Config
from performance_testing.result import Result
class Tool:
def __init__(self, config='config.yml', result_directory='result'):
self.config = Config(config_path=config)
self.result = Result(result_directory)
def start_testing(self):
pass
def run(self):
print('Start tests ...')
for url in self.config.urls:
full_url = self.config.host + url
self.result.file.write_line('URL: %s\n' % url)
for i in range(0, self.config.requests):
self.result.file.write_line(' %i - %.3f\n' % (i, web.request(full_url)))
print('Finished tests!')
| Use Config and Result class in Tool | Use Config and Result class in Tool
| Python | mit | BakeCode/performance-testing,BakeCode/performance-testing | import os
import yaml
from performance_testing.errors import ConfigFileError, ConfigKeyError
from performance_testing import web
- from datetime import datetime as date
- from time import time
+ from performance_testing.config import Config
+ from performance_testing.result import Result
class Tool:
def __init__(self, config='config.yml', result_directory='result'):
- self.read_config(config_file=config)
+ self.config = Config(config_path=config)
+ self.result = Result(result_directory)
- self.create_result_file(directory=result_directory)
-
- def read_config(self, config_file):
- try:
- config_stream = open(config_file, 'r')
- config_data = yaml.load(config_stream)
- config_stream.close()
- self.host = config_data['host']
- self.requests = config_data['requests']
- self.clients = config_data['clients']
- self.time = config_data['time']
- self.urls = config_data['urls']
- except KeyError as ex:
- raise ConfigKeyError(ex.args[0])
- except IOError:
- raise ConfigFileError(config_file)
-
- def create_result_file(self, directory):
- datetime = date.fromtimestamp(time())
- file_name = '%d-%d-%d_%d-%d-%d' % (datetime.year,
- datetime.month,
- datetime.day,
- datetime.hour,
- datetime.minute,
- datetime.second)
- file_path = os.path.join(directory, file_name)
- if not os.path.exists(directory):
- os.makedirs(directory)
- open(file_path, 'a').close()
- self.result_file = file_path
def start_testing(self):
pass
def run(self):
- file_stream = open(self.result_file, 'w')
print('Start tests ...')
- for url in self.urls:
+ for url in self.config.urls:
- full_url = self.host + url
+ full_url = self.config.host + url
- file_stream.write('URL: %s\n' % url)
+ self.result.file.write_line('URL: %s\n' % url)
- for i in range(0, self.requests):
+ for i in range(0, self.config.requests):
- file_stream.write(' %i - %.3f\n' % (i, web.request(full_url)))
+ self.result.file.write_line(' %i - %.3f\n' % (i, web.request(full_url)))
print('Finished tests!')
| Use Config and Result class in Tool | ## Code Before:
import os
import yaml
from performance_testing.errors import ConfigFileError, ConfigKeyError
from performance_testing import web
from datetime import datetime as date
from time import time
class Tool:
def __init__(self, config='config.yml', result_directory='result'):
self.read_config(config_file=config)
self.create_result_file(directory=result_directory)
def read_config(self, config_file):
try:
config_stream = open(config_file, 'r')
config_data = yaml.load(config_stream)
config_stream.close()
self.host = config_data['host']
self.requests = config_data['requests']
self.clients = config_data['clients']
self.time = config_data['time']
self.urls = config_data['urls']
except KeyError as ex:
raise ConfigKeyError(ex.args[0])
except IOError:
raise ConfigFileError(config_file)
def create_result_file(self, directory):
datetime = date.fromtimestamp(time())
file_name = '%d-%d-%d_%d-%d-%d' % (datetime.year,
datetime.month,
datetime.day,
datetime.hour,
datetime.minute,
datetime.second)
file_path = os.path.join(directory, file_name)
if not os.path.exists(directory):
os.makedirs(directory)
open(file_path, 'a').close()
self.result_file = file_path
def start_testing(self):
pass
def run(self):
file_stream = open(self.result_file, 'w')
print('Start tests ...')
for url in self.urls:
full_url = self.host + url
file_stream.write('URL: %s\n' % url)
for i in range(0, self.requests):
file_stream.write(' %i - %.3f\n' % (i, web.request(full_url)))
print('Finished tests!')
## Instruction:
Use Config and Result class in Tool
## Code After:
import os
import yaml
from performance_testing.errors import ConfigFileError, ConfigKeyError
from performance_testing import web
from performance_testing.config import Config
from performance_testing.result import Result
class Tool:
def __init__(self, config='config.yml', result_directory='result'):
self.config = Config(config_path=config)
self.result = Result(result_directory)
def start_testing(self):
pass
def run(self):
print('Start tests ...')
for url in self.config.urls:
full_url = self.config.host + url
self.result.file.write_line('URL: %s\n' % url)
for i in range(0, self.config.requests):
self.result.file.write_line(' %i - %.3f\n' % (i, web.request(full_url)))
print('Finished tests!')
|
5fc699b89eae0c41923a813ac48281729c4d80b8 | orderable_inlines/inlines.py | orderable_inlines/inlines.py | from django.contrib.admin import StackedInline, TabularInline
from django.template.defaultfilters import slugify
class OrderableInlineMixin(object):
class Media:
js = (
'js/jquery.browser.min.js',
'js/orderable-inline-jquery-ui.js',
'js/orderable-inline.js',
)
css = {
'all': [
'css/orderable-inline.css'
]
}
def get_fieldsets(self, request, obj=None):
if self.declared_fieldsets:
return self.declared_fieldsets
form = self.get_formset(request, obj, fields=None).form
fields = list(form.base_fields) + list(self.get_readonly_fields(request, obj))
return [
(None, {
'fields': fields,
'classes': self.fieldset_css_classes + ['orderable-field-%s' % self.orderable_field]
})
]
class OrderableStackedInline(OrderableInlineMixin, StackedInline):
fieldset_css_classes = ['orderable-stacked']
class OrderableTabularInline(OrderableInlineMixin, TabularInline):
fieldset_css_classes = ['orderable-tabular']
template = 'orderable_inlines/edit_inline/tabular.html'
| from django.contrib.admin import StackedInline, TabularInline
from django.template.defaultfilters import slugify
class OrderableInlineMixin(object):
class Media:
js = (
'js/jquery.browser.min.js',
'js/orderable-inline-jquery-ui.js',
'js/orderable-inline.js',
)
css = {
'all': [
'css/orderable-inline.css'
]
}
def get_fieldsets(self, request, obj=None):
form = self.get_formset(request, obj, fields=None).form
fields = list(form.base_fields) + list(self.get_readonly_fields(request, obj))
return [
(None, {
'fields': fields,
'classes': self.fieldset_css_classes + ['orderable-field-%s' % self.orderable_field]
})
]
class OrderableStackedInline(OrderableInlineMixin, StackedInline):
fieldset_css_classes = ['orderable-stacked']
class OrderableTabularInline(OrderableInlineMixin, TabularInline):
fieldset_css_classes = ['orderable-tabular']
template = 'orderable_inlines/edit_inline/tabular.html'
| Make this hack compatible with Django 1.9 | Make this hack compatible with Django 1.9
| Python | bsd-2-clause | frx0119/django-orderable-inlines,frx0119/django-orderable-inlines | from django.contrib.admin import StackedInline, TabularInline
from django.template.defaultfilters import slugify
class OrderableInlineMixin(object):
class Media:
js = (
'js/jquery.browser.min.js',
'js/orderable-inline-jquery-ui.js',
'js/orderable-inline.js',
)
css = {
'all': [
'css/orderable-inline.css'
]
}
def get_fieldsets(self, request, obj=None):
- if self.declared_fieldsets:
- return self.declared_fieldsets
form = self.get_formset(request, obj, fields=None).form
fields = list(form.base_fields) + list(self.get_readonly_fields(request, obj))
return [
(None, {
'fields': fields,
'classes': self.fieldset_css_classes + ['orderable-field-%s' % self.orderable_field]
})
]
class OrderableStackedInline(OrderableInlineMixin, StackedInline):
fieldset_css_classes = ['orderable-stacked']
class OrderableTabularInline(OrderableInlineMixin, TabularInline):
fieldset_css_classes = ['orderable-tabular']
template = 'orderable_inlines/edit_inline/tabular.html'
| Make this hack compatible with Django 1.9 | ## Code Before:
from django.contrib.admin import StackedInline, TabularInline
from django.template.defaultfilters import slugify
class OrderableInlineMixin(object):
class Media:
js = (
'js/jquery.browser.min.js',
'js/orderable-inline-jquery-ui.js',
'js/orderable-inline.js',
)
css = {
'all': [
'css/orderable-inline.css'
]
}
def get_fieldsets(self, request, obj=None):
if self.declared_fieldsets:
return self.declared_fieldsets
form = self.get_formset(request, obj, fields=None).form
fields = list(form.base_fields) + list(self.get_readonly_fields(request, obj))
return [
(None, {
'fields': fields,
'classes': self.fieldset_css_classes + ['orderable-field-%s' % self.orderable_field]
})
]
class OrderableStackedInline(OrderableInlineMixin, StackedInline):
fieldset_css_classes = ['orderable-stacked']
class OrderableTabularInline(OrderableInlineMixin, TabularInline):
fieldset_css_classes = ['orderable-tabular']
template = 'orderable_inlines/edit_inline/tabular.html'
## Instruction:
Make this hack compatible with Django 1.9
## Code After:
from django.contrib.admin import StackedInline, TabularInline
from django.template.defaultfilters import slugify
class OrderableInlineMixin(object):
class Media:
js = (
'js/jquery.browser.min.js',
'js/orderable-inline-jquery-ui.js',
'js/orderable-inline.js',
)
css = {
'all': [
'css/orderable-inline.css'
]
}
def get_fieldsets(self, request, obj=None):
form = self.get_formset(request, obj, fields=None).form
fields = list(form.base_fields) + list(self.get_readonly_fields(request, obj))
return [
(None, {
'fields': fields,
'classes': self.fieldset_css_classes + ['orderable-field-%s' % self.orderable_field]
})
]
class OrderableStackedInline(OrderableInlineMixin, StackedInline):
fieldset_css_classes = ['orderable-stacked']
class OrderableTabularInline(OrderableInlineMixin, TabularInline):
fieldset_css_classes = ['orderable-tabular']
template = 'orderable_inlines/edit_inline/tabular.html'
|