hexsha
stringlengths
40
40
size
int64
5
2.06M
ext
stringclasses
10 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
248
max_stars_repo_name
stringlengths
5
125
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
sequencelengths
1
10
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
3
248
max_issues_repo_name
stringlengths
5
125
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
sequencelengths
1
10
max_issues_count
int64
1
67k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
3
248
max_forks_repo_name
stringlengths
5
125
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
sequencelengths
1
10
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
5
2.06M
avg_line_length
float64
1
1.02M
max_line_length
int64
3
1.03M
alphanum_fraction
float64
0
1
count_classes
int64
0
1.6M
score_classes
float64
0
1
count_generators
int64
0
651k
score_generators
float64
0
1
count_decorators
int64
0
990k
score_decorators
float64
0
1
count_async_functions
int64
0
235k
score_async_functions
float64
0
1
count_documentation
int64
0
1.04M
score_documentation
float64
0
1
4b44ab85799151e1020d7bb62b4190682ce5fa39
1,974
py
Python
src/entities/report/actions/consult.py
LuisArmando-TestCoder/ShareGraph
fa89d37c8fe522c526b903fe25bd1e22fd769425
[ "MIT" ]
null
null
null
src/entities/report/actions/consult.py
LuisArmando-TestCoder/ShareGraph
fa89d37c8fe522c526b903fe25bd1e22fd769425
[ "MIT" ]
null
null
null
src/entities/report/actions/consult.py
LuisArmando-TestCoder/ShareGraph
fa89d37c8fe522c526b903fe25bd1e22fd769425
[ "MIT" ]
null
null
null
from utilities.getStore import getStore filePath = "./entities/bill/store.json" productFilePath = "./entities/product/store.json" def getProduct(name): for product in getStore(productFilePath): if product["name"] == name: return product def getProductBillsAverage(name): productPriceSummation = 0 productAmount = 0 for bill in getStore(filePath): for sell in bill: if sell["name"] == name: productAmount += 1 productPriceSummation += sell["amount"] * getProduct( name )["price"] return productPriceSummation / (productAmount if productAmount else 1) def getSellsAverage(bills): # print("The average of sells has being") allProductsPricesSummation = 0 allSellsAmount = 0 for bill in bills: allSellsAmount += 1 for sell in bill: allProductsPricesSummation = sell["amount"] * getProduct(sell["name"])["price"] print(f"The average of sells has being {allProductsPricesSummation / allSellsAmount}") def getProductsAverage(): print("The average of sells for each product") for product in getStore(productFilePath): print(f"For {product['name']} the average sells are {getProductBillsAverage(product['name'])}") def getHighestSellWithBill(bills): maximumBillIndex = 0 maximumBill = 0 for billIndex in range(len(bills)): billSummation = 0 for sell in bills[billIndex]: billSummation += sell["amount"] * getProduct( sell["name"] )["price"] if billSummation > maximumBill: maximumBill = billSummation maximumBillIndex = billIndex print(f"The highest sell is {maximumBill}") print(f"For the following bill {bills[maximumBillIndex]}") def main(): bills = getStore(filePath) getSellsAverage(bills) getProductsAverage() getHighestSellWithBill(bills)
28.608696
103
0.644377
0
0
0
0
0
0
0
0
462
0.234043
4b44b5778d0cc6b0adc1458cef3d5591585dd53d
1,826
py
Python
discordbot.py
kinotan/discordpy-startup
1505c4f78deff7f793de75985e669ee84a78a3f2
[ "MIT" ]
null
null
null
discordbot.py
kinotan/discordpy-startup
1505c4f78deff7f793de75985e669ee84a78a3f2
[ "MIT" ]
null
null
null
discordbot.py
kinotan/discordpy-startup
1505c4f78deff7f793de75985e669ee84a78a3f2
[ "MIT" ]
null
null
null
#discord.pyのインポート from asyncio import sleep import discord client = discord.Client() #BOTログイン処理 @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') await client.change_presence(game=discord.Game(name='!delchat *')) # BOT動作プログラム @client.event async def on_message(message): # 送り主がBotだった場合反応したくないので if client.user != message.author: # 削除コマンド if message.content.startswith("!delchat "): #役職比較 if discord.utils.get(message.author.roles, name="admin"): # メッセージを格納 delcmd = message.content # 入力メッセージのリスト化 delcmd_ = delcmd.split() # 入力メッセージのint化 delcmd_int = int(delcmd_[1]) # 入力メッセージの単語数 delcmd_c = len(delcmd_) if delcmd_c == 2 and delcmd_int <= 50 and delcmd_int > 1: # メッセージ取得 msgs = [msg async for msg in client.logs_from(message.channel, limit=(delcmd_int+1))] await client.delete_messages(msgs) delmsg = await client.send_message(message.channel, '削除が完了しました') await sleep(5) await client.delete_message(delmsg) else: # エラーメッセージを送ります delmsg = await client.send_message(message.channel, "コマンドが間違っています。[!delchat *] *:2~50") await sleep(5) await client.delete_message(delmsg) else: # エラーメッセージを送ります delmsg = await client.send_message(message.channel, "admin権限がありません。") await sleep(5) await client.delete_message(delmsg) client.run("***")
37.265306
107
0.545455
0
0
0
0
1,963
0.921596
1,935
0.908451
587
0.275587
4b453f03ad86e35ca9832f88f32cb3f426b3e7ef
1,191
py
Python
api/migrations/0002_element_pokemon.py
danielchikara/pokemon_in_hom
5da9baa3f87e012ae0d4278668409e1668bf87a6
[ "MIT" ]
null
null
null
api/migrations/0002_element_pokemon.py
danielchikara/pokemon_in_hom
5da9baa3f87e012ae0d4278668409e1668bf87a6
[ "MIT" ]
null
null
null
api/migrations/0002_element_pokemon.py
danielchikara/pokemon_in_hom
5da9baa3f87e012ae0d4278668409e1668bf87a6
[ "MIT" ]
null
null
null
# Generated by Django 3.2.11 on 2022-01-08 03:50 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('api', '0001_initial'), ] operations = [ migrations.CreateModel( name='Element', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100)), ('description', models.TextField()), ], ), migrations.CreateModel( name='Pokemon', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100)), ('image', models.CharField(max_length=700)), ('id_pokedex', models.IntegerField()), ('description', models.TextField()), ('id_element', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='element_pokemons', to='api.element')), ], ), ]
35.029412
146
0.572628
1,064
0.893367
0
0
0
0
0
0
201
0.168766
4b45b1a0f99a331cd3db5bcbd3c80d4d359c59e4
5,468
py
Python
src/CLI/actioner/sonic-cli-ip-prefix.py
project-arlo/sonic-mgmt-framework
562cd84ff3fec9ca705c7df621742f2daa61ce71
[ "Apache-2.0" ]
7
2019-10-17T06:12:02.000Z
2021-09-08T11:16:19.000Z
src/CLI/actioner/sonic-cli-ip-prefix.py
noolex/sonic-mgmt-framework
5493889adc47fc584b04dca1a0cc0a2007211df4
[ "Apache-2.0" ]
207
2019-06-24T04:48:11.000Z
2020-05-06T05:51:37.000Z
src/CLI/actioner/sonic-cli-ip-prefix.py
noolex/sonic-mgmt-framework
5493889adc47fc584b04dca1a0cc0a2007211df4
[ "Apache-2.0" ]
20
2019-06-27T19:24:45.000Z
2021-07-15T21:12:30.000Z
#!/usr/bin/python ########################################################################### # # Copyright 2019 Dell, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ########################################################################### import sys import time import json import ast import cli_client as cc from rpipe_utils import pipestr from scripts.render_cli import show_cli_output import urllib3 urllib3.disable_warnings() def generate_ipprefix_uri(args, delete): _action = "PERMIT" _mode = set_name = ge = le = _maskrange_length = _ip_prefix = '' ge_val = le_val = prefix_exits = le_exits = ge_exits = is_error = i = 0 for arg in args: if "permit" == arg: _action = "PERMIT" elif "deny" == arg: _action = "DENY" elif "prefix-list" == arg: set_name = args[i+1] if len(args) > 4: _ip_prefix = args[i+3] prefix_exits = 1 elif "ge" == arg: ge_exits = 1 ge_val = int(args[i+1]) ge = args[i+1] elif "le" == arg: le_exits = 1 le_val = int(args[i+1]) le = args[i+1] elif "ip" == arg: _mode = "IPV4" max = "32" elif "ipv6" == arg: _mode = "IPV6" max = "128" else: temp = 1 i = i + 1 if prefix_exits: _prefix, _mask = _ip_prefix.split("/") mask_val = int(_mask) if (ge_exits == 0 and le_exits == 0): _maskrange_length = "exact" elif (ge_exits == 1 and le_exits == 0): if (ge_val <= mask_val): is_error = 1 _maskrange_length = ge + ".." + max elif (ge_exits == 0 and le_exits == 1): if (mask_val > le_val): is_error = 1 _maskrange_length = _mask+".."+le else: if ((ge_val <= mask_val) or (mask_val > le_val) or (ge_val > le_val)): is_error = 1 _maskrange_length = ge+".."+le if is_error: print ("%Error: Invalid prefix range, make sure: len < ge <= le") exit(1) if delete: keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets/prefix-set={prefix_list_name}/prefixes/prefix={prefix}%2F{mask},{masklength_range}', prefix_list_name=set_name, prefix=_prefix, mask=_mask, masklength_range=_maskrange_length) body = None else: keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets') body = {"openconfig-routing-policy:prefix-sets":{"prefix-set":[{"name": set_name,"config":{"name": set_name, "mode": _mode},"prefixes":{"prefix":[{"ip-prefix": _ip_prefix,"masklength-range": _maskrange_length,"config": { "ip-prefix": _ip_prefix,"masklength-range": _maskrange_length,"openconfig-routing-policy-ext:action": _action}}]}}]}} else: keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets/prefix-set={prefix_list_name}', prefix_list_name=set_name) body = None return keypath, body def invoke(func, args): body = None aa = cc.ApiClient() if func == 'ip_prefix_create': keypath, body = generate_ipprefix_uri(args, 0) return aa.patch(keypath, body) elif func == 'ip_prefix_delete': keypath, body = generate_ipprefix_uri(args, 1) return aa.delete(keypath) elif func == 'ip_prefix_show_all': keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets') return aa.get(keypath) elif func == 'ip_prefix_show_specific': keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets/prefix-set={name}',name=args[1]) return aa.get(keypath) elif func == 'ipv6_prefix_show_all': keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets') return aa.get(keypath) elif func == 'ipv6_prefix_show_specific': keypath = cc.Path('/restconf/data/openconfig-routing-policy:routing-policy/defined-sets/prefix-sets/prefix-set={name}',name=args[1]) return aa.get(keypath) else: return aa.cli_not_implemented(func) def run(func, args): try: response = invoke(func,args) if response.ok(): if response.content is not None: # Get Command Output api_response = response.content if api_response is None: print("Failed") return #print api_response show_cli_output(args[0], api_response) else: print response.error_message() return except Exception as e: print "%Error: " + str(e) return if __name__ == '__main__': pipestr().write(sys.argv) run(sys.argv[1], sys.argv[2:])
34.389937
285
0.599854
0
0
0
0
0
0
0
0
2,018
0.369056
4b465725a6717037599028f9aa649996198118b6
268
py
Python
exec-shell.py
zqsheng/snippet
cb14300fc62c616d48e6552ad93c6d33b5e8c9a1
[ "Apache-2.0" ]
1
2018-09-10T11:31:33.000Z
2018-09-10T11:31:33.000Z
exec-shell.py
zqsheng/snippet
cb14300fc62c616d48e6552ad93c6d33b5e8c9a1
[ "Apache-2.0" ]
null
null
null
exec-shell.py
zqsheng/snippet
cb14300fc62c616d48e6552ad93c6d33b5e8c9a1
[ "Apache-2.0" ]
null
null
null
import os import time exec_count = 100 cmds = [] cmds.append("tar -cPzf /opt/web.tar.gz /opt/web/ /opt/soft") cmds.append("rm -f /opt/web.tar.gz") for i in range(exec_count): for cmd in cmds: if os.system(cmd) != 0: break time.sleep(1)
24.363636
60
0.604478
0
0
0
0
0
0
0
0
70
0.261194
4b489a2a948e0d5e9116987af060d05a349b77d2
123
py
Python
handlers/__init__.py
IronWolf-K/nonebot_plugin_fr24
f3752882598de54f41cd9b27456dd3b6f88971e2
[ "MIT" ]
null
null
null
handlers/__init__.py
IronWolf-K/nonebot_plugin_fr24
f3752882598de54f41cd9b27456dd3b6f88971e2
[ "MIT" ]
null
null
null
handlers/__init__.py
IronWolf-K/nonebot_plugin_fr24
f3752882598de54f41cd9b27456dd3b6f88971e2
[ "MIT" ]
null
null
null
from .help import help_handler from .pre import pre_handler from .now import now_handler from .filter import filter_handler
30.75
34
0.845528
0
0
0
0
0
0
0
0
0
0
4b48a1ce648ccd7eddf1077ee9304a100d815be4
581
py
Python
day6_10/day6b.py
invincible-akshay/advent_of_code2020
81f207c6f7218ff235c31d67e1b4659cc482297c
[ "MIT" ]
null
null
null
day6_10/day6b.py
invincible-akshay/advent_of_code2020
81f207c6f7218ff235c31d67e1b4659cc482297c
[ "MIT" ]
null
null
null
day6_10/day6b.py
invincible-akshay/advent_of_code2020
81f207c6f7218ff235c31d67e1b4659cc482297c
[ "MIT" ]
null
null
null
import utils.fileutils as futils inp = futils.read_list("../data/day6.txt") nums_dict = dict() group_size, res_count = 0, 0 for line in inp: if line == "": # res_count += len(nums_set) for k, v in nums_dict.items(): if v == group_size: res_count += 1 nums_dict = dict() group_size = 0 continue group_size += 1 for ch in line: nums_dict[ch] = 1 + nums_dict.get(ch, 0) for k, v in nums_dict.items(): if v == group_size: res_count += 1 print("Sum of counts: {0}".format(res_count))
25.26087
48
0.567986
0
0
0
0
0
0
0
0
68
0.11704
4b48d527c36dcd783a13d6a5609545147bc8c89c
45,165
py
Python
platform/hwconf_data/zgm13/PythonSnippet/ExporterModel.py
lenloe1/v2.7
9ac9c4a7bb37987af382c80647f42d84db5f2e1d
[ "Zlib" ]
null
null
null
platform/hwconf_data/zgm13/PythonSnippet/ExporterModel.py
lenloe1/v2.7
9ac9c4a7bb37987af382c80647f42d84db5f2e1d
[ "Zlib" ]
1
2020-08-25T02:36:22.000Z
2020-08-25T02:36:22.000Z
platform/hwconf_data/zgm13/PythonSnippet/ExporterModel.py
lenloe1/v2.7
9ac9c4a7bb37987af382c80647f42d84db5f2e1d
[ "Zlib" ]
1
2020-08-25T01:56:04.000Z
2020-08-25T01:56:04.000Z
from . import types from . import dep from . import RuntimeModel from . import Metadata class Property(object): def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None): ''' :param name: Name of the property (string) :param type: PropertyType describing the type of property :param visible: Boolean, whether the property is visible in the UI or not ''' self.name = name self.type = type self.dependencies = [] self.generation = [] self.visible = visible self.readonly = readonly self.category='' self.subcategory='' self.namespace = namespace self.id = '.'.join((str(namespace).upper(), self.name)) self.label = description self.description = long_description self.defaultvalue = '' self.transient = False self.parent = None self.mode = None self.define_name = define_name if define_name else name self.is_advanced = False self.allowedconflicts = [] self.generate_if_hidden = True def set_parent_module(self, mod): self.parent = mod # Set default category to name of module if not self.category: self.category = self.parent.name def set_namespace(self, namespace): ''' Sets the namespace on a property :param namespace: :return: ''' self.namespace = namespace self.id = '.'.join((str(namespace).upper(), self.name, 'PROP')) def set_visibility(self, visible): self.visible = visible def set_readonly(self, readonly): self.readonly = readonly def add_dependency(self, dependency): self.dependencies.append(dependency) def get_dependencies(self): return self.dependencies def generateXML(self): ''' Generate the Studio XML for this property :return: etree.Element containing the Studio XML describing the property ''' print("Not able to gen XML from base property!") print(self.name) print(self.type) return None class StringProperty(Property): ''' Property which can take on a string value ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None): Property.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.id = '.'.join((str(namespace).upper(), self.name, 'STRING')) def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'STRING')) class ArrayProperty(Property): ''' Property which can take on an array value ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None): Property.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.id = '.'.join((str(namespace).upper(), self.name, 'ARRAY')) def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'ARRAY')) class IntegerProperty(Property): ''' Property which can take on integer values ''' def __init__(self, name, description, min, max, default, namespace='', visible=False, readonly=False, define_name=None, long_description=None): Property.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.min = int(min) self.max = int(max) self.defaultvalue = int(default) self.id = '.'.join((str(namespace).upper(), self.name, 'INT')) self.format = None def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'INT')) def set_format(self, format): self.format = format class Enum: ''' Container class for an item inside of an EnumProperty ''' def __init__(self, value, index, define_value=None, visible=True): self.value = value self.visible = visible self.index = index if define_value is not None: self.define_value = define_value else: self.define_value = value class EnumProperty(Property): ''' Property allowing a selection from a list of options ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None): Property.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.values = {} self.id = '.'.join((str(namespace).upper(), self.name, 'ENUM')) def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'ENUM')) def add_enum(self, value, define_value=None, visible=True): ''' Add an option to the selection list :param value: String value for the option (visible in UI) :param visible: Whether or not this option will be visible in the UI :param define_value: Name which will be generated as #def value :return: None ''' self.values[len(self.values.keys())] = Enum(value, len(self.values.keys()), define_value=define_value, visible=visible) class ModeProperty(EnumProperty): def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None, hide_properties=True): Property.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.values = {} self.id = '.'.join((str(namespace).upper(), self.name, 'ENUM')) self.hide_properties = hide_properties class BoolProperty(EnumProperty): ''' Property allowing you to select a binary setting ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None): EnumProperty.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.id = '.'.join((str(namespace).upper(), self.name, 'BOOL')) self.add_enum('False', define_value="0") self.add_enum('True', define_value="1") def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'BOOL')) def set_default_to_false(self): self.defaultvalue = 'False' def set_default_to_true(self): self.defaultvalue = 'True' class CheckboxProperty(Property): ''' Property allowing you to select a binary setting using a checkbox ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None): Property.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name) self.id = '.'.join((str(namespace).upper(), self.name, 'CHECKBOX')) def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'CHECKBOX')) class ModuleProperty(EnumProperty): ''' Property allowing you to select a peripheral available on the current chip ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, long_description=None): EnumProperty.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.id = '.'.join((str(namespace).upper(), self.name, 'MOD')) self.allowedModules = [] self.inherit_options = False self.define_value_prefix = '' self.owned_mode = None self.define_name_postfix = '' def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'MOD')) def add_allowed_module(self, module_namespace): ''' Adds a module 'namespace' to the allowed modules for this property :param module_namespace: regular expression for which modules can be selected by this property :return: None ''' self.allowedModules.append(module_namespace) def mask_with_module_list(self, module_name_list): ''' Updates the list of allowed modules for this property by comparing a list with the property's allowed modules. :param module_name_list: list of module names available on this part :return: ''' self.values = {} self.add_enum('None') for mod_name in module_name_list: for allowed_mod in self.allowedModules: if mod_name.rstrip('0123456789') == allowed_mod: define_value = mod_name self.add_enum(mod_name, define_value=define_value) class PinProperty(EnumProperty): ''' Property allowing you to select any GPIO pin available ''' def __init__(self, name, description, namespace='', visible=False, readonly=False, define_name=None, disabled_label=None, long_description=None): EnumProperty.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.referenced_module = None self.referenced_route = None self.em4 = False self.id = '.'.join((str(namespace).upper(), self.name, 'PIN')) if not disabled_label: disabled_label = 'Disabled' self.disabled_label = disabled_label self.add_enum(disabled_label, define_value="Disabled") def set_namespace(self, namespace): self.id = '.'.join((str(namespace).upper(), self.name, 'PIN')) def mask_with_pin_list(self, pin_list): ''' Updates the available enum values with the values from pin_list :param pin_list: list of pin names available on the part :return: None ''' self.values={} self.add_enum(self.disabled_label, define_value="Disabled") for pin in pin_list: self.add_enum(pin) def set_reference_route(self, route): self.referenced_route = route def set_reference_module(self, module): self.referenced_module = module def set_reference(self, module, route): self.referenced_module = module self.referenced_route = route class PRSChannelProperty(EnumProperty): """ Property allowing you to select PRS channel available from the PRS module """ def __init__(self, name, description, channel_count, custom_name="", namespace='', visible=False, readonly=False, define_name=None, long_description=None, gpio=True): EnumProperty.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name, long_description=long_description) self.add_enum("Disabled") self.channel_count = channel_count self.custom_name = custom_name self.gpio = gpio for i in range(channel_count): self.add_enum("CH" + str(i), define_value=str(i)) class AportBusProperty(EnumProperty): """ APORT bus select """ def __init__(self, name, description, signal=None, define_name_prefix=None, define_value_prefix=None, namespace='', visible=True, readonly=False, define_name=None): EnumProperty.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name) self.signal = signal self.define_name_prefix = define_name_prefix self.define_value_prefix = define_value_prefix self.extra_enums = [] self.bus_props = {} def add_extra_enum(self, value, define_value=None): self.extra_enums.append((value, define_value)) def mask_with_bus_list(self, bus_list, superset=False): ''' Updates the list of allowed buses for this property. :param bus_list: list of buses available on this part :return: ''' self.values = {} # Find existing referenced bus property bus_property_prefix = "{}_".format(self.define_name_prefix.lower()) for bus_name in bus_list: aportname = busname_to_aportname(self.parent.name, bus_name) # Add bus to bus dropdown self.add_enum("APORT bus {}".format(bus_name), define_value="APORT{}".format(aportname)) # Add channel dropdown for bus bus_property_name = bus_property_prefix + bus_name bus_prop = self.bus_props.get(bus_property_name) if not bus_prop: bus_prop = PinProperty(bus_property_name, "Channel on APORT bus {} ({})".format(bus_name, self.signal), define_name=self.define_name_prefix, visible=False) bus_prop.set_reference(self.parent.name, "{}_{}".format(self.signal, bus_name)) bus_prop.category = self.category bus_prop.subcategory = self.subcategory self.parent.add_property(bus_prop) self.bus_props[bus_property_name] = bus_prop else: bus_prop.set_visibility(False) # Add extra values to bus dropdown for value, define_value in self.extra_enums: self.add_enum(value, define_value=define_value) class AportScanProperty(EnumProperty): scan_props = {} """ APORT scan select """ def __init__(self, name, description, define_name_prefix=None, define_value_prefix=None, namespace='', visible=True, readonly=False, define_name=None): EnumProperty.__init__(self, name, description, namespace=namespace, visible=visible, readonly=readonly, define_name=define_name) self.define_name_prefix = define_name_prefix self.define_value_prefix = define_value_prefix self.scan_mask = None self.start = 0 self.end = 0 def attach_to_scan_mask(self, scan_mask_property): self.scan_mask = scan_mask_property def set_range(self, start, end): self.start = start self.end = end def mask_with_bus_list(self, bus_list, superset=False): ''' Updates the list of allowed buses for this property. :param bus_list: list of buses available on this part :return: ''' self.values = {} if not self.scan_props.get(self.parent.name): self.scan_props[self.parent.name] = {} bus_channels = {} aports = {} updated_scan_props = [] # print(bus_list) for signal, buses in bus_list.items(): for bus_name, routes in buses.items(): aport = busname_to_aportname(self.parent.name, bus_name)[:-1] bus_name = bus_name[:-1] aports[bus_name] = aport if bus_name not in bus_channels: bus_channels[bus_name] = set() bus_channels[bus_name] = bus_channels[bus_name] | set(routes) for name, route_prop in self.scan_props[self.parent.name].items(): # Hide props by default route_prop.set_visibility(False) for bus, routes in bus_channels.items(): channels_available = [False, False, False, False] for route in sorted(routes, key=lambda r: r.number): channels_available[int(route.number / 8)] = True for i in range(4): start = i * 8 end = (i + 1) * 8 - 1 if channels_available[i]: self.add_enum("APORT bus {} channel {}-{}".format(bus, start, end), "APORT{}CH{}TO{}".format(aports[bus], start, end)) else: self.add_enum("APORT bus {} channel {}-{} (no pins available)".format(bus, start, end), "APORT{}CH{}TO{}".format(aports[bus], start, end)) if superset: for route in sorted(routes, key=lambda r: r.number): route_prop_name = "{}_{}_ch{}".format(self.name.lower().rsplit('_', 1)[0], bus, route.number) route_prop = self.scan_props[self.parent.name].get(route_prop_name) if not route_prop: route_prop = CheckboxProperty(route_prop_name, "Enable channel {}".format(route.number), namespace=self.parent.namespace, visible=False) channel_range_start = int(route.number / 8) * 8 channel_range_end = channel_range_start + 7 route_prop.category = self.category route_prop.subcategory = "APORT bus {} channel {}-{}".format(bus, channel_range_start, channel_range_end) self.scan_props[self.parent.name][route_prop_name] = route_prop self.parent.add_property(route_prop) else: for route in sorted(routes, key=lambda r: r.number): route_prop_name = "{}_{}_ch{}".format(self.name.lower().rsplit('_', 1)[0], bus, route.number) route_prop = self.scan_props[self.parent.name].get(route_prop_name) route_prop.label = "Enable channel {} ({})".format(route.number, route.padName) route_prop.set_visibility(True) updated_scan_props.append(route_prop) if not superset: return updated_scan_props class AportScanMaskProperty(IntegerProperty): """ APORT scan mask """ def __init__(self, name, description, min, max, default, namespace='', visible=False, readonly=False, define_name=None): IntegerProperty.__init__(self, name, description, min, max, default, namespace, visible, readonly, define_name) self.channel_selectors = [] self.other_scan_masks = [] self.set_format("0x{:08X}UL") self.channel_start = 0 def add_channel_selector(self, channel_selector): self.channel_selectors.append(channel_selector) class AportBondedMaskProperty(IntegerProperty): """ APORT bonded mask """ def __init__(self, name, description, min, max, default, namespace='', visible=False, readonly=False, define_name=None): IntegerProperty.__init__(self, name, description, min, max, default, namespace, visible, readonly, define_name) self.set_format("0x{:08X}UL") self.channel_start = 0 self.aport = "1" self.input_props = [] def mask_with_bus_list(self, bus_list, superset=False): ''' Updates the list of allowed buses for this property. :param bus_list: list of buses available on this part :return: ''' updated_inputs = [] bus_channels = {} if not superset: for signal, buses in bus_list.items(): for bus_name, routes in buses.items(): bus_name = bus_name[:-1] if bus_name not in bus_channels: bus_channels[bus_name] = set() bus_channels[bus_name] = bus_channels[bus_name] | set(routes) for route in sorted(bus_channels[aportname_to_busname(self.parent.name, self.aport)], key=lambda r: r.number): route_prop = self.input_props[int(route.number) % 32] route_prop.label = "Enable channel {} ({})".format(route.number, route.padName) route_prop.set_visibility(True) updated_inputs.append(route_prop) return updated_inputs class StudioModule(object): """docstring for StudioModule""" def __init__(self, basename, modules): super(StudioModule, self).__init__() self.basename = basename self.modules = {} for m in modules: # Allow both lists of frameworks and single framework to be specified if isinstance(m.frameworks,list): for framework in m.frameworks: self.modules[framework.value] = m.name else: self.modules[m.frameworks.value] = m.name def getModuleId(self, framework): print ("%s: %s" % (self.basename, self.modules.keys())) if framework not in self.modules: return None return "%s.%s" % (self.basename, self.modules[framework]) def __str__(self): return self.basename class Module(object): ''' Class for describing a HALConfig module or device peripheral. A module is basically a collection of properties. ''' def __init__(self, name, core=False, visible=False, namespace=None): # Name is a required argument self.name = name self.displayname = name # namespace defaults to module base (i.e. module without the instance number) if namespace: self.namespace = namespace else: self.namespace = name.rstrip('0123456789') # No description by default self.description = "" # Core signifies a module contributed by the die self.core = core # Visible controls whether the module shows up in the UI self.visible = visible # List of properties on this module self.properties = [] # Category is the category where to put the module on the UI. Default for core is 'Core'. self.category = ' Peripherals' if core else " HAL" # Define generated with the module being active (selected) or not self.enabled_define = None # Compatibility of module self.compatibility = dep.Dependency() # Association with an on-chip peripheral self.peripheral = None # Studio module specifier self.studio_module = None # By default, module has no custom name property self.has_custom_name = False self.model = None self.family = None # Contribute 'standard' properties for every module, allowing SDKs to take control in a hwconf doc # (even though they shouldn't) inuse = BoolProperty('usedbysdk', 'SDK is taking control over this module', visible=False) self.add_property(inuse) hidden = BoolProperty('hiddenbysdk', 'SDK is hiding this module', visible=False) self.add_property(hidden) hidden = BoolProperty('showadvanced', 'Show advanced options', visible=False) self.add_property(hidden) forceenable = BoolProperty('forceenable', 'Forcefully enabled in model', visible=False) self.add_property(forceenable) owner = StringProperty('owner', 'Owned by', visible=True, readonly=True) owner.transient = True self.add_property(owner) if self.core and (self.namespace != self.name): # Add custom name property if this is a parameterized core module (e.g. USARTn, TIMERn...) customname = StringProperty('customname', 'Custom name', visible=True, readonly=False) self.add_property(customname) self.has_custom_name = True def __str__(self): if self.studio_module: return str(self.studio_module) return "none" def add_property(self, prop): ''' Add a property to this module :type prop: Property :param prop: property to add :return: None ''' # Regular list append for now # TODO: change to property merge on properties with same ID prop.set_namespace(self.namespace) prop.set_parent_module(self) self.properties.append(prop) def load_halconfig_model(self, available_module_names_list, family=None): ''' Load a HAL config model :param model: a HAL config model :param family: a halconfig_dependency Family object describing for which family this module is loaded or str containing family name :return: None ''' if not family: raise ValueError("Family is not set") if isinstance(family, str): self.family = dep.Family(family_str=family) else: self.family = family self.family.available_mods = available_module_names_list if hasattr(self.model, 'compatibility'): self.compatibility = self.model.compatibility if hasattr(self.model, "peripheral"): self.peripheral = self.model.peripheral if hasattr(self.model, "category"): self.category = self.model.category if hasattr(self.model, "displayname"): self.displayname = self.model.displayname if hasattr(self.model, "description"): self.description = self.model.description if hasattr(self.model, "studio_module"): self.studio_module = StudioModule(self.model.studio_module["basename"], \ self.model.studio_module["modules"]) if hasattr(self.model, 'modes'): mode_prop = ModeProperty(self.model.modes["define"], "mode", visible=True, hide_properties=self.model.modes.get('hide_properties', True)) for val in self.model.modes["values"]: if isinstance(val, types.EnumValue): if val.dependency: if val.dependency.applies_to(family=self.family): mode_prop.add_enum(val.display_name, define_value=val.define_value) else: mode_prop.add_enum(val.display_name, define_value=val.define_value) else: mode_prop.add_enum(val) self.add_property(mode_prop) if hasattr(self.model, 'enable'): self.enabled_define = self.model.enable["define"] for prop, options in self.model.options.items(): current_opt_set = None # If one property has several option elements, iterate to find which option element has the correct dependency if isinstance(options, list): for opt in options: # Skip documentation option if opt.get("documentation"): continue if opt.get("dependency"): if opt.get("dependency").applies_to_family(self.family): if opt.get("dependency").applies_to_module(self.name): current_opt_set = opt break else: if options.get("dependency"): if options.get("dependency").applies_to_family(self.family): if options.get("dependency").applies_to_module(self.name): current_opt_set = options else: current_opt_set = options if current_opt_set is not None: self._load_halconfig_property(prop, current_opt_set, self.family, self.model) self.post_load() def _load_halconfig_property(self, prop, opts, family, model): """ :param prop: a HAL config property :param opts: dictionary containing a set of options for current prop :param family: a halconfig_dependency Family object describing for which family this module is loaded :return: None """ prop_obj = None extra_properties = [] if opts['type'] == 'enable': self.enabled_define = prop elif opts['type'] == 'boolean': prop_obj = BoolProperty(prop, opts['description'], visible=True) elif opts['type'] == 'integer': prop_obj = IntegerProperty(prop, opts['description'], opts['min'], opts['max'], 0, visible=True) elif isinstance(opts['type'], str) and 'int' in opts['type'] and '_t' in opts['type']: prop_obj = IntegerProperty(prop, opts['description'], opts['min'], opts['max'], 0, visible=True) elif opts['type'] == 'string': prop_obj = StringProperty(prop, opts['description'], visible=True) elif opts['type'] == 'array': prop_obj = ArrayProperty(prop, opts['description'], visible=True) elif opts['type'] == 'enum': if opts['values']: prop_obj = EnumProperty(prop, opts['description'], visible=True) for val in opts['values']: if isinstance(val, types.EnumValue): if val.dependency: if val.dependency.applies_to_family(family=family): prop_obj.add_enum(val.display_name, define_value=val.define_value) else: prop_obj.add_enum(val.display_name, define_value=val.define_value) else: prop_obj.add_enum(val) elif isinstance(opts['type'], types.Pin): prop_obj = PinProperty(prop, opts['description'], visible=True, disabled_label=opts['type'].disabled_label) if opts['type'].signal: # Pin is connected to a PORTIO signal prop_obj.set_reference(self.name, opts['type'].signal) if opts['type'].em4: prop_obj.em4 = True elif isinstance(opts['type'], types.Peripheral): prop_obj = ModuleProperty(prop, opts['description'], visible=True) for filter in opts['type'].filter: prop_obj.add_allowed_module(filter) prop_obj.inherit_options = opts['type'].inherit_options prop_obj.define_value_prefix = opts['type'].define_value_prefix prop_obj.define_name_postfix = opts['type'].define_name_postfix if hasattr(opts['type'], 'mode'): prop_obj.owned_mode = opts['type'].mode elif isinstance(opts['type'], types.PinArray): prop_obj = IntegerProperty(opts['type'].count_define, opts['description'], opts['type'].min, opts['type'].max, opts['type'].default, visible=True) init_string = "" for i in range(opts['type'].min, opts['type'].max): visible = True if i < opts['type'].default else False item_property = PinProperty(opts['type'].item_define.replace("%n", str(i)), opts['type'].item_description.replace("%n", str(i)), visible=visible) if opts.get('allowedconflicts') is not None: item_property.allowedconflicts = opts['allowedconflicts'] if visible: init_string += ("{{ {0}, {1} }}, ".format(opts['type'].item_port_define.replace("%n", str(i)), opts['type'].item_pin_define.replace("%n", str(i)))) extra_properties.append(item_property) if init_string: # Strip last comma space from default value init_string = init_string[:-2] init_property = ArrayProperty(opts['type'].init_define, "{} init".format(prop), visible=False) init_property.defaultvalue = init_string init_property.transient = True extra_properties.append(init_property) elif isinstance(opts['type'], types.PRSChannelLocation): prs_chan_count = Metadata.get_prs_chan_with_gpio_count(family.get_name()) prop_obj = PRSChannelProperty(opts['type'].define, opts['description'], prs_chan_count, custom_name=opts['type'].custom_name, gpio=opts['type'].gpio, visible=True) if dep.Dependency(platform=dep.Platform.SERIES0).applies_to_family(family): # Make PRS dropdown readonly on Series 0, since changing it will affect unrelated modules that # also use PRS. Users will have to use PORTIO view to select PRS location. readonly = True else: readonly = False if opts['type'].gpio: disabled_property = StringProperty( "prs_disabled_chn_{}_pin".format(opts['type'].custom_name if opts['type'].custom_name else ""), "PRS channel output pin", visible=True, readonly=True, long_description="No PRS channel selected") if opts.get('category') is not None: disabled_property.category = opts['category'] if opts.get('subcategory') is not None: disabled_property.subcategory = opts['subcategory'] extra_properties.append(disabled_property) for i in range(prs_chan_count): item_property = PinProperty(opts['type'].name + str(i), opts['type'].output_description.replace("%n", str(i)), visible=False, readonly=readonly, define_name=opts['type'].name) if dep.Dependency(platform=dep.Platform.SERIES2).applies_to_family(family): item_property.set_reference("PRS", "ASYNCH" + str(i)) else: item_property.set_reference("PRS", "CH" + str(i)) if opts.get('category') is not None: item_property.category = opts['category'] if opts.get('subcategory') is not None: item_property.subcategory = opts['subcategory'] extra_properties.append(item_property) elif isinstance(opts['type'], types.AportSingleChannel): obj = opts['type'] prop_obj = AportBusProperty(obj.define, opts['description'], signal=obj.signal, define_name_prefix=obj.define_name_prefix, define_value_prefix=obj.define_value_prefix) for val in obj.extra_values: if isinstance(val, types.EnumValue): if val.dependency: if val.dependency.applies_to_family(family=family): prop_obj.add_extra_enum(val.display_name, define_value=val.define_value) else: prop_obj.add_extra_enum(val.display_name, define_value=val.define_value) else: prop_obj.add_extra_enum(val) elif isinstance(opts['type'], types.AportScanMode): obj = opts['type'] prop_obj = AportScanMaskProperty(prop, opts['description'], 0, 0xFFFFFFFF, 0, visible=True, readonly=True) prop_obj.channel_start = obj.channel_start define_prefix = prop.rsplit('_', 1)[0] range_start = int(obj.channel_start / 8) for i in range(range_start, range_start + 4): start = i * 8 end = (i + 1) * 8 - 1 input_number = "{}TO{}".format(start, end) input_name = "{}_INPUT{}".format(define_prefix, input_number) input_prop = AportScanProperty(input_name, "Input {} to {}".format(start, end), define_value_prefix=obj.define_value_prefix.replace("%n", input_number)) if opts.get('mode') is not None: input_prop.mode = opts['mode'] input_prop.set_range(start, end) if opts.get('subcategory') is not None: input_prop.subcategory = opts['subcategory'] if opts.get('category') is not None: input_prop.category = opts['category'] input_prop.attach_to_scan_mask(prop_obj) prop_obj.add_channel_selector(input_prop) extra_properties.append(input_prop) for i in range(obj.channel_start, obj.channel_start + 32): pin_prop = PinProperty("{}_INPUT{}".format(define_prefix, i), "Input {}".format(i), visible=True, readonly=True) pin_prop.category = opts['category'] + " Pinout" pin_prop.transient = True extra_properties.append(pin_prop) for p in self.properties: if isinstance(p, AportScanMaskProperty): prop_obj.other_scan_masks.append(p) p.other_scan_masks.append(prop_obj) elif isinstance(opts['type'], types.AportBondedMode): obj = opts['type'] prop_obj = AportBondedMaskProperty(prop, opts['description'], 0, 0xFFFFFFFF, 0, visible=True, readonly=True) prop_obj.channel_start = obj.channel_start prop_obj.aport = obj.aport define_prefix = prop.rsplit('_', 1)[0] for i in range(obj.channel_start, obj.channel_start + 32): input_prop_name = "{}_{}_ch{}".format(prop_obj.name.lower().rsplit('_', 1)[0], aportname_to_busname(self.name, obj.aport), i % 32) input_prop = CheckboxProperty(input_prop_name, "Enable channel {}".format(i), namespace=self.namespace, visible=False) input_prop.category = opts['category'] input_prop.subcategory = opts['subcategory'] extra_properties.append(input_prop) prop_obj.input_props.append(input_prop) pin_prop = PinProperty("{}_INPUT{}".format(define_prefix, i), "Input {}".format(i), visible=True, readonly=True) pin_prop.category = opts['category'] + " Pinout" pin_prop.transient = True extra_properties.append(pin_prop) else: print("ERROR: unknown property type {} in HAL config model for {}".format(opts['type'], model.name)) if prop_obj is not None: if opts.get('mode') is not None: prop_obj.mode = opts['mode'] if opts.get('generate_if_hidden') is not None: prop_obj.generate_if_hidden = opts['generate_if_hidden'] # Hiding properties that don't belong to the default mode mode_prop = next((prop for prop in self.get_properties() if isinstance(prop, ModeProperty)), None) if mode_prop and mode_prop.hide_properties: if hasattr(prop_obj, 'mode'): if isinstance(prop_obj.mode, list): prop_obj.set_visibility(True if mode_prop.values[0].define_value in prop_obj.mode else False) elif prop_obj.mode: prop_obj.set_visibility(True if prop_obj.mode == mode_prop.values[0].define_value else False) # If _model specifically states visibility, this overrides hiding by default mode if opts.get("visible") is not None: prop_obj.set_visibility(opts['visible']) if opts.get("advanced", False): # Hide advanced properties by default prop_obj.is_advanced = opts.get("advanced", False) prop_obj.set_visibility(False) if opts.get("readonly") is not None: prop_obj.set_readonly(opts['readonly']) if opts.get('defaultValue') is not None: prop_obj.defaultvalue = opts['defaultValue'] if opts.get('overrideDefaultValue') is not None: f = family.get_name().lower() for override_for, value in opts.get('overrideDefaultValue').items(): if f.startswith(override_for.lower()): prop_obj.defaultvalue = value if opts.get('longdescription') is not None: prop_obj.description = opts['longdescription'] elif opts.get("default") is not None: prop_obj.defaultvalue = opts['default'] if opts.get('subcategory') is not None: prop_obj.subcategory = opts['subcategory'] if opts.get('category') is not None: prop_obj.category = opts['category'] if opts.get('allowedconflicts') is not None: prop_obj.allowedconflicts = opts['allowedconflicts'] self.add_property(prop_obj) for property in extra_properties: self.add_property(property) def get_property(self, name): """ Look up property on this module :param name: Regular expression needing to match the name of the property :return: Property if found, None elsewhere """ return next((x for x in self.properties if name == x.name), None) def get_properties(self): ''' :return: Collection of properties in this module ''' return self.properties def activate_runtime(self, state): # Install default hooks for prop in self.properties: if isinstance(prop, ModuleProperty): if prop.inherit_options: RuntimeModel.set_change_handler(prop, RuntimeModel.owning_module_property_callback, on_enable=True) if isinstance(prop, ModeProperty): RuntimeModel.set_change_handler(prop, RuntimeModel.module_mode_callback) if isinstance(prop, PinProperty): if prop.referenced_route is None: RuntimeModel.set_change_handler(prop, RuntimeModel.pin_selection_callback) else: RuntimeModel.configure_route_handler(prop, state) if isinstance(prop, AportBusProperty): RuntimeModel.configure_aport_single_route_handler(prop, state) if isinstance(prop, AportScanMaskProperty): RuntimeModel.configure_aport_scan(prop, state) if isinstance(prop, AportBondedMaskProperty): RuntimeModel.configure_aport_bonded(prop, state) if prop.name == "owner": RuntimeModel.set_change_handler(prop, RuntimeModel.owner_changed_callback) if prop.name == "usedbysdk": RuntimeModel.set_change_handler(prop, RuntimeModel.module_usedbysdk_callback) if prop.name == "hiddenbysdk": RuntimeModel.set_change_handler(prop, RuntimeModel.module_hiddenbysdk_callback) if prop.name == "showadvanced": RuntimeModel.set_change_handler(prop, RuntimeModel.module_showadvanced_callback) if isinstance(prop, PRSChannelProperty): RuntimeModel.set_change_handler(prop, RuntimeModel.prs_channel_changed_callback, on_enable=True) RuntimeModel.set_enable_handler(self, RuntimeModel.module_enabled_callback) # Install user hooks self.set_runtime_hooks() def set_runtime_hooks(self): """ To be overridden by the implementing HAL Config module :return: None """ pass def post_load(self): """ To be overridden by the implementing HAL Config module :return: None """ pass def get_property(mod, property_name): """ Get a property model object by searching for property name :param mod: module on which to look for the property :param property_name: name of the property :return: ExporterModel.Property (or superclass) if found, None else. """ if mod is None: return None prop = mod.get_property(property_name) return prop def override_module(module_list, old, new): """ Override a module in the module_list with another instance :param old: :param new: :return: """ if old.name != new.name: print("ERROR: Not replacing module with same module") return for k,v in enumerate(module_list): if v == old: module_list[k] = new def mask_peripheral_selectors_with_module_list(module_list, module_names): for module_name, module in module_list.items(): for property in module.properties: if isinstance(property, ModuleProperty): property.mask_with_module_list(list(module_names)) def busname_to_aportname(module_name, busname): if 'IDAC' in module_name: idx = 1 elif len(busname) > 2: idx = 0 else: idx = ord(busname[0].upper()) - 64 aportname = "{}{}".format(idx, busname[-1]) return aportname def aportname_to_busname(module_name, aportname): if len(aportname) == 2: diff = aportname[1] aportname = aportname[0] else: diff = '' if 'IDAC' in module_name: busname = 'C' elif aportname == '0': busname = module_name else: busname = chr(ord(aportname) + 16) return "{}{}".format(busname, diff) def clear(): AportScanProperty.scan_props = {}
44.497537
179
0.613949
43,280
0.958264
0
0
0
0
0
0
8,896
0.196967
4b4b052054557003ef2b409b1b9f8cb5ed96012e
48,837
py
Python
tests/schematics_proto3_tests_pb2.py
mlga/schematics-proto3
588fe5bc212e203688166638a1c52dfeda931403
[ "MIT" ]
null
null
null
tests/schematics_proto3_tests_pb2.py
mlga/schematics-proto3
588fe5bc212e203688166638a1c52dfeda931403
[ "MIT" ]
11
2020-04-09T13:33:54.000Z
2020-08-19T17:38:26.000Z
tests/schematics_proto3_tests_pb2.py
mlga/schematics-proto3
588fe5bc212e203688166638a1c52dfeda931403
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: tests/schematics_proto3_tests.proto import sys _b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1')) from google.protobuf.internal import enum_type_wrapper from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() from google.protobuf import timestamp_pb2 as google_dot_protobuf_dot_timestamp__pb2 from google.protobuf import wrappers_pb2 as google_dot_protobuf_dot_wrappers__pb2 DESCRIPTOR = _descriptor.FileDescriptor( name='tests/schematics_proto3_tests.proto', package='schematics_proto3.tests', syntax='proto3', serialized_options=None, serialized_pb=_b('\n#tests/schematics_proto3_tests.proto\x12\x17schematics_proto3.tests\x1a\x1fgoogle/protobuf/timestamp.proto\x1a\x1egoogle/protobuf/wrappers.proto\"e\n\x06Nested\x12\x34\n\x05inner\x18\x01 \x01(\x0b\x32%.schematics_proto3.tests.Nested.Inner\x12\r\n\x05other\x18\x02 \x01(\t\x1a\x16\n\x05Inner\x12\r\n\x05value\x18\x01 \x01(\t\">\n\rWrappedDouble\x12-\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1c.google.protobuf.DoubleValue\"<\n\x0cWrappedFloat\x12,\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1b.google.protobuf.FloatValue\"<\n\x0cWrappedInt64\x12,\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1b.google.protobuf.Int64Value\">\n\rWrappedUInt64\x12-\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1c.google.protobuf.UInt64Value\"<\n\x0cWrappedInt32\x12,\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1b.google.protobuf.Int32Value\">\n\rWrappedUInt32\x12-\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1c.google.protobuf.UInt32Value\":\n\x0bWrappedBool\x12+\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1a.google.protobuf.BoolValue\">\n\rWrappedString\x12-\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1c.google.protobuf.StringValue\"<\n\x0cWrappedBytes\x12,\n\x07wrapped\x18\x01 \x01(\x0b\x32\x1b.google.protobuf.BytesValue\"6\n\tTimestamp\x12)\n\x05value\x18\x01 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\">\n\x11RepeatedTimestamp\x12)\n\x05value\x18\x01 \x03(\x0b\x32\x1a.google.protobuf.Timestamp\"w\n\x0eOneOfTimestamp\x12.\n\x06value1\x18\x01 \x01(\x0b\x32\x1c.google.protobuf.StringValueH\x00\x12,\n\x06value2\x18\x02 \x01(\x0b\x32\x1a.google.protobuf.TimestampH\x00\x42\x07\n\x05inner\"\x17\n\x06\x44ouble\x12\r\n\x05value\x18\x01 \x01(\x01\"\x16\n\x05\x46loat\x12\r\n\x05value\x18\x01 \x01(\x02\"\x16\n\x05Int64\x12\r\n\x05value\x18\x01 \x01(\x03\"\x17\n\x06UInt64\x12\r\n\x05value\x18\x01 \x01(\x04\"\x16\n\x05Int32\x12\r\n\x05value\x18\x01 \x01(\x05\"\x17\n\x06UInt32\x12\r\n\x05value\x18\x01 \x01(\r\"\x15\n\x04\x42ool\x12\r\n\x05value\x18\x01 \x01(\x08\"\x17\n\x06String\x12\r\n\x05value\x18\x01 \x01(\t\"\x16\n\x05\x42ytes\x12\r\n\x05value\x18\x01 \x01(\x0c\"\"\n\x11RepeatedPrimitive\x12\r\n\x05value\x18\x01 \x03(\t\"f\n\x0eRepeatedNested\x12<\n\x05inner\x18\x01 \x03(\x0b\x32-.schematics_proto3.tests.RepeatedNested.Inner\x1a\x16\n\x05Inner\x12\r\n\x05value\x18\x01 \x01(\t\"=\n\x0fRepeatedWrapped\x12*\n\x05value\x18\x01 \x03(\x0b\x32\x1b.google.protobuf.Int32Value\"=\n\x0eOneOfPrimitive\x12\x10\n\x06value1\x18\x01 \x01(\x04H\x00\x12\x10\n\x06value2\x18\x02 \x01(\tH\x00\x42\x07\n\x05inner\"\x9c\x01\n\x0bOneOfNested\x12<\n\x06value1\x18\x01 \x01(\x0b\x32*.schematics_proto3.tests.OneOfNested.InnerH\x00\x12.\n\x06value2\x18\x02 \x01(\x0b\x32\x1c.google.protobuf.StringValueH\x00\x1a\x16\n\x05Inner\x12\r\n\x05value\x18\x01 \x01(\tB\x07\n\x05inner\":\n\nSimpleEnum\x12,\n\x05value\x18\x01 \x01(\x0e\x32\x1d.schematics_proto3.tests.Enum\"<\n\x0cRepeatedEnum\x12,\n\x05value\x18\x01 \x03(\x0e\x32\x1d.schematics_proto3.tests.Enum\"u\n\tOneOfEnum\x12.\n\x06value1\x18\x01 \x01(\x0b\x32\x1c.google.protobuf.StringValueH\x00\x12/\n\x06value2\x18\x02 \x01(\x0e\x32\x1d.schematics_proto3.tests.EnumH\x00\x42\x07\n\x05inner**\n\x04\x45num\x12\x0b\n\x07UNKNOWN\x10\x00\x12\t\n\x05\x46IRST\x10\x01\x12\n\n\x06SECOND\x10\x02\x62\x06proto3') , dependencies=[google_dot_protobuf_dot_timestamp__pb2.DESCRIPTOR,google_dot_protobuf_dot_wrappers__pb2.DESCRIPTOR,]) _ENUM = _descriptor.EnumDescriptor( name='Enum', full_name='schematics_proto3.tests.Enum', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='UNKNOWN', index=0, number=0, serialized_options=None, type=None), _descriptor.EnumValueDescriptor( name='FIRST', index=1, number=1, serialized_options=None, type=None), _descriptor.EnumValueDescriptor( name='SECOND', index=2, number=2, serialized_options=None, type=None), ], containing_type=None, serialized_options=None, serialized_start=1922, serialized_end=1964, ) _sym_db.RegisterEnumDescriptor(_ENUM) Enum = enum_type_wrapper.EnumTypeWrapper(_ENUM) UNKNOWN = 0 FIRST = 1 SECOND = 2 _NESTED_INNER = _descriptor.Descriptor( name='Inner', full_name='schematics_proto3.tests.Nested.Inner', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Nested.Inner.value', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=208, serialized_end=230, ) _NESTED = _descriptor.Descriptor( name='Nested', full_name='schematics_proto3.tests.Nested', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='inner', full_name='schematics_proto3.tests.Nested.inner', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='other', full_name='schematics_proto3.tests.Nested.other', index=1, number=2, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[_NESTED_INNER, ], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=129, serialized_end=230, ) _WRAPPEDDOUBLE = _descriptor.Descriptor( name='WrappedDouble', full_name='schematics_proto3.tests.WrappedDouble', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedDouble.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=232, serialized_end=294, ) _WRAPPEDFLOAT = _descriptor.Descriptor( name='WrappedFloat', full_name='schematics_proto3.tests.WrappedFloat', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedFloat.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=296, serialized_end=356, ) _WRAPPEDINT64 = _descriptor.Descriptor( name='WrappedInt64', full_name='schematics_proto3.tests.WrappedInt64', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedInt64.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=358, serialized_end=418, ) _WRAPPEDUINT64 = _descriptor.Descriptor( name='WrappedUInt64', full_name='schematics_proto3.tests.WrappedUInt64', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedUInt64.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=420, serialized_end=482, ) _WRAPPEDINT32 = _descriptor.Descriptor( name='WrappedInt32', full_name='schematics_proto3.tests.WrappedInt32', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedInt32.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=484, serialized_end=544, ) _WRAPPEDUINT32 = _descriptor.Descriptor( name='WrappedUInt32', full_name='schematics_proto3.tests.WrappedUInt32', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedUInt32.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=546, serialized_end=608, ) _WRAPPEDBOOL = _descriptor.Descriptor( name='WrappedBool', full_name='schematics_proto3.tests.WrappedBool', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedBool.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=610, serialized_end=668, ) _WRAPPEDSTRING = _descriptor.Descriptor( name='WrappedString', full_name='schematics_proto3.tests.WrappedString', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedString.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=670, serialized_end=732, ) _WRAPPEDBYTES = _descriptor.Descriptor( name='WrappedBytes', full_name='schematics_proto3.tests.WrappedBytes', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='wrapped', full_name='schematics_proto3.tests.WrappedBytes.wrapped', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=734, serialized_end=794, ) _TIMESTAMP = _descriptor.Descriptor( name='Timestamp', full_name='schematics_proto3.tests.Timestamp', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Timestamp.value', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=796, serialized_end=850, ) _REPEATEDTIMESTAMP = _descriptor.Descriptor( name='RepeatedTimestamp', full_name='schematics_proto3.tests.RepeatedTimestamp', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.RepeatedTimestamp.value', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=852, serialized_end=914, ) _ONEOFTIMESTAMP = _descriptor.Descriptor( name='OneOfTimestamp', full_name='schematics_proto3.tests.OneOfTimestamp', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value1', full_name='schematics_proto3.tests.OneOfTimestamp.value1', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='value2', full_name='schematics_proto3.tests.OneOfTimestamp.value2', index=1, number=2, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ _descriptor.OneofDescriptor( name='inner', full_name='schematics_proto3.tests.OneOfTimestamp.inner', index=0, containing_type=None, fields=[]), ], serialized_start=916, serialized_end=1035, ) _DOUBLE = _descriptor.Descriptor( name='Double', full_name='schematics_proto3.tests.Double', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Double.value', index=0, number=1, type=1, cpp_type=5, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1037, serialized_end=1060, ) _FLOAT = _descriptor.Descriptor( name='Float', full_name='schematics_proto3.tests.Float', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Float.value', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1062, serialized_end=1084, ) _INT64 = _descriptor.Descriptor( name='Int64', full_name='schematics_proto3.tests.Int64', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Int64.value', index=0, number=1, type=3, cpp_type=2, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1086, serialized_end=1108, ) _UINT64 = _descriptor.Descriptor( name='UInt64', full_name='schematics_proto3.tests.UInt64', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.UInt64.value', index=0, number=1, type=4, cpp_type=4, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1110, serialized_end=1133, ) _INT32 = _descriptor.Descriptor( name='Int32', full_name='schematics_proto3.tests.Int32', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Int32.value', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1135, serialized_end=1157, ) _UINT32 = _descriptor.Descriptor( name='UInt32', full_name='schematics_proto3.tests.UInt32', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.UInt32.value', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1159, serialized_end=1182, ) _BOOL = _descriptor.Descriptor( name='Bool', full_name='schematics_proto3.tests.Bool', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Bool.value', index=0, number=1, type=8, cpp_type=7, label=1, has_default_value=False, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1184, serialized_end=1205, ) _STRING = _descriptor.Descriptor( name='String', full_name='schematics_proto3.tests.String', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.String.value', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1207, serialized_end=1230, ) _BYTES = _descriptor.Descriptor( name='Bytes', full_name='schematics_proto3.tests.Bytes', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.Bytes.value', index=0, number=1, type=12, cpp_type=9, label=1, has_default_value=False, default_value=_b(""), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1232, serialized_end=1254, ) _REPEATEDPRIMITIVE = _descriptor.Descriptor( name='RepeatedPrimitive', full_name='schematics_proto3.tests.RepeatedPrimitive', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.RepeatedPrimitive.value', index=0, number=1, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1256, serialized_end=1290, ) _REPEATEDNESTED_INNER = _descriptor.Descriptor( name='Inner', full_name='schematics_proto3.tests.RepeatedNested.Inner', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.RepeatedNested.Inner.value', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=208, serialized_end=230, ) _REPEATEDNESTED = _descriptor.Descriptor( name='RepeatedNested', full_name='schematics_proto3.tests.RepeatedNested', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='inner', full_name='schematics_proto3.tests.RepeatedNested.inner', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[_REPEATEDNESTED_INNER, ], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1292, serialized_end=1394, ) _REPEATEDWRAPPED = _descriptor.Descriptor( name='RepeatedWrapped', full_name='schematics_proto3.tests.RepeatedWrapped', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.RepeatedWrapped.value', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1396, serialized_end=1457, ) _ONEOFPRIMITIVE = _descriptor.Descriptor( name='OneOfPrimitive', full_name='schematics_proto3.tests.OneOfPrimitive', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value1', full_name='schematics_proto3.tests.OneOfPrimitive.value1', index=0, number=1, type=4, cpp_type=4, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='value2', full_name='schematics_proto3.tests.OneOfPrimitive.value2', index=1, number=2, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ _descriptor.OneofDescriptor( name='inner', full_name='schematics_proto3.tests.OneOfPrimitive.inner', index=0, containing_type=None, fields=[]), ], serialized_start=1459, serialized_end=1520, ) _ONEOFNESTED_INNER = _descriptor.Descriptor( name='Inner', full_name='schematics_proto3.tests.OneOfNested.Inner', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.OneOfNested.Inner.value', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=208, serialized_end=230, ) _ONEOFNESTED = _descriptor.Descriptor( name='OneOfNested', full_name='schematics_proto3.tests.OneOfNested', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value1', full_name='schematics_proto3.tests.OneOfNested.value1', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='value2', full_name='schematics_proto3.tests.OneOfNested.value2', index=1, number=2, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[_ONEOFNESTED_INNER, ], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ _descriptor.OneofDescriptor( name='inner', full_name='schematics_proto3.tests.OneOfNested.inner', index=0, containing_type=None, fields=[]), ], serialized_start=1523, serialized_end=1679, ) _SIMPLEENUM = _descriptor.Descriptor( name='SimpleEnum', full_name='schematics_proto3.tests.SimpleEnum', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.SimpleEnum.value', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1681, serialized_end=1739, ) _REPEATEDENUM = _descriptor.Descriptor( name='RepeatedEnum', full_name='schematics_proto3.tests.RepeatedEnum', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value', full_name='schematics_proto3.tests.RepeatedEnum.value', index=0, number=1, type=14, cpp_type=8, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=1741, serialized_end=1801, ) _ONEOFENUM = _descriptor.Descriptor( name='OneOfEnum', full_name='schematics_proto3.tests.OneOfEnum', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='value1', full_name='schematics_proto3.tests.OneOfEnum.value1', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='value2', full_name='schematics_proto3.tests.OneOfEnum.value2', index=1, number=2, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ _descriptor.OneofDescriptor( name='inner', full_name='schematics_proto3.tests.OneOfEnum.inner', index=0, containing_type=None, fields=[]), ], serialized_start=1803, serialized_end=1920, ) _NESTED_INNER.containing_type = _NESTED _NESTED.fields_by_name['inner'].message_type = _NESTED_INNER _WRAPPEDDOUBLE.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._DOUBLEVALUE _WRAPPEDFLOAT.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._FLOATVALUE _WRAPPEDINT64.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._INT64VALUE _WRAPPEDUINT64.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._UINT64VALUE _WRAPPEDINT32.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._INT32VALUE _WRAPPEDUINT32.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._UINT32VALUE _WRAPPEDBOOL.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._BOOLVALUE _WRAPPEDSTRING.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._STRINGVALUE _WRAPPEDBYTES.fields_by_name['wrapped'].message_type = google_dot_protobuf_dot_wrappers__pb2._BYTESVALUE _TIMESTAMP.fields_by_name['value'].message_type = google_dot_protobuf_dot_timestamp__pb2._TIMESTAMP _REPEATEDTIMESTAMP.fields_by_name['value'].message_type = google_dot_protobuf_dot_timestamp__pb2._TIMESTAMP _ONEOFTIMESTAMP.fields_by_name['value1'].message_type = google_dot_protobuf_dot_wrappers__pb2._STRINGVALUE _ONEOFTIMESTAMP.fields_by_name['value2'].message_type = google_dot_protobuf_dot_timestamp__pb2._TIMESTAMP _ONEOFTIMESTAMP.oneofs_by_name['inner'].fields.append( _ONEOFTIMESTAMP.fields_by_name['value1']) _ONEOFTIMESTAMP.fields_by_name['value1'].containing_oneof = _ONEOFTIMESTAMP.oneofs_by_name['inner'] _ONEOFTIMESTAMP.oneofs_by_name['inner'].fields.append( _ONEOFTIMESTAMP.fields_by_name['value2']) _ONEOFTIMESTAMP.fields_by_name['value2'].containing_oneof = _ONEOFTIMESTAMP.oneofs_by_name['inner'] _REPEATEDNESTED_INNER.containing_type = _REPEATEDNESTED _REPEATEDNESTED.fields_by_name['inner'].message_type = _REPEATEDNESTED_INNER _REPEATEDWRAPPED.fields_by_name['value'].message_type = google_dot_protobuf_dot_wrappers__pb2._INT32VALUE _ONEOFPRIMITIVE.oneofs_by_name['inner'].fields.append( _ONEOFPRIMITIVE.fields_by_name['value1']) _ONEOFPRIMITIVE.fields_by_name['value1'].containing_oneof = _ONEOFPRIMITIVE.oneofs_by_name['inner'] _ONEOFPRIMITIVE.oneofs_by_name['inner'].fields.append( _ONEOFPRIMITIVE.fields_by_name['value2']) _ONEOFPRIMITIVE.fields_by_name['value2'].containing_oneof = _ONEOFPRIMITIVE.oneofs_by_name['inner'] _ONEOFNESTED_INNER.containing_type = _ONEOFNESTED _ONEOFNESTED.fields_by_name['value1'].message_type = _ONEOFNESTED_INNER _ONEOFNESTED.fields_by_name['value2'].message_type = google_dot_protobuf_dot_wrappers__pb2._STRINGVALUE _ONEOFNESTED.oneofs_by_name['inner'].fields.append( _ONEOFNESTED.fields_by_name['value1']) _ONEOFNESTED.fields_by_name['value1'].containing_oneof = _ONEOFNESTED.oneofs_by_name['inner'] _ONEOFNESTED.oneofs_by_name['inner'].fields.append( _ONEOFNESTED.fields_by_name['value2']) _ONEOFNESTED.fields_by_name['value2'].containing_oneof = _ONEOFNESTED.oneofs_by_name['inner'] _SIMPLEENUM.fields_by_name['value'].enum_type = _ENUM _REPEATEDENUM.fields_by_name['value'].enum_type = _ENUM _ONEOFENUM.fields_by_name['value1'].message_type = google_dot_protobuf_dot_wrappers__pb2._STRINGVALUE _ONEOFENUM.fields_by_name['value2'].enum_type = _ENUM _ONEOFENUM.oneofs_by_name['inner'].fields.append( _ONEOFENUM.fields_by_name['value1']) _ONEOFENUM.fields_by_name['value1'].containing_oneof = _ONEOFENUM.oneofs_by_name['inner'] _ONEOFENUM.oneofs_by_name['inner'].fields.append( _ONEOFENUM.fields_by_name['value2']) _ONEOFENUM.fields_by_name['value2'].containing_oneof = _ONEOFENUM.oneofs_by_name['inner'] DESCRIPTOR.message_types_by_name['Nested'] = _NESTED DESCRIPTOR.message_types_by_name['WrappedDouble'] = _WRAPPEDDOUBLE DESCRIPTOR.message_types_by_name['WrappedFloat'] = _WRAPPEDFLOAT DESCRIPTOR.message_types_by_name['WrappedInt64'] = _WRAPPEDINT64 DESCRIPTOR.message_types_by_name['WrappedUInt64'] = _WRAPPEDUINT64 DESCRIPTOR.message_types_by_name['WrappedInt32'] = _WRAPPEDINT32 DESCRIPTOR.message_types_by_name['WrappedUInt32'] = _WRAPPEDUINT32 DESCRIPTOR.message_types_by_name['WrappedBool'] = _WRAPPEDBOOL DESCRIPTOR.message_types_by_name['WrappedString'] = _WRAPPEDSTRING DESCRIPTOR.message_types_by_name['WrappedBytes'] = _WRAPPEDBYTES DESCRIPTOR.message_types_by_name['Timestamp'] = _TIMESTAMP DESCRIPTOR.message_types_by_name['RepeatedTimestamp'] = _REPEATEDTIMESTAMP DESCRIPTOR.message_types_by_name['OneOfTimestamp'] = _ONEOFTIMESTAMP DESCRIPTOR.message_types_by_name['Double'] = _DOUBLE DESCRIPTOR.message_types_by_name['Float'] = _FLOAT DESCRIPTOR.message_types_by_name['Int64'] = _INT64 DESCRIPTOR.message_types_by_name['UInt64'] = _UINT64 DESCRIPTOR.message_types_by_name['Int32'] = _INT32 DESCRIPTOR.message_types_by_name['UInt32'] = _UINT32 DESCRIPTOR.message_types_by_name['Bool'] = _BOOL DESCRIPTOR.message_types_by_name['String'] = _STRING DESCRIPTOR.message_types_by_name['Bytes'] = _BYTES DESCRIPTOR.message_types_by_name['RepeatedPrimitive'] = _REPEATEDPRIMITIVE DESCRIPTOR.message_types_by_name['RepeatedNested'] = _REPEATEDNESTED DESCRIPTOR.message_types_by_name['RepeatedWrapped'] = _REPEATEDWRAPPED DESCRIPTOR.message_types_by_name['OneOfPrimitive'] = _ONEOFPRIMITIVE DESCRIPTOR.message_types_by_name['OneOfNested'] = _ONEOFNESTED DESCRIPTOR.message_types_by_name['SimpleEnum'] = _SIMPLEENUM DESCRIPTOR.message_types_by_name['RepeatedEnum'] = _REPEATEDENUM DESCRIPTOR.message_types_by_name['OneOfEnum'] = _ONEOFENUM DESCRIPTOR.enum_types_by_name['Enum'] = _ENUM _sym_db.RegisterFileDescriptor(DESCRIPTOR) Nested = _reflection.GeneratedProtocolMessageType('Nested', (_message.Message,), dict( Inner = _reflection.GeneratedProtocolMessageType('Inner', (_message.Message,), dict( DESCRIPTOR = _NESTED_INNER, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Nested.Inner) )) , DESCRIPTOR = _NESTED, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Nested) )) _sym_db.RegisterMessage(Nested) _sym_db.RegisterMessage(Nested.Inner) WrappedDouble = _reflection.GeneratedProtocolMessageType('WrappedDouble', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDDOUBLE, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedDouble) )) _sym_db.RegisterMessage(WrappedDouble) WrappedFloat = _reflection.GeneratedProtocolMessageType('WrappedFloat', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDFLOAT, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedFloat) )) _sym_db.RegisterMessage(WrappedFloat) WrappedInt64 = _reflection.GeneratedProtocolMessageType('WrappedInt64', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDINT64, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedInt64) )) _sym_db.RegisterMessage(WrappedInt64) WrappedUInt64 = _reflection.GeneratedProtocolMessageType('WrappedUInt64', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDUINT64, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedUInt64) )) _sym_db.RegisterMessage(WrappedUInt64) WrappedInt32 = _reflection.GeneratedProtocolMessageType('WrappedInt32', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDINT32, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedInt32) )) _sym_db.RegisterMessage(WrappedInt32) WrappedUInt32 = _reflection.GeneratedProtocolMessageType('WrappedUInt32', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDUINT32, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedUInt32) )) _sym_db.RegisterMessage(WrappedUInt32) WrappedBool = _reflection.GeneratedProtocolMessageType('WrappedBool', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDBOOL, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedBool) )) _sym_db.RegisterMessage(WrappedBool) WrappedString = _reflection.GeneratedProtocolMessageType('WrappedString', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDSTRING, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedString) )) _sym_db.RegisterMessage(WrappedString) WrappedBytes = _reflection.GeneratedProtocolMessageType('WrappedBytes', (_message.Message,), dict( DESCRIPTOR = _WRAPPEDBYTES, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.WrappedBytes) )) _sym_db.RegisterMessage(WrappedBytes) Timestamp = _reflection.GeneratedProtocolMessageType('Timestamp', (_message.Message,), dict( DESCRIPTOR = _TIMESTAMP, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Timestamp) )) _sym_db.RegisterMessage(Timestamp) RepeatedTimestamp = _reflection.GeneratedProtocolMessageType('RepeatedTimestamp', (_message.Message,), dict( DESCRIPTOR = _REPEATEDTIMESTAMP, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.RepeatedTimestamp) )) _sym_db.RegisterMessage(RepeatedTimestamp) OneOfTimestamp = _reflection.GeneratedProtocolMessageType('OneOfTimestamp', (_message.Message,), dict( DESCRIPTOR = _ONEOFTIMESTAMP, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.OneOfTimestamp) )) _sym_db.RegisterMessage(OneOfTimestamp) Double = _reflection.GeneratedProtocolMessageType('Double', (_message.Message,), dict( DESCRIPTOR = _DOUBLE, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Double) )) _sym_db.RegisterMessage(Double) Float = _reflection.GeneratedProtocolMessageType('Float', (_message.Message,), dict( DESCRIPTOR = _FLOAT, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Float) )) _sym_db.RegisterMessage(Float) Int64 = _reflection.GeneratedProtocolMessageType('Int64', (_message.Message,), dict( DESCRIPTOR = _INT64, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Int64) )) _sym_db.RegisterMessage(Int64) UInt64 = _reflection.GeneratedProtocolMessageType('UInt64', (_message.Message,), dict( DESCRIPTOR = _UINT64, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.UInt64) )) _sym_db.RegisterMessage(UInt64) Int32 = _reflection.GeneratedProtocolMessageType('Int32', (_message.Message,), dict( DESCRIPTOR = _INT32, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Int32) )) _sym_db.RegisterMessage(Int32) UInt32 = _reflection.GeneratedProtocolMessageType('UInt32', (_message.Message,), dict( DESCRIPTOR = _UINT32, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.UInt32) )) _sym_db.RegisterMessage(UInt32) Bool = _reflection.GeneratedProtocolMessageType('Bool', (_message.Message,), dict( DESCRIPTOR = _BOOL, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Bool) )) _sym_db.RegisterMessage(Bool) String = _reflection.GeneratedProtocolMessageType('String', (_message.Message,), dict( DESCRIPTOR = _STRING, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.String) )) _sym_db.RegisterMessage(String) Bytes = _reflection.GeneratedProtocolMessageType('Bytes', (_message.Message,), dict( DESCRIPTOR = _BYTES, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.Bytes) )) _sym_db.RegisterMessage(Bytes) RepeatedPrimitive = _reflection.GeneratedProtocolMessageType('RepeatedPrimitive', (_message.Message,), dict( DESCRIPTOR = _REPEATEDPRIMITIVE, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.RepeatedPrimitive) )) _sym_db.RegisterMessage(RepeatedPrimitive) RepeatedNested = _reflection.GeneratedProtocolMessageType('RepeatedNested', (_message.Message,), dict( Inner = _reflection.GeneratedProtocolMessageType('Inner', (_message.Message,), dict( DESCRIPTOR = _REPEATEDNESTED_INNER, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.RepeatedNested.Inner) )) , DESCRIPTOR = _REPEATEDNESTED, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.RepeatedNested) )) _sym_db.RegisterMessage(RepeatedNested) _sym_db.RegisterMessage(RepeatedNested.Inner) RepeatedWrapped = _reflection.GeneratedProtocolMessageType('RepeatedWrapped', (_message.Message,), dict( DESCRIPTOR = _REPEATEDWRAPPED, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.RepeatedWrapped) )) _sym_db.RegisterMessage(RepeatedWrapped) OneOfPrimitive = _reflection.GeneratedProtocolMessageType('OneOfPrimitive', (_message.Message,), dict( DESCRIPTOR = _ONEOFPRIMITIVE, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.OneOfPrimitive) )) _sym_db.RegisterMessage(OneOfPrimitive) OneOfNested = _reflection.GeneratedProtocolMessageType('OneOfNested', (_message.Message,), dict( Inner = _reflection.GeneratedProtocolMessageType('Inner', (_message.Message,), dict( DESCRIPTOR = _ONEOFNESTED_INNER, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.OneOfNested.Inner) )) , DESCRIPTOR = _ONEOFNESTED, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.OneOfNested) )) _sym_db.RegisterMessage(OneOfNested) _sym_db.RegisterMessage(OneOfNested.Inner) SimpleEnum = _reflection.GeneratedProtocolMessageType('SimpleEnum', (_message.Message,), dict( DESCRIPTOR = _SIMPLEENUM, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.SimpleEnum) )) _sym_db.RegisterMessage(SimpleEnum) RepeatedEnum = _reflection.GeneratedProtocolMessageType('RepeatedEnum', (_message.Message,), dict( DESCRIPTOR = _REPEATEDENUM, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.RepeatedEnum) )) _sym_db.RegisterMessage(RepeatedEnum) OneOfEnum = _reflection.GeneratedProtocolMessageType('OneOfEnum', (_message.Message,), dict( DESCRIPTOR = _ONEOFENUM, __module__ = 'tests.schematics_proto3_tests_pb2' # @@protoc_insertion_point(class_scope:schematics_proto3.tests.OneOfEnum) )) _sym_db.RegisterMessage(OneOfEnum) # @@protoc_insertion_point(module_scope)
33.750518
3,198
0.757377
0
0
0
0
0
0
0
0
12,389
0.253681
4b4dceb98d231438a803f497f8f31de32f299051
241
py
Python
sps_demo/accounts/api/serializers.py
JuanDM93/sps_django
df47c7ee63a1e99468644a6f428a6cdabc7ac6ae
[ "MIT" ]
null
null
null
sps_demo/accounts/api/serializers.py
JuanDM93/sps_django
df47c7ee63a1e99468644a6f428a6cdabc7ac6ae
[ "MIT" ]
1
2021-07-27T06:46:05.000Z
2021-07-27T06:46:05.000Z
sps_demo/accounts/api/serializers.py
JuanDM93/sps_django
df47c7ee63a1e99468644a6f428a6cdabc7ac6ae
[ "MIT" ]
null
null
null
from rest_framework.serializers import ModelSerializer from accounts.models import Account class AccountSerializer(ModelSerializer): class Meta: model = Account fields = [ 'account_id', 'limit', ]
21.909091
54
0.6639
145
0.60166
0
0
0
0
0
0
19
0.078838
4b4ddf5eeb83ed879035c41d407475a7baf89592
6,320
py
Python
benchmark/HIGGS/explore/contour_nll.py
victor-estrade/SystGradDescent
822e7094290301ec47a99433381a8d6406798aff
[ "MIT" ]
2
2019-03-20T09:05:02.000Z
2019-03-20T15:23:44.000Z
benchmark/HIGGS/explore/contour_nll.py
victor-estrade/SystGradDescent
822e7094290301ec47a99433381a8d6406798aff
[ "MIT" ]
null
null
null
benchmark/HIGGS/explore/contour_nll.py
victor-estrade/SystGradDescent
822e7094290301ec47a99433381a8d6406798aff
[ "MIT" ]
null
null
null
# coding: utf-8 from __future__ import division from __future__ import print_function from __future__ import absolute_import from __future__ import unicode_literals import os import logging import datetime import numpy as np import pandas as pd import matplotlib.pyplot as plt import seaborn as sns from config import SAVING_DIR from config import SEED from visual import set_plot_config set_plot_config() from utils.log import set_logger from utils.log import flush from utils.log import print_line from utils.evaluation import evaluate_minuit from problem.higgs import HiggsConfigTesOnly as Config from problem.higgs import get_minimizer from problem.higgs import get_minimizer_no_nuisance from problem.higgs import get_generators_torch from problem.higgs import HiggsNLL as NLLComputer from ..common import N_BINS def do_iter(config, model, i_iter, valid_generator, test_generator, root_dir, n_bins=N_BINS): logger = logging.getLogger() directory = os.path.join(root_dir, model.name, f"iter_{i_iter}") os.makedirs(directory, exist_ok=True) logger.info(f"saving dir = {directory}") logger.info('Generate testing data') X_test, y_test, w_test = test_generator.generate(*config.TRUE, n_samples=config.N_TESTING_SAMPLES, no_grad=True) logger.info('Set up NLL computer') compute_summaries = model.summary_computer(n_bins=n_bins) compute_nll = NLLComputer(compute_summaries, valid_generator, X_test, w_test, config=config) basic_check(compute_nll, config) basic_contourplot(compute_nll, config, directory) # MINIMIZE NLL logger.info('Prepare minuit minimizer') minimizer = get_minimizer(compute_nll, config.CALIBRATED, config.CALIBRATED_ERROR) some_dict = evaluate_minuit(minimizer, config.TRUE, directory, suffix="") # FOCUSED contour plot nll_func = lambda mu, tes : compute_nll(tes, config.TRUE.jes, config.TRUE.les, mu) x = minimizer.values[3] y = minimizer.values[0] x_err = minimizer.errors[3] y_err = minimizer.errors[0] focused_contour(x, y, x_err, y_err, nll_func, directory, xlabel="mu", ylabel='tes') nll_func = lambda mu, jes : compute_nll(config.TRUE.tes, jes, config.TRUE.les, mu) x = minimizer.values[3] y = minimizer.values[1] x_err = minimizer.errors[3] y_err = minimizer.errors[1] focused_contour(x, y, x_err, y_err, nll_func, directory, xlabel="mu", ylabel='jes') nll_func = lambda mu, les : compute_nll(config.TRUE.tes, config.TRUE.jes, les, mu) x = minimizer.values[3] y = minimizer.values[2] x_err = minimizer.errors[3] y_err = minimizer.errors[2] focused_contour(x, y, x_err, y_err, nll_func, directory, xlabel="mu", ylabel='les') def basic_check(compute_nll, config): logger = logging.getLogger() nll = compute_nll(*config.CALIBRATED) logger.info(f"Calib nll = {nll}") nll = compute_nll(*config.TRUE) logger.info(f"TRUE nll = {nll}") def basic_contourplot(compute_nll, config, directory): logger = logging.getLogger() ARRAY_SIZE = 10 # MESH NLL logger.info(f"basic mu-tes contour plot...") mu_array = np.linspace(0.5, 1.5, ARRAY_SIZE) tes_array = np.linspace(0.95, 1.05, ARRAY_SIZE) mu_mesh, tes_mesh = np.meshgrid(mu_array, tes_array) nll_func = lambda mu, tes : compute_nll(tes, config.TRUE.jes, config.TRUE.les, mu) nll_mesh = np.array([nll_func(mu, tes) for mu, tes in zip(mu_mesh.ravel(), tes_mesh.ravel())]).reshape(mu_mesh.shape) plot_contour(mu_mesh, tes_mesh, nll_mesh, directory, xlabel="mu", ylabel="tes") logger.info(f"basic mu-jes contour plot...") jes_array = np.linspace(0.95, 1.05, ARRAY_SIZE) mu_mesh, jes_mesh = np.meshgrid(mu_array, jes_array) nll_func = lambda mu, jes : compute_nll(config.TRUE.tes, jes, config.TRUE.les, mu) nll_mesh = np.array([nll_func(mu, jes) for mu, jes in zip(mu_mesh.ravel(), jes_mesh.ravel())]).reshape(mu_mesh.shape) plot_contour(mu_mesh, jes_mesh, nll_mesh, directory, xlabel="mu", ylabel="jes") logger.info(f"basic mu-les contour plot...") les_array = np.linspace(0.95, 1.05, ARRAY_SIZE) mu_mesh, les_mesh = np.meshgrid(mu_array, les_array) nll_func = lambda mu, les : compute_nll(config.TRUE.tes, config.TRUE.jes, les, mu) nll_mesh = np.array([nll_func(mu, les) for mu, les in zip(mu_mesh.ravel(), les_mesh.ravel())]).reshape(mu_mesh.shape) plot_contour(mu_mesh, les_mesh, nll_mesh, directory, xlabel="mu", ylabel="les") logger.info(f"basic tes-jes contour plot...") tes_mesh, jes_mesh = np.meshgrid(tes_array, jes_array) nll_func = lambda tes, jes : compute_nll(tes, jes, config.TRUE.les, config.TRUE.mu) nll_mesh = np.array([nll_func(tes, jes) for tes, jes in zip(tes_mesh.ravel(), jes_mesh.ravel())]).reshape(tes_mesh.shape) plot_contour(tes_mesh, jes_mesh, nll_mesh, directory, xlabel="tes", ylabel="jes") def plot_contour(x, y, z, directory, xlabel="mu", ylabel="tes"): logger = logging.getLogger() fig, ax = plt.subplots() CS = ax.contour(x, y, z) ax.clabel(CS, inline=1, fontsize=10) ax.set_xlabel(xlabel) ax.set_ylabel(ylabel) now = datetime.datetime.now().strftime("%Y-%m-%d_%H-%M-%S\n") fig.title(now) fname = f"{xlabel}-{ylabel}_contour_plot.png" path = os.path.join(directory, fname) plt.savefig(path) plt.clf() plt.close(fig) logger.info(f"saved at {path}") def focused_contour(x, y, x_err, y_err, nll_func, directory, xlabel="mu", ylabel='tes'): logger = logging.getLogger() ARRAY_SIZE = 10 logger.info(f"focused {xlabel}-{ylabel} contour plot...") x_array = np.linspace(x-3*x_err, x+3*x_err, ARRAY_SIZE) y_array = np.linspace(y-3*y_err, y+3*y_err, ARRAY_SIZE) x_mesh, y_mesh = np.meshgrid(x_array, y_array) z_mesh = np.array([nll_func(x, y) for x, y in zip(x_mesh.ravel(), y_mesh.ravel())]).reshape(x_mesh.shape) fig, ax = plt.subplots() CS = ax.contour(x_mesh, y_mesh, z_mesh) ax.clabel(CS, inline=1, fontsize=10) ax.set_xlabel(xlabel) ax.set_ylabel(ylabel) now = datetime.datetime.now().strftime("%Y-%m-%d_%H-%M-%S\n") fig.title(now) fname = f"{xlabel}-{ylabel}_focused_contour_plot.png" path = os.path.join(directory, fname) plt.savefig(path) plt.clf() plt.close(fig) logger.info(f"saved at {path}")
39.012346
125
0.708861
0
0
0
0
0
0
0
0
626
0.099051
4b4e8e6fb685efd1c0bdeed695e0d638ddd30af1
1,376
py
Python
backend/framework/qlf/dashboard/migrations/0012_auto_20180921_1717.py
desihub/qlf
a9c455f7aee41d7901c89ae90dd821c617340a86
[ "BSD-3-Clause" ]
8
2017-09-08T00:24:20.000Z
2019-02-03T07:31:03.000Z
backend/framework/qlf/dashboard/migrations/0012_auto_20180921_1717.py
desihub/qlf
a9c455f7aee41d7901c89ae90dd821c617340a86
[ "BSD-3-Clause" ]
77
2017-06-15T21:39:09.000Z
2019-07-13T19:41:27.000Z
backend/framework/qlf/dashboard/migrations/0012_auto_20180921_1717.py
desihub/qlf
a9c455f7aee41d7901c89ae90dd821c617340a86
[ "BSD-3-Clause" ]
5
2017-09-10T02:25:03.000Z
2019-02-06T20:55:59.000Z
# -*- coding: utf-8 -*- # Generated by Django 1.11 on 2018-09-21 17:17 from __future__ import unicode_literals import django.contrib.postgres.fields from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('dashboard', '0011_auto_20180727_1800'), ] operations = [ migrations.CreateModel( name='Fibermap', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('ra_obs', django.contrib.postgres.fields.ArrayField(base_field=models.FloatField(), size=None)), ('dec_obs', django.contrib.postgres.fields.ArrayField(base_field=models.FloatField(), size=None)), ('fiber', django.contrib.postgres.fields.ArrayField(base_field=models.FloatField(), size=None)), ('objtype', django.contrib.postgres.fields.ArrayField(base_field=models.CharField(max_length=15), size=None)), ('exposure', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='fibermap_exposure', to='dashboard.Exposure')), ], ), migrations.RemoveField( model_name='qa', name='job', ), migrations.DeleteModel( name='QA', ), ]
38.222222
152
0.635174
1,149
0.835029
0
0
0
0
0
0
218
0.15843
4b53328e075db009dbb8d21c3c121da0a2ce955a
476
py
Python
qcodes/instrument_drivers/rohde_schwarz/HMC8042.py
LGruenhaupt/Qcodes
ffb74dae53c13c4885e61b5a2df3f833d524de04
[ "MIT" ]
1
2019-12-07T01:25:49.000Z
2019-12-07T01:25:49.000Z
qcodes/instrument_drivers/rohde_schwarz/HMC8042.py
Dominik-Vogel/Qcodes
b4cf7d58bc1bf3be97af6bf48f57cb6b87d588bb
[ "MIT" ]
12
2020-10-13T16:53:37.000Z
2020-10-14T17:16:22.000Z
qcodes/instrument_drivers/rohde_schwarz/HMC8042.py
Dominik-Vogel/Qcodes
b4cf7d58bc1bf3be97af6bf48f57cb6b87d588bb
[ "MIT" ]
1
2020-05-03T22:47:40.000Z
2020-05-03T22:47:40.000Z
from .private.HMC804x import _RohdeSchwarzHMC804x from qcodes.utils.deprecate import deprecate_moved_to_qcd @deprecate_moved_to_qcd(alternative="qcodes_contrib_drivers.drivers.RohdeSchwarz.HMC8042.RohdeSchwarzHMC8042") class RohdeSchwarzHMC8042(_RohdeSchwarzHMC804x): """ This is the qcodes driver for the Rohde & Schwarz HMC8042 Power Supply """ def __init__(self, name, address, **kwargs): super().__init__(name, address, num_channels=2, **kwargs)
39.666667
110
0.781513
254
0.533613
0
0
365
0.766807
0
0
159
0.334034
4b5611952b114a3d2cf44eadfe8d22e693d8c643
682
py
Python
python_fundamentals/Multiple_Sum_Average/index.py
justnclrk/Python
0922961cbd94694a69ae8132a5c33baf552d8d89
[ "MIT" ]
null
null
null
python_fundamentals/Multiple_Sum_Average/index.py
justnclrk/Python
0922961cbd94694a69ae8132a5c33baf552d8d89
[ "MIT" ]
8
2020-06-06T01:02:06.000Z
2022-03-12T00:24:13.000Z
python_fundamentals/Multiple_Sum_Average/index.py
justnclrk/Python
0922961cbd94694a69ae8132a5c33baf552d8d89
[ "MIT" ]
null
null
null
# Multiples -- Part I - Write code that prints all the odd numbers from 1 to 1000. Use the for loop and don't use a list to do this exercise for i in range(1, 1000, 2): print(i) # Multiples -- Part II - Create another program that prints all the multiples of 5 from 5 to 1,000,000 for m in range(5, 1000000, 5): print(m) # Sum List -- Create a program that prints the sum of all the values in the list: a = [1, 2, 5, 10, 255, 3] a = [1, 2, 5, 10, 255, 3] b = sum(a) print(b) # Average List -- Create a program that prints the average of the values in the list: c = [1, 2, 5, 10, 255, 3] c = [1, 2, 5, 10, 255, 3] dSum = sum(c) eLen = len(c) fAvg = (dSum / eLen) print(fAvg)
40.117647
140
0.64956
0
0
0
0
0
0
0
0
460
0.674487
4b5646cef1fca290360a2f8a03244f3cf60a9b62
2,817
py
Python
examples/gen9_valset_test.py
mgesteiro/pyubx2
02fd8fa2863b88ed2d746b5800717a1b6b213181
[ "BSD-3-Clause" ]
null
null
null
examples/gen9_valset_test.py
mgesteiro/pyubx2
02fd8fa2863b88ed2d746b5800717a1b6b213181
[ "BSD-3-Clause" ]
null
null
null
examples/gen9_valset_test.py
mgesteiro/pyubx2
02fd8fa2863b88ed2d746b5800717a1b6b213181
[ "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python3 """ Demo example to test CFG-VALSET ublox message - generation 9 @author: mgesteiro """ import sys import time from serial import Serial, SerialException, SerialTimeoutException from pyubx2 import ( UBXMessage, GET, SET, VALSET_RAM, UBXMessageError, UBXTypeError, UBXParseError, ) def message_valsetuart1baudrate_set(baudrate): """ Function to generate a CFG-VALSET CFG-UART1-BAUDRATE set UBX message """ # https://www.u-blox.com/en/docs/UBX-18010854#page=86&zoom=auto,-74,499 # CFG-UART1-BAUDRATE Key = 0x40520001 return UBXMessage( "CFG", "CFG-VALSET", SET, payload=b"\x00" + VALSET_RAM # version + int(0).to_bytes(2, byteorder="little", signed=False) # layers + 0x40520001 .to_bytes(4, byteorder="little", signed=False) # reserved0 + baudrate.to_bytes(4, byteorder="little", signed=False), # key # value ) def message_valsetuart1baudrate_response(): """ Function to generate a ACK-ACK-ACK UBX message """ # https://www.u-blox.com/en/docs/UBX-18010854#page=52&zoom=auto,-74,379 return UBXMessage("ACK", "ACK-ACK", GET, clsID=0x06, msgID=0x8A) if __name__ == "__main__": PORTNAME = "/dev/tty.usbserial-A50285BI" BAUDRATE = 230400 try: print("\nBuilding CFG-UART1-BAUDRATE VALSET message:") msg = message_valsetuart1baudrate_set(BAUDRATE) print(f" GENERATED: {msg.serialize().hex()}") print( " EXPECTED: b562068a0c00000100000100524000840300b7ef" + " (Note: valid for 230400 baudrate)" ) print(f" {msg}\n") print(f"This demo will now set your module's UART1 to {BAUDRATE} (only in RAM)") try: input("press <ENTER> to continue, CTRL-C to abort!\n") except KeyboardInterrupt: print("\nExecution aborted.\n") sys.exit(0) sport = Serial(PORTNAME, BAUDRATE, timeout=2) time.sleep(0.250) # stabilize print( f"Sending set message to {PORTNAME} at {BAUDRATE} " + "(edit the code to change these values)\n" ) sport.flushInput() sport.write(msg.serialize()) print("Receiving response ...") raw = sport.read(512) START = raw.find(b"\xB5\x62") data = raw[START : START + 10] # expected ACK msg = message_valsetuart1baudrate_response() print(f" RECEIVED: {data.hex()}") print(f" EXPECTED: {msg.serialize().hex()}") print(f" {UBXMessage.parse(data)}\n") except ( UBXMessageError, UBXTypeError, UBXParseError, SerialException, SerialTimeoutException, ) as err: print(f"Something broke 💥🤷‍♂️: {err}\n")
28.17
88
0.606674
0
0
0
0
0
0
0
0
1,205
0.425946
4b57edd76cfedc441b5ed69fe2a9fd78c4dbd2d2
3,996
py
Python
main.py
saswatsamal/Snake-Game
2c0f427fd6001f09d26a4586ce55453af706c355
[ "CC0-1.0" ]
2
2021-04-25T07:34:14.000Z
2021-04-30T15:24:55.000Z
main.py
saswatsamal/Snake-Game
2c0f427fd6001f09d26a4586ce55453af706c355
[ "CC0-1.0" ]
null
null
null
main.py
saswatsamal/Snake-Game
2c0f427fd6001f09d26a4586ce55453af706c355
[ "CC0-1.0" ]
null
null
null
import pygame import time import sys, random pygame.init() yellow = (255, 255, 102) green = (0, 255, 0) black = (0,0,0) width = 1280 height = 720 gameDisplay = pygame.display.set_mode((width, height)) pygame.display.set_caption('Snake Game By Saswat Samal') clock = pygame.time.Clock() snake_block = 10 snake_speed = 15 font_style = pygame.font.SysFont("ubuntu", 25) score_font = pygame.font.SysFont("ubuntu", 20) def main_menu(): while 1: for event in pygame.event.get(): if event.type == pygame.QUIT: pygame.quit() sys.exit() if event.type == pygame.MOUSEBUTTONDOWN: gameLoop() gameDisplay.fill(black) main_menu_message = font_style.render('Press anywhere to start the game' , True , (255,255,255)) font_pos = main_menu_message.get_rect(center=(width//2, height//2)) gameDisplay.blit(main_menu_message , font_pos) pygame.display.update() def gameScore(score): value = score_font.render("Your Score: " + str(score), True, green) gameDisplay.blit(value, [width/2, 0]) def our_snake(snake_block, snake_list): for x in snake_list: pygame.draw.rect(gameDisplay, green, [x[0], x[1], snake_block, snake_block]) def message(msg, color): mesg = font_style.render(msg, True, color) gameDisplay.blit(mesg, [width / 6, height / 3]) def gameLoop(): game_over = False game_close = False x1 = width / 2 y1 = height / 2 x1_change = 0 y1_change = 0 snake_List = [] Length_of_snake = 1 foodx = round(random.randrange(0, width - snake_block) / 10.0) * 10.0 foody = round(random.randrange(0, height - snake_block) / 10.0) * 10.0 while not game_over: while game_close == True: gameDisplay.fill(black) message("Game Over! Press P to Play Again and Press Q to Quit the game. ", green) gameScore(Length_of_snake - 1) pygame.display.update() for event in pygame.event.get(): if event.type == pygame.KEYDOWN: if event.key == pygame.K_q: game_over = True game_close = False if event.key == pygame.K_p: gameLoop() for event in pygame.event.get(): if event.type == pygame.QUIT: game_over = True if event.type == pygame.KEYDOWN: if event.key == pygame.K_LEFT: x1_change = -snake_block y1_change = 0 elif event.key == pygame.K_RIGHT: x1_change = snake_block y1_change = 0 elif event.key == pygame.K_UP: y1_change = -snake_block x1_change = 0 elif event.key == pygame.K_DOWN: y1_change = snake_block x1_change = 0 if x1 >= width or x1 < 0 or y1 >= height or y1 < 0: game_close = True x1 += x1_change y1 += y1_change gameDisplay.fill(black) pygame.draw.rect(gameDisplay, yellow, [foodx, foody, snake_block, snake_block]) snake_Head = [] snake_Head.append(x1) snake_Head.append(y1) snake_List.append(snake_Head) if len(snake_List) > Length_of_snake: del snake_List[0] for x in snake_List[:-1]: if x == snake_Head: game_close = True our_snake(snake_block, snake_List) gameScore(Length_of_snake - 1) pygame.display.update() if x1 == foodx and y1 == foody: foodx = round(random.randrange(0, width - snake_block) / 10.0) * 10.0 foody = round(random.randrange(0, height - snake_block) / 10.0) * 10.0 Length_of_snake += 1 clock.tick(snake_speed) pygame.quit() quit() main_menu()
27.369863
104
0.56006
0
0
0
0
0
0
0
0
157
0.039289
4b59cb1bcbcd0c6d58e12de2aac812a57e139151
918
py
Python
SoftMax_Regression.py
chunish/tfboy-is-on-the-way
7cd4c1f7c0c1dd94189377ee0751f2c232a1e98c
[ "Apache-2.0" ]
null
null
null
SoftMax_Regression.py
chunish/tfboy-is-on-the-way
7cd4c1f7c0c1dd94189377ee0751f2c232a1e98c
[ "Apache-2.0" ]
null
null
null
SoftMax_Regression.py
chunish/tfboy-is-on-the-way
7cd4c1f7c0c1dd94189377ee0751f2c232a1e98c
[ "Apache-2.0" ]
null
null
null
import tensorflow as tf from tensorflow.examples.tutorials.mnist import input_data mnist = input_data.read_data_sets("MNIST_data/", one_hot = True) sess = tf.InteractiveSession() x = tf.placeholder(tf.float32, [None, 784]) W = tf.Variable(tf.zeros([784, 10])) b = tf.Variable(tf.zeros([10])) y = tf.nn.softmax(tf.matmul(x, W) + b) y_ = tf.placeholder(tf.float32, [None, 10]) # 真实概率 cross_entropy = tf.reduce_mean(-tf.reduce_sum(y_ * tf.log(y), reduction_indices = [1])) train_step = tf.train.GradientDescentOptimizer(0.5).minimize(cross_entropy) tf.global_variables_initializer().run() for i in range(10000): batch_xs, batch_ys = mnist.train.next_batch(100) train_step.run({x: batch_xs, y_: batch_ys}) correct_prediction = tf.equal(tf.argmax(y, 1), tf.argmax(y_, 1)) accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32)) print(accuracy.eval({x: mnist.test.images, y_: mnist.test.labels}))
31.655172
87
0.734205
0
0
0
0
0
0
0
0
27
0.029158
4b5a4dc0a470a6b6a0219e69281685d307bd50e5
465
py
Python
ltc/analyzer/migrations/0006_graphitevariable_function.py
v0devil/jltom
b302a39a187b8e1154c6deda636a4db8b30bb40b
[ "MIT" ]
4
2016-12-30T13:26:59.000Z
2017-04-26T12:07:36.000Z
ltc/analyzer/migrations/0006_graphitevariable_function.py
v0devil/jltom
b302a39a187b8e1154c6deda636a4db8b30bb40b
[ "MIT" ]
null
null
null
ltc/analyzer/migrations/0006_graphitevariable_function.py
v0devil/jltom
b302a39a187b8e1154c6deda636a4db8b30bb40b
[ "MIT" ]
null
null
null
# Generated by Django 2.2.20 on 2021-05-27 14:53 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('analyzer', '0005_auto_20210526_1755'), ] operations = [ migrations.AddField( model_name='graphitevariable', name='function', field=models.CharField(choices=[('A', 'avg'), ('MA', 'max'), ('MI', 'min')], default='A', max_length=12), ), ]
24.473684
117
0.584946
371
0.797849
0
0
0
0
0
0
140
0.301075
4b5c74257ca507c7289c021413a4bdff6ed7d6a6
2,865
py
Python
Python/061.py
jaimeliew1/Project_Euler_Solutions
963c9c6d6571cade8f87341f97a6a2cd1af202bb
[ "MIT" ]
null
null
null
Python/061.py
jaimeliew1/Project_Euler_Solutions
963c9c6d6571cade8f87341f97a6a2cd1af202bb
[ "MIT" ]
1
2018-04-16T21:01:50.000Z
2018-04-16T21:01:50.000Z
Python/061.py
jaimeliew1/Project_Euler_Solutions
963c9c6d6571cade8f87341f97a6a2cd1af202bb
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- """ Solution to Project Euler problem X Author: Jaime Liew https://github.com/jaimeliew1/Project_Euler_Solutions """ import math def isTri(n): return ((math.sqrt(1+8*n)-1)/2).is_integer() def isSqr(n): return (math.sqrt(n)).is_integer() def isPent(n): return ((1+math.sqrt(1+24*n))/6).is_integer() def isHex(n): return ((1+math.sqrt(1+8*n))/4).is_integer() def isHept(n): return ((3+math.sqrt(9+40*n))/10).is_integer() def isOct(n): return ((2+math.sqrt(4+12*n))/6).is_integer() isPoly = [isTri, isSqr, isPent, isHex,isHept,isOct] class Jnum: id = 0 #each nth bit is 1 if it is an nGon number n = 0 isMultiPoly = False def __init__(self, num): self.n = num for i in (f(num) for f in isPoly): self.id = (self.id << 1) | i if bin(self.id).count('1') > 1: self.isMultiPoly = True def __eq__(self,other): return self.n == other.n def __ne__(self,other): return self.n != other.n def checkThisSet(thisSet,depth,maxDepth, numSet): for q in (q for q in numSet if q not in thisSet): workingBit = 0 qIsCandidate = True if str(thisSet[-1].n)[2:] == str(q.n)[:2]: #if cyclical workingBit = 0 for i in (thisSet + [q]): if workingBit & (i.id) == 0: workingBit |= (i.id) else: qIsCandidate = False break else: qIsCandidate = False if qIsCandidate: if depth == maxDepth-1: if str(thisSet[0].n)[:2] == str(q.n)[2:]: #if cyclical back to start return list(thisSet + [q]) else: return [Jnum(0)] furtherTesting = checkThisSet(list(thisSet + [q]),depth +1, maxDepth, numSet) if furtherTesting != [Jnum(0)]: return furtherTesting return [Jnum(0)] def run(): ### generate set of possible candidates numSet = [] for i in range(1000, 10000): a = Jnum(i) if a.id != 0: if a.isMultiPoly: temp = a for k, bit in enumerate(bin(a.id)[2:].zfill(6)[::-1]): if bit == '1': temp.id = 1<<k numSet.append(Jnum(a.n)) numSet[-1].id = 1<<k else: numSet.append(a) #print("there are ",len(numSet)," candidate numbers.\n") ### Recursive search loop for i in numSet: currentSet = checkThisSet(list([i]), 1, 6, numSet) if currentSet != [Jnum(0)]: break Sum = 0 for i in currentSet: #print(i.n, bin(i.id)[2:].zfill(6)) Sum += i.n return Sum if __name__ == "__main__": print(run())
25.131579
89
0.506806
433
0.151134
0
0
0
0
0
0
391
0.136475
4b5cbe04dec6e7c55b09522e71410c55307b8fa0
302
py
Python
shop/views/shop_views.py
cuescience/cuescience-shop
bf5ea159f9277d1d6ab7acfcad3f2517723a225c
[ "MIT" ]
null
null
null
shop/views/shop_views.py
cuescience/cuescience-shop
bf5ea159f9277d1d6ab7acfcad3f2517723a225c
[ "MIT" ]
null
null
null
shop/views/shop_views.py
cuescience/cuescience-shop
bf5ea159f9277d1d6ab7acfcad3f2517723a225c
[ "MIT" ]
null
null
null
from shop.models import Product from django.shortcuts import render_to_response from django.template import RequestContext def index_view(request): products = Product.objects.all() return render_to_response("cuescience_shop/index.html", RequestContext(request, {"products": products}))
20.133333
108
0.788079
0
0
0
0
0
0
0
0
38
0.125828
4b5d16684195ca44a761cc1ab6755c005952e4d5
163
py
Python
qrogue/test/statevector_tests.py
7Magic7Mike7/Qrogue
70bd5671a77981c1d4b633246321ba44f13c21ff
[ "MIT" ]
4
2021-12-14T19:13:43.000Z
2022-02-16T13:25:38.000Z
qrogue/test/statevector_tests.py
7Magic7Mike7/Qrogue
70bd5671a77981c1d4b633246321ba44f13c21ff
[ "MIT" ]
null
null
null
qrogue/test/statevector_tests.py
7Magic7Mike7/Qrogue
70bd5671a77981c1d4b633246321ba44f13c21ff
[ "MIT" ]
1
2022-01-04T18:35:51.000Z
2022-01-04T18:35:51.000Z
import numpy as np from qrogue.game.logic.actors import StateVector stv = StateVector([1 / np.sqrt(2), 0 + 0j, 0 + 0j, 1 / np.sqrt(2)]) #stv.extend(1) print(stv)
23.285714
67
0.680982
0
0
0
0
0
0
0
0
14
0.08589
4b5fd841b1005516ab298b5be16fb1dd41b071b3
3,190
py
Python
taehyoungram/images/views.py
TaeHyoungKwon/taehyoungram
055c9effdaa718d60e7627196754ea6b48dded20
[ "MIT" ]
null
null
null
taehyoungram/images/views.py
TaeHyoungKwon/taehyoungram
055c9effdaa718d60e7627196754ea6b48dded20
[ "MIT" ]
7
2020-02-12T01:23:48.000Z
2022-03-11T23:26:02.000Z
taehyoungram/images/views.py
TaeHyoungKwon/taehyoungram
055c9effdaa718d60e7627196754ea6b48dded20
[ "MIT" ]
null
null
null
from rest_framework.views import APIView from rest_framework.response import Response from rest_framework import status from .models import Image, Comment, Like from .serializers import ImageSerializer, CommentSerializer, LikeSerializer class Feed(APIView): def get(self, request, format=None): user = request.user follwoing_users = user.following.all() image_list = [] for following_user in follwoing_users: user_images = following_user.images.all()[:2] for image in user_images: image_list.append(image) sorted_list = sorted(image_list, key=lambda image:image.created_at, reverse=True) serializer = ImageSerializer(sorted_list, many=True) return Response(serializer.data) class LikeImage(APIView): def post(self, request, image_id, format=None): try: found_image = Image.objects.get(id=image_id) except Image.DoesNotExist : return Response(status=status.HTTP_404_NOT_FOUND) try: pre_exisiting_like = Like.objects.get( creator=request.user, image=found_image ) return Response(status=status.HTTP_304_NOT_MODIFIED) except Like.DoesNotExist: new_like = Like.objects.create( creator=request.user, image=found_image ) new_like.save() return Response(status=status.HTTP_201_CREATED) class UnLikeImage(APIView): def delete(self, request, image_id, format=None): user = request.user try: found_image = Image.objects.get(id=image_id) except: return Response(status=status.HTTP_404_NOT_FOUND) try: pre_existing_like = Like.objects.get( creator=user, image=found_image ) pre_existing_like.delete() return Response(status=status.HTTP_204_NO_CONTENT) except Like.DoesNotExist: return Response(status=status.HTTP_304_NOT_MODIFIED) class CommentOnImage(APIView): def post(self, request, image_id, format=None): user = request.user try: found_image = Image.objects.get(id=image_id) except Image.DoesNotExist: return Response(status=status.HTTP_404_NOT_FOUND) serializer = CommentSerializer(data=request.data) if serializer.is_valid(): serializer.save(creator=user, image=found_image) return Response(data=serializer.data, status=status.HTTP_201_CREATED) else: return Response(data=serializer.errors, status=status.HTTP_400_BAD_REQUEST) class Comment(APIView): def delete(self, request, comment_id, format=None): s user = request.user try: comment = Comment.objects.get(id=comment_id, creator=user) comment.delete() return Response(status=status.HTTP_204_NO_CONTENT) except Comment.DoesNotExist: return Response(status=status.HTTP_404_NOT_FOUND)
27.73913
89
0.626646
2,927
0.917555
0
0
0
0
0
0
0
0
4b6108a9e3aec6079aa6d738f5836676b16bd14f
4,195
py
Python
examples/zoomed.py
ColmTalbot/psd-covariance-matrices
59631dd1860e9cf28658df1ce28b10f6f8d59868
[ "MIT" ]
null
null
null
examples/zoomed.py
ColmTalbot/psd-covariance-matrices
59631dd1860e9cf28658df1ce28b10f6f8d59868
[ "MIT" ]
null
null
null
examples/zoomed.py
ColmTalbot/psd-covariance-matrices
59631dd1860e9cf28658df1ce28b10f6f8d59868
[ "MIT" ]
2
2021-07-01T02:00:10.000Z
2021-08-02T07:29:03.000Z
#!/usr/bin/env python """ Compute the comparison of the analytic and experimental PSD matrices. This will generate Figure 1. This is probably the only example that will run in a reasonable time without a GPU. For more details on the method see https://arxiv.org/abs/2106.13785. """ import numpy as np import matplotlib.pyplot as plt from bilby.core.utils import create_white_noise, create_frequency_series from scipy.signal.windows import tukey from scipy.interpolate import interp1d from tqdm.auto import trange from coarse_psd_matrix.utils import ( compute_psd_matrix, create_parser, fetch_psd_data, ) from coarse_psd_matrix.plotting import plot_psd_matrix from matplotlib import rcParams rcParams["font.family"] = "serif" rcParams["font.serif"] = "Computer Modern Roman" rcParams["font.size"] = 20 rcParams["text.usetex"] = True rcParams["grid.alpha"] = 0 if __name__ == "__main__": parser = create_parser() args = parser.parse_args() interferometer = args.interferometer outdir = args.outdir duration = args.duration medium_duration = args.medium_duration sampling_frequency = args.sampling_frequency low_frequency = args.low_frequency tukey_alpha = args.tukey_alpha minimum_frequency = 480 maximum_frequency = 530 event = args.event data = fetch_psd_data( interferometer_name=interferometer, event=event, duration=duration, sampling_frequency=sampling_frequency, low_frequency=low_frequency, tukey_alpha=tukey_alpha, medium_duration=medium_duration, outdir=outdir, ) svd = compute_psd_matrix( interferometer_name=interferometer, event=event, duration=duration, sampling_frequency=sampling_frequency, low_frequency=low_frequency, tukey_alpha=tukey_alpha, minimum_frequency=minimum_frequency, maximum_frequency=maximum_frequency, medium_duration=medium_duration, outdir=outdir, ) psd = data["medium_psd"][: sampling_frequency // 2 * medium_duration + 1] original_frequencies = create_frequency_series( duration=medium_duration, sampling_frequency=sampling_frequency ) new_frequencies = create_frequency_series( duration=256, sampling_frequency=sampling_frequency ) psd = interp1d(original_frequencies, psd)(new_frequencies) short_window = tukey(duration * sampling_frequency, tukey_alpha) short_window /= np.mean(short_window ** 2) ** 0.5 analytic_psd_matrix = (svd[0] * svd[1]) @ svd[2] estimated_psd_matrix = np.zeros_like(analytic_psd_matrix) nfft = duration * sampling_frequency start_idx = minimum_frequency * duration stop_idx = maximum_frequency * duration n_average = 1024 * 1024 // 64 for _ in trange(n_average): white_noise, frequencies = create_white_noise( sampling_frequency=2048, duration=256 ) coloured_noise = white_noise * psd ** 0.5 td_noise = np.fft.irfft(coloured_noise).reshape((-1, nfft)) fd_noise = np.fft.rfft(td_noise * short_window) reduced_noise = fd_noise[:, start_idx : stop_idx + 1] estimated_psd_matrix += np.einsum( "ki,kj->ij", reduced_noise, reduced_noise.conjugate() ) / 2 total_averages = n_average * len(reduced_noise) estimated_psd_matrix /= total_averages rcParams["font.family"] = "serif" rcParams["font.serif"] = "Computer Modern Roman" rcParams["font.size"] = 20 rcParams["text.usetex"] = True rcParams["grid.alpha"] = 0 fig, axes = plt.subplots(nrows=2, figsize=(10, 16)) kwargs = dict( minimum_frequency=minimum_frequency, maximum_frequency=maximum_frequency, duration=duration, vmin=-53, vmax=-41.8, tick_step=10, ) plot_psd_matrix(estimated_psd_matrix, axes[0], **kwargs) plot_psd_matrix(analytic_psd_matrix, axes[1], **kwargs) axes[0].text(-25, 190, "(a)") axes[1].text(-25, 190, "(b)") plt.tight_layout() plt.savefig(f"{outdir}/zoom_{tukey_alpha}.pdf") if tukey_alpha == 0.1: plt.savefig("figure_1.pdf") plt.close()
32.269231
77
0.695828
0
0
0
0
0
0
0
0
555
0.1323
4b624ab13f54c8cfd7032b48000491920f6d9a27
5,581
py
Python
web_spider/web_spider/pipelines.py
syun0216/simple_ci
83d31cb04357fe0bd428ab8f09c2db81a06eb723
[ "MIT" ]
null
null
null
web_spider/web_spider/pipelines.py
syun0216/simple_ci
83d31cb04357fe0bd428ab8f09c2db81a06eb723
[ "MIT" ]
null
null
null
web_spider/web_spider/pipelines.py
syun0216/simple_ci
83d31cb04357fe0bd428ab8f09c2db81a06eb723
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # Define your item pipelines here # # Don't forget to add your pipeline to the ITEM_PIPELINES setting # See: https://doc.scrapy.org/en/latest/topics/item-pipeline.html import pymysql class WebSpiderPipeline(object): def init_insert_db(self,key,table_name): pass def process_item(self, item, spider): # print(item['name']) connection = pymysql.connect(host='127.0.0.1', user='root', password='123456', db='mydb', charset='utf8mb4', cursorclass=pymysql.cursors.DictCursor) try: with connection.cursor() as cursor: # Create a new record if item['type'] == 'toutiao': insert_sql = """INSERT INTO `dongqiudi` (`id`, `name`,`url`,`time`,`comment`,`image`) VALUES (%s, %s,%s,%s,%s,%s) ON DUPLICATE KEY UPDATE id=VALUES(id), name=VALUES(name), url=VALUES (url), time=VALUES (time), comment=VALUES (comment), image=VALUES (image)""" cursor.execute(insert_sql, (item['id'], item['name'], item['url'], item['time'], item['comment'], item['image'])) elif item['type'] == 'rank': insert_sql = """INSERT INTO `rank` (`rank`,`team_avatar`,`team_name`,`round`,`win`,`draw`,`lost`,`goal`,`fumble`,`GD`,`integral`,`rel`,`rel_name`) VALUES (%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s) ON DUPLICATE KEY UPDATE rank=VALUES (rank), team_avatar=VALUES (team_avatar), team_name=VALUES (team_name), round=VALUES (round), win=VALUES (win), draw=VALUES (draw), lost=VALUES (lost), goal=VALUES (goal), fumble=VALUES (fumble), GD=VALUES (GD), integral=VALUES (integral), rel=VALUES (rel), rel_name=VALUES (rel_name) """ cursor.execute(insert_sql, (item['rank'], item['team_avatar'], item['team_name'], item['round'], item['win'], item['draw'],item['lost'],item['goal'],item['fumble'],item['GD'],item['integral'],item['rel'],item['rel_name'])) elif item['type'] == 'goal': insert_sql = """INSERT INTO `player_goal_rank` (`rank`,`data`,`player_avatar`,`player_name`,`team_avatar`,`team_name`,`rel`,`rel_name`) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) ON DUPLICATE KEY UPDATE rank=VALUES (rank), data=VALUES (data), player_avatar=VALUES (player_avatar), player_name=VALUES (player_name), team_avatar=VALUES (team_avatar), team_name=VALUES (team_name), rel=VALUES (rel), rel_name=VALUES (rel_name) """ cursor.execute(insert_sql, (item['rank'], item['data'], item['player_avatar'], item['player_name'],item['team_avatar'], item['team_name'], item['rel'], item['rel_name'])) elif item['type'] == 'assist': insert_sql = """INSERT INTO `player_assist_rank` (`rank`,`data`,`player_avatar`,`player_name`,`team_avatar`,`team_name`,`rel`,`rel_name`) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) ON DUPLICATE KEY UPDATE rank=VALUES (rank), data=VALUES (data), player_avatar=VALUES (player_avatar), player_name=VALUES (player_name), team_avatar=VALUES (team_avatar), team_name=VALUES (team_name), rel=VALUES (rel), rel_name=VALUES (rel_name) """ cursor.execute(insert_sql, (item['rank'], item['data'], item['player_avatar'], item['player_name'],item['team_avatar'], item['team_name'], item['rel'], item['rel_name'])) # connection is not autocommit by default. So you must commit to save # your changes. connection.commit() finally: connection.close() pass
52.650943
226
0.392403
5,364
0.961118
0
0
0
0
0
0
4,061
0.727647
4b63b53c7d4c16192575ee81d07d9b767d76c245
239
py
Python
etl_framework/config_mixins/AddDestinationMixin.py
pantheon-ci-bot/etl-framework
36d4c0d5c26ddd7c0bb2d2b99e3138b50a21c46f
[ "MIT" ]
2
2017-03-01T20:09:06.000Z
2019-02-08T17:10:16.000Z
etl_framework/config_mixins/AddDestinationMixin.py
pantheon-ci-bot/etl-framework
36d4c0d5c26ddd7c0bb2d2b99e3138b50a21c46f
[ "MIT" ]
40
2015-10-10T15:02:21.000Z
2020-03-17T22:32:04.000Z
etl_framework/config_mixins/AddDestinationMixin.py
pantheon-ci-bot/etl-framework
36d4c0d5c26ddd7c0bb2d2b99e3138b50a21c46f
[ "MIT" ]
2
2018-11-14T21:50:58.000Z
2022-03-07T20:59:27.000Z
"""mixin to add destinations to config attributes""" class AddDestinationMixin(object): """stuff""" def add_destination_chooser(self, destination_chooser_mappings): """adds destination_choosers to config""" pass
23.9
68
0.707113
184
0.769874
0
0
0
0
0
0
104
0.435146
4b65eb4040ecf11e53140f9d3ec6fb5084fff907
6,298
py
Python
src/utilities/download_bc.py
geoschem/integrated_methane_inversion
0615e3b76c111beadaf0d7fb5b9fa99aa782f403
[ "MIT" ]
null
null
null
src/utilities/download_bc.py
geoschem/integrated_methane_inversion
0615e3b76c111beadaf0d7fb5b9fa99aa782f403
[ "MIT" ]
3
2022-02-14T20:42:35.000Z
2022-03-29T18:11:40.000Z
src/utilities/download_bc.py
geoschem/integrated_methane_inversion
0615e3b76c111beadaf0d7fb5b9fa99aa782f403
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 """ Description: ------------ This Python script (assumes Python3) downloads boundary conditions files from AWS S3 to a target directory for the requested date range. Remarks: -------- (1) Jiawei Zhuang found that it is much faster to issue aws s3 cp commands from a bash script than a Python script. Therefore, in this routine we create a bash script with all of the download commands that will be executed by the main routine. """ # Imports import os import sys import subprocess # Exit with error if we are not using Python3 assert sys.version_info.major >= 3, "ERROR: Python 3 is required to run download_bc.py!" # Define global variables DATA_DOWNLOAD_SCRIPT = "./auto_generated_download_script.sh" def list_missing_files(start_date, end_date, destination): """ Creates list of BC files in date range that do not already exist at destination. Args: ----- start_date : str Initial date of simulation. end_date : str Final date of simulation. destination : str Target directory for downloaded files """ missing_files = [] start_str = str(start_date) start_year = start_str[:4] start_month = start_str[4:6] start_day = start_str[6:8] end_str = str(end_date) end_year = end_str[:4] end_month = end_str[4:6] end_day = end_str[6:8] month_days = [31, [28, 29], 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] file_prefix = "GEOSChem.BoundaryConditions." file_suffix = "_0000z.nc4" for year in range(int(start_year), int(end_year) + 1): # skip years with definite no data if year < 2018: print( "Skipping BC data download for ", str(year), ": no data from this year" ) continue init_month = 1 final_month = 12 if year == int(start_year): # only get desired months from incomplete years init_month = int(start_month) if year == int(end_year): final_month = int(end_month) for month in range(init_month, final_month + 1): # skip months with definite no data if year == 2018 and month < 4: print( "Skipping BC data download for ", str(year), "/0", str(month), ": no data from this month", ) continue # add 0 to month string if necessary month_prefix = "0" if month < 10 else "" init_day = 1 final_day = month_days[month - 1] # leap day if month == 2: if year % 4 == 0: final_day = final_day[1] else: final_day = final_day[0] if month == int(start_month) and year == int(start_year): # only get desired days from incomplete months init_day = int(start_day) if month == int(end_month) and year == int(end_year): final_day = int(end_day) for day in range(init_day, final_day + 1): # add 0 to day string if necessary day_prefix = "0" if day < 10 else "" # check if file for this day already exists file_name = ( file_prefix + str(year) + month_prefix + str(month) + day_prefix + str(day) + file_suffix ) # add file to download list if needed if not os.path.exists(destination + "/" + file_name): missing_files.append(file_name) return missing_files def create_download_script(paths, destination): """ Creates a data download script to obtain missing files Args: ----- paths : dict Output of function list_missing_files. """ # Create the data download script with open(DATA_DOWNLOAD_SCRIPT, "w") as f: # Write shebang line to script print("#!/bin/bash\n", file=f) print("# This script was generated by download_bc.py\n", file=f) cmd_prefix = "aws s3 cp --only-show-errors --request-payer=requester " remote_root = "s3://imi-boundary-conditions/" # make destination if needed if not os.path.exists(destination): os.mkdir(destination) # Write download commands for only the missing data files for path in paths: cmd = cmd_prefix + remote_root + path + " " + destination print(cmd, file=f) print(file=f) # Close file and make it executable f.close() os.chmod(DATA_DOWNLOAD_SCRIPT, 0o755) def download_the_data(start_date, end_date, destination): """ Downloads required boundary conditions files from AWS. Args: ----- start_date : str Initial date of simulation. end_date : str Final date of simulation. destination : str Target directory for downloaded files """ # Get a list of missing data paths paths = list_missing_files(start_date, end_date, destination) # Create script to download missing files from AWS S3 create_download_script(paths, destination) # Run the data download script and return the status # Remove the file afterwards status = subprocess.call(DATA_DOWNLOAD_SCRIPT) os.remove(DATA_DOWNLOAD_SCRIPT) # Raise an exception if the data was not successfully downloaded if status != 0: err_msg = "Error downloading data from AWS!" raise Exception(err_msg) def main(): """ Main program. Gets command-line arguments and calls function download_the_data to initiate a data-downloading process. Calling sequence: ----------------- ./download_data.py start_date end_date destination Example call: ------------- ./download_data.py 20200501 20200531 /home/ubuntu/ExtData/BoundaryConditions """ download_the_data(sys.argv[1], sys.argv[2], sys.argv[3]) if __name__ == "__main__": main()
30.872549
88
0.580343
0
0
0
0
0
0
0
0
2,909
0.461893
4b675eebc011dcbf30a99943e8204d368b2ed0b9
216
py
Python
backend/core/__init__.py
itimor/one-ops
f1111735de252012752dfabe11598e9690c89257
[ "MIT" ]
2
2020-09-25T05:52:55.000Z
2021-01-14T07:06:17.000Z
backend/core/__init__.py
itimor/one-ops
f1111735de252012752dfabe11598e9690c89257
[ "MIT" ]
6
2021-03-19T10:20:05.000Z
2021-09-22T19:30:21.000Z
backend/core/__init__.py
itimor/one-ops
f1111735de252012752dfabe11598e9690c89257
[ "MIT" ]
1
2022-02-24T01:37:06.000Z
2022-02-24T01:37:06.000Z
from __future__ import absolute_import, unicode_literals # This will make sure the app is always imported when # Django starts so that shared_task will use this app. # import pymysql # # pymysql.install_as_MySQLdb()
30.857143
56
0.805556
0
0
0
0
0
0
0
0
154
0.712963
4b6ab5ac671d440fac187ec117fb8831906a3c8f
39
tac
Python
semantics_and_generate/samples/link1.tac
AHEADer/my_decaf_compiler
42ba9f140c5fda3cd2b4fdb727745d2cfd39c923
[ "MIT" ]
1
2018-01-03T03:35:38.000Z
2018-01-03T03:35:38.000Z
semantics_and_generate/samples/link1.tac
AHEADer/my_decaf_compiler
42ba9f140c5fda3cd2b4fdb727745d2cfd39c923
[ "MIT" ]
null
null
null
semantics_and_generate/samples/link1.tac
AHEADer/my_decaf_compiler
42ba9f140c5fda3cd2b4fdb727745d2cfd39c923
[ "MIT" ]
null
null
null
____notmain: BeginFunc 0 ; EndFunc ;
9.75
14
0.717949
0
0
0
0
0
0
0
0
0
0
4b6add180192d528e3ed133e29c757a81886beb8
483
py
Python
NoteBooks/Curso de Python/Python/Paradigmas/Object Oriented Programming/Modelando Objetos_2.py
Alejandro-sin/Learning_Notebooks
161d6bed4c7b1d171b45f61c0cc6fa91e9894aad
[ "MIT" ]
1
2021-02-26T13:12:22.000Z
2021-02-26T13:12:22.000Z
NoteBooks/Curso de Python/Python/Paradigmas/Object Oriented Programming/Modelando Objetos_2.py
Alejandro-sin/Learning_Notebooks
161d6bed4c7b1d171b45f61c0cc6fa91e9894aad
[ "MIT" ]
null
null
null
NoteBooks/Curso de Python/Python/Paradigmas/Object Oriented Programming/Modelando Objetos_2.py
Alejandro-sin/Learning_Notebooks
161d6bed4c7b1d171b45f61c0cc6fa91e9894aad
[ "MIT" ]
null
null
null
""" Ejercicio para operación entre currencies """ """ Representación del currency""" class Curency: def __init__(self, name, symbol, factor): self.name = name self.symbol = symbol self.factor = factor # No me queda muy claro el uso de esta función, sirve para mostrar puntualmente qué? # def __repr__(self): # info = self.name # info2 = self.symbol # return info, info2 euro = Curency("Euro","EU","3.2") print(euro)
17.888889
84
0.621118
143
0.293634
0
0
0
0
0
0
289
0.593429
4b6b2b2466b7f50264d915b0b9ab9925c879719e
587
py
Python
kora/install/blender.py
wannaphong/kora
8a9034097d07b14094e077769c02a0b4857d179b
[ "MIT" ]
91
2020-05-26T05:54:51.000Z
2022-03-09T07:33:44.000Z
kora/install/blender.py
wannaphong/kora
8a9034097d07b14094e077769c02a0b4857d179b
[ "MIT" ]
12
2020-10-03T10:09:11.000Z
2021-03-06T23:12:21.000Z
kora/install/blender.py
wannaphong/kora
8a9034097d07b14094e077769c02a0b4857d179b
[ "MIT" ]
16
2020-07-07T18:39:29.000Z
2021-03-06T03:46:49.000Z
import os from IPython import get_ipython # need this fix first os.environ["LD_PRELOAD"] = "" os.system("apt remove libtcmalloc-minimal4") os.system("apt install libtcmalloc-minimal4") os.environ["LD_PRELOAD"] = "/usr/lib/x86_64-linux-gnu/libtcmalloc_minimal.so.4.3.0" os.system("dpkg -L libtcmalloc-minimal4") # then install blender url = "https://download.blender.org/release/Blender2.83/blender-2.83.0-linux64.tar.xz" os.system(f"curl {url} | tar xJ") os.system("ln -s /content/blender-2.83.0-linux64/blender /usr/local/bin/blender") # show result get_ipython().system("blender -v")
36.6875
86
0.749574
0
0
0
0
0
0
0
0
419
0.713799
4b6cb29835a30f52c4bb14e8a53ca4a8d4a5cdb7
2,352
py
Python
parks/models.py
ExpertOfNone/expert_of_none
9ff4e4279a570712766546122c014c754f753485
[ "MIT" ]
null
null
null
parks/models.py
ExpertOfNone/expert_of_none
9ff4e4279a570712766546122c014c754f753485
[ "MIT" ]
null
null
null
parks/models.py
ExpertOfNone/expert_of_none
9ff4e4279a570712766546122c014c754f753485
[ "MIT" ]
null
null
null
from django.db import models from django_countries.fields import CountryField from localflavor.us.models import USStateField from base.models import EONBaseModel class Amenity(EONBaseModel): """Common amenities, i.e. Pool, Lake, Hiking Trail, Primitive Camping, etc.""" name = models.CharField(max_length=50) code = models.CharField(max_length=5) description = models.TextField(null=True, blank=True) def __str__(self): return self.name class Park(EONBaseModel): """General Park Information Utilized for Reference""" PARK_TYPE_STATE = 'state' PARK_TYPE_NATIONAL = 'national' PARK_TYPE_CITY = 'city' PARK_TYPE_OTHER = 'other' PARK_TYPE_CHOICES = ( (PARK_TYPE_STATE, 'State'), (PARK_TYPE_NATIONAL, 'National'), (PARK_TYPE_CITY, 'City'), (PARK_TYPE_OTHER, 'Other') ) park_type = models.CharField(max_length=20, choices=PARK_TYPE_CHOICES) name = models.CharField(max_length=100) description = models.TextField(null=True, blank=True) photos = models.ManyToManyField('base.Photo', through='ParkPhoto') address_one = models.CharField(max_length=50) address_two = models.CharField(max_length=50, null=True, blank=True) city = models.CharField(max_length=50) state = USStateField(blank=True, null=True) country = CountryField() postal_code = models.CharField(blank=True, null=True, max_length=20) amenities = models.ManyToManyField(Amenity, through='ParkAmenity') topic = models.ForeignKey('base.Topic', on_delete=models.SET_NULL, null=True) def __str__(self): return '{name} - {park_type}'.format( name=self.name, park_type=self.get_park_type_display(), ) class ParkAmenity(EONBaseModel): park = models.ForeignKey(Park, on_delete=models.CASCADE) amenity = models.ForeignKey(Amenity, on_delete=models.CASCADE) additional_info = models.TextField(blank=True) class ParkPhoto(EONBaseModel): photo = models.ForeignKey('base.Photo', on_delete=models.CASCADE, related_name='park_photos') park = models.ForeignKey(Park, on_delete=models.DO_NOTHING, related_name='park_photos') def __str__(self): return '{park_name} - Photo:{photo_name}'.format( park_name=self.park.name, photo_name=self.photo.name, )
31.36
97
0.702381
2,177
0.925595
0
0
0
0
0
0
333
0.141582
4b6d6ef8de7836397b32a70984d9c9488bd0f64f
3,891
py
Python
roomfinder_dispo/roomfinder_dispo/dispo.py
GuillaumeMorini/roomfinder
d756bba6e50a7361ecf9cf529af4a1775a0e836b
[ "Apache-2.0" ]
14
2017-01-23T02:58:53.000Z
2020-12-21T14:05:07.000Z
roomfinder_dispo/roomfinder_dispo/dispo.py
GuillaumeMorini/roomfinder
d756bba6e50a7361ecf9cf529af4a1775a0e836b
[ "Apache-2.0" ]
2
2017-01-23T09:46:54.000Z
2017-09-11T10:15:07.000Z
roomfinder_dispo/roomfinder_dispo/dispo.py
GuillaumeMorini/roomfinder
d756bba6e50a7361ecf9cf529af4a1775a0e836b
[ "Apache-2.0" ]
9
2017-01-23T02:55:27.000Z
2020-05-20T18:38:18.000Z
#!/usr/bin/env python2.7 import sys reload(sys) sys.setdefaultencoding("utf-8") from flask import Flask, render_template, request, jsonify import argparse import datetime import os, sys import requests from socket import error as SocketError import errno import json import pika import uuid app = Flask(__name__) @app.route("/book", methods=["GET"]) def book(): starttime=request.args.get('starttime', '') endtime=request.args.get('endtime', '') user_name=request.args.get('user_name', '') user_email=request.args.get('user_email', '') room_name=request.args.get('room_name', '') if starttime is None or endtime is None or user_name is None or user_email is None or room_name is None: return "no parameter provided to book request\n" data = { "cmd": "book", "data": {"starttime": starttime, "endtime": endtime, "user_name": user_name, "user_email": user_email, "room_name": room_name} } message = json.dumps(data) return send_message_to_queue(message) @app.route("/dispo", methods=["GET"]) def dispo(): key=request.args.get('key', '') sys.stderr.write( "key: "+str(key)+'\r\n') if key is not None and str(key) is not "": data = { "cmd": "dispo", "data": {"key": key} } message = json.dumps(data) return send_message_to_queue(message) return "no parameter provided to dispo request\n" def on_response(ch, method, props, body): global corr_id global response if corr_id == props.correlation_id: response = body def send_message_to_queue(message): global corr_id global response global connection global channel global callback_queue response=None connection = pika.BlockingConnection(pika.ConnectionParameters(host=rabbitmq,port=int(rabbitmq_port),heartbeat_interval=30)) channel = connection.channel() result=channel.queue_declare(exclusive=True) callback_queue = result.method.queue channel.basic_consume(on_response, no_ack=True, queue=callback_queue) corr_id=str(uuid.uuid4()) response = None corr_id = str(uuid.uuid4()) channel.basic_publish( exchange='', routing_key="rpc_queue", properties=pika.BasicProperties( reply_to = callback_queue, correlation_id = corr_id), body=message) print(" [x] Sent data to RabbitMQ") while response is None: connection.process_data_events() print(" [x] Get response from RabbitMQ") print "response: "+str(response) return response if __name__ == '__main__': from argparse import ArgumentParser parser = ArgumentParser("Room Finder Dispo Service") parser.add_argument("-r","--rabbitmq", help="IP or hostname for rabbitmq server, e.g. 'rabbit.domain.com'.") parser.add_argument("-p","--port", help="tcp port for rabitmq server, e.g. '2765'.") args = parser.parse_args() rabbitmq = args.rabbitmq if (rabbitmq == None): rabbitmq = os.getenv("roomfinder_rabbitmq_server") if (rabbitmq == None): get_rabbitmq_server = raw_input("What is the rabbitmq server IP or hostname? ") rabbitmq = get_rabbitmq_server rabbitmq_port = args.port if (rabbitmq_port == None): rabbitmq_port = os.getenv("roomfinder_rabbitmq_port") if (rabbitmq_port == None): get_rabbitmq_port = raw_input("What is the rabbitmq TCP port? ") rabbitmq_port = get_rabbitmq_port try: app.run(host='0.0.0.0', port=int("5000")) except: try: app.run(host='0.0.0.0', port=int("5000")) except: print "Dispo web services error"
31.893443
138
0.627602
0
0
0
0
1,148
0.29504
0
0
765
0.196608
4b6e6530e3333cd913c07220255b812f35a812cc
167
py
Python
Sea/model/components/Component3D.py
FRidh/Sea
b474e93a449570a9ba3b915c4d80f814feee2545
[ "BSD-3-Clause" ]
2
2015-07-02T13:34:09.000Z
2015-09-28T09:07:52.000Z
Sea/model/components/Component3D.py
FRidh/Sea
b474e93a449570a9ba3b915c4d80f814feee2545
[ "BSD-3-Clause" ]
null
null
null
Sea/model/components/Component3D.py
FRidh/Sea
b474e93a449570a9ba3b915c4d80f814feee2545
[ "BSD-3-Clause" ]
1
2022-01-22T03:01:54.000Z
2022-01-22T03:01:54.000Z
import numpy as np from ComponentStructural import ComponentStructural from ..subsystems import SubsystemStructural class Component3D(ComponentStructural): pass
20.875
51
0.844311
48
0.287425
0
0
0
0
0
0
0
0
4b6f1aec2b3a7aa82fa7792516bb55e9223b7c08
1,063
py
Python
bot.py
federicosapienza/InboxNotionTelegramBot
031d5e78cd352dfb692b93f3e0b421695f1dc18e
[ "MIT" ]
null
null
null
bot.py
federicosapienza/InboxNotionTelegramBot
031d5e78cd352dfb692b93f3e0b421695f1dc18e
[ "MIT" ]
null
null
null
bot.py
federicosapienza/InboxNotionTelegramBot
031d5e78cd352dfb692b93f3e0b421695f1dc18e
[ "MIT" ]
null
null
null
from telegram.ext import Updater, CommandHandler, MessageHandler, Filters, ConversationHandler import logging from utils import TELEGRAM_TOKEN from handlers import start, ask_new_url, get_url, get_description, cancel from handlers import URL_URL, URL_DESCRIPTION logging.basicConfig(format='%(levelname)s - %(message)s', level=logging.DEBUG) logger = logging.getLogger(__name__) updater = None def start_bot(): global updater updater = Updater(TELEGRAM_TOKEN, use_context=True) dispatcher = updater.dispatcher dispatcher.add_handler(CommandHandler('start', start)) conversation_url_handler = ConversationHandler( entry_points=[CommandHandler('url', ask_new_url)], states={ URL_URL: [MessageHandler(Filters.text, get_url)], URL_DESCRIPTION: [MessageHandler(Filters.text, get_description)], }, fallbacks=[MessageHandler(Filters.command, cancel)] ) dispatcher.add_handler(conversation_url_handler) updater.start_polling(timeout=30) updater.idle() start_bot()
25.926829
94
0.739417
0
0
0
0
0
0
0
0
41
0.03857
4b6fc4e98137fcd105847298b470d6ad64f71618
841
py
Python
examples/face.py
birkenfeld/python-gr
1d6cd36616a73c8e569b8348869e6e30f3830ec4
[ "RSA-MD" ]
null
null
null
examples/face.py
birkenfeld/python-gr
1d6cd36616a73c8e569b8348869e6e30f3830ec4
[ "RSA-MD" ]
null
null
null
examples/face.py
birkenfeld/python-gr
1d6cd36616a73c8e569b8348869e6e30f3830ec4
[ "RSA-MD" ]
null
null
null
#!/usr/bin/env python """ Simple surface plot example """ from gr import * from math import * x = [-2 + i * 0.5 for i in range(0, 29)] y = [-7 + i * 0.5 for i in range(0, 29)] z = list(range(0, 841)) for i in range(0, 29): for j in range(0, 29): r1 = sqrt((x[j] - 5)**2 + y[i]**2) r2 = sqrt((x[j] + 5)**2 + y[i]**2) z[i * 29 - 1 + j] = (exp(cos(r1)) + exp(cos(r2)) - 0.9) * 25 setcharheight(24.0/500) settextalign(TEXT_HALIGN_CENTER, TEXT_VALIGN_TOP) textext(0.5, 0.9, "Surface Example") (tbx, tby) = inqtextext(0.5, 0.9, "Surface Example") fillarea(tbx, tby) setwindow(-2, 12, -7, 7) setspace(-80, 200, 45, 70) setcharheight(14.0/500) axes3d(1, 0, 20, -2, -7, -80, 2, 0, 2, -0.01) axes3d(0, 1, 0, 12, -7, -80, 0, 2, 0, 0.01) titles3d("X-Axis", "Y-Axis", "Z-Axis") surface(x, y, z, 3) surface(x, y, z, 1) updatews()
22.72973
64
0.567182
0
0
0
0
0
0
0
0
114
0.135553
4b70e648e25de3717c9f7effa2fbe1723459da27
344
py
Python
tests/system/examples/dask/assets/dask_function.py
Hedingber/mlrun
e2269718fcc7caa7e1aa379ac28495830b45f9da
[ "Apache-2.0" ]
1
2021-02-17T08:12:33.000Z
2021-02-17T08:12:33.000Z
tests/system/examples/dask/assets/dask_function.py
Hedingber/mlrun
e2269718fcc7caa7e1aa379ac28495830b45f9da
[ "Apache-2.0" ]
1
2020-12-31T14:36:29.000Z
2020-12-31T14:36:29.000Z
tests/system/examples/dask/assets/dask_function.py
Hedingber/mlrun
e2269718fcc7caa7e1aa379ac28495830b45f9da
[ "Apache-2.0" ]
1
2021-08-30T21:43:38.000Z
2021-08-30T21:43:38.000Z
# function that will be distributed def inc(x): return x + 2 # wrapper function, uses the dask client object def main(context, x=1, y=2): context.logger.info(f"params: x={x},y={y}") print(f"params: x={x},y={y}") x = context.dask_client.submit(inc, x) print(x) print(x.result()) context.log_result("y", x.result())
24.571429
47
0.627907
0
0
0
0
0
0
0
0
129
0.375
4b714a892a0b336d54d129baf723bfd26bcf8c4a
1,495
py
Python
app/core.py
antmicro/raw-image-data-previewer
1fc14848a27ce628047cf3e473a9f30f83c9892d
[ "Apache-2.0" ]
5
2021-06-08T15:37:23.000Z
2021-06-10T15:41:21.000Z
app/core.py
antmicro/raw-image-data-previewer
1fc14848a27ce628047cf3e473a9f30f83c9892d
[ "Apache-2.0" ]
37
2021-03-12T12:48:56.000Z
2021-12-09T11:41:05.000Z
app/core.py
antmicro/raw-image-data-previewer
1fc14848a27ce628047cf3e473a9f30f83c9892d
[ "Apache-2.0" ]
9
2021-03-22T14:03:37.000Z
2021-12-31T07:22:04.000Z
"""Main functionalities.""" from .image.image import (Image, RawDataContainer) from .image.color_format import AVAILABLE_FORMATS from .parser.factory import ParserFactory import cv2 as cv import os def load_image(file_path, color_format, width): try: image = Image.from_file(file_path) parser = ParserFactory.create_object( determine_color_format(color_format)) except Exception as e: print(type(e).__name__, e) image = parser.parse(image.data_buffer, determine_color_format(color_format), width) return image def get_displayable(image): if image.color_format is None: raise Exception("Image should be already parsed!") parser = ParserFactory.create_object(image.color_format) return parser.get_displayable(image) def determine_color_format(format_string): if format_string in AVAILABLE_FORMATS.keys(): return AVAILABLE_FORMATS[format_string] else: raise NotImplementedError( "Provided string is not name of supported format.") def save_image_as_file(image, file_path): directory = file_path.replace('\\', '/') if directory.rfind('/') == -1: directory = './' else: directory = directory[:directory.rfind("/")] if not os.path.isdir(directory): os.makedirs(directory) try: cv.imwrite(file_path, cv.cvtColor(image, cv.COLOR_RGB2BGR)) except Exception as e: print(type(e).__name__, e)
26.22807
69
0.681605
0
0
0
0
0
0
0
0
127
0.08495
4b73785d9dd8a4aaaf6a1aac49dbeb16165c0050
1,423
py
Python
demo/demo/urls.py
AlanCoding/Example-Django-App
1cca52b720d1b117ccce780618d9af94f584ac2c
[ "MIT" ]
null
null
null
demo/demo/urls.py
AlanCoding/Example-Django-App
1cca52b720d1b117ccce780618d9af94f584ac2c
[ "MIT" ]
null
null
null
demo/demo/urls.py
AlanCoding/Example-Django-App
1cca52b720d1b117ccce780618d9af94f584ac2c
[ "MIT" ]
null
null
null
"""demo URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.10/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.conf import settings from django.contrib import admin from django.views.generic import TemplateView import debug_toolbar from rockband import rocking_urls # from movies import urls as movie_urls from async import async_urls urlpatterns = [ url(r'^$', TemplateView.as_view(template_name="index.html")), url(r'^admin/', admin.site.urls), # Rock band urls url(r'^bands/', include(rocking_urls)), # asynchronous demo app url(r'^async/', include(async_urls)), # url(r'$movies/', include(movie_urls)) # Django auth views url('^', include('django.contrib.auth.urls')), ] # For the debug toolbar if settings.DEBUG: urlpatterns += [ url(r'^__debug__/', include(debug_toolbar.urls)), ]
33.093023
79
0.702741
0
0
0
0
0
0
0
0
886
0.622628
4b7544498643883f50311519a373ed59f4faa469
3,478
py
Python
app/urls.py
etihadprime/etihadwebclass
3b46d9068afeb0806198ef08fe26849ab9a09bb9
[ "Apache-2.0" ]
null
null
null
app/urls.py
etihadprime/etihadwebclass
3b46d9068afeb0806198ef08fe26849ab9a09bb9
[ "Apache-2.0" ]
6
2021-03-19T03:55:20.000Z
2021-09-22T19:06:06.000Z
app/urls.py
etihadprime/etihadwebclass
3b46d9068afeb0806198ef08fe26849ab9a09bb9
[ "Apache-2.0" ]
null
null
null
from django.urls import path from .views import teacherregister,studentregister,login_view,logout from . import views from .views import ( ClassroomCreateView,ClassroomListView,ClassroomDetailView,ClassroomUpdateView,ClassroomDeleteView, SubjectCreateView,SubjectListView,SubjectDetailView,SubjectUpdateView,SubjectDeleteView, ClassMemberCreateView,ClassMemberListView,ClassMemberDetailView,ClassMemberUpdateView,ClassMemberDeleteView, TimetableCreateView,TimetableListView,TimetableDetailView,TimetableUpdateView,TimetableDeleteView,CrudView,chatroom ) urlpatterns = [ path('', views.index, name='index'), path('health', views.health, name='health'), path('404', views.handler404, name='404'), path('500', views.handler500, name='500'), path('signup/teacher', teacherregister,name='register-teacher'), path('signup/student', studentregister,name='register-student'), path('accounts/login/', login_view, name='login'), path('logout/', logout,name='logout'), #Classroom path('classroom/new', ClassroomCreateView.as_view(),name='classroom-create'), path('classroom_list', ClassroomListView.as_view(),name='classroom-list'), path('classroom/<str:pk>/', ClassroomDetailView.as_view(),name='classroom-detail'), path('classroom/<str:pk>/update', ClassroomUpdateView.as_view(),name='classroom-update'), path('classroom/<str:pk>/delete', ClassroomDeleteView.as_view(),name='classroom-delete'), #path('Classroom/<int:pk>/image',ChildImageUpdateView.as_view(),name='Classroom-image'), #Subject path('subject/new', SubjectCreateView.as_view(),name='subject-create'), path('subject_list', SubjectListView.as_view(),name='subject-list'), path('subject/<int:pk>/', SubjectDetailView.as_view(),name='subject-detail'), path('subject/<int:pk>/update', SubjectUpdateView.as_view(),name='subject-update'), path('subject/<int:pk>/delete', SubjectDeleteView.as_view(),name='subject-delete'), # Class Members path('classmember/new', ClassMemberCreateView.as_view(),name='classmember-create'), path('classmember_list', ClassMemberListView.as_view(),name='classmember-list'), path('classmember/<str:pk>/', ClassMemberDetailView.as_view(),name='classmember-detail'), path('classmember/<str:pk>/update', ClassMemberUpdateView.as_view(),name='classmember-update'), path('classmember/<str:pk>/delete', ClassMemberDeleteView.as_view(),name='classmember-delete'), # TimeTable path('timetable/new', TimetableCreateView.as_view(),name='timetable-create'), path('timetable_list', TimetableListView.as_view(),name='timetable-list'), path('timetable/<int:pk>/', TimetableDetailView.as_view(),name='timetable-detail'), path('timetable/<int:pk>/update', TimetableUpdateView.as_view(),name='timetable-update'), path('timetable/<int:pk>/delete', TimetableDeleteView.as_view(),name='timetable-delete'), # chatroom path('chat/new',chatroom,name='chatroom'), path('crud/',CrudView.as_view(), name='crud_ajax'), ]
70.979592
115
0.648074
0
0
0
0
0
0
0
0
1,110
0.319149
4b76dabace6084b6df07b8d27c9db12c437ca835
44,634
py
Python
qaboard/qa.py
Samsung/qaboard
a2290f33da2bbd87cacf95822e1c85376083dfa1
[ "Apache-2.0" ]
51
2019-12-02T07:25:58.000Z
2022-03-23T13:27:11.000Z
qaboard/qa.py
Samsung/qaboard
a2290f33da2bbd87cacf95822e1c85376083dfa1
[ "Apache-2.0" ]
25
2020-01-20T16:13:49.000Z
2022-02-19T17:07:38.000Z
qaboard/qa.py
Samsung/qaboard
a2290f33da2bbd87cacf95822e1c85376083dfa1
[ "Apache-2.0" ]
15
2020-01-17T21:21:17.000Z
2022-02-23T10:13:48.000Z
#!/usr/bin/env python """ CLI tool to runs various tasks related to QA. """ import os import time from pathlib import Path import sys import traceback import json import yaml import uuid import datetime import click from .run import RunContext from .runners import runners, Job, JobGroup from .runners.lsf import LsfPriority from .conventions import batch_dir, batch_dir, make_batch_dir, make_batch_conf_dir, make_hash from .conventions import serialize_config, deserialize_config, get_settings from .utils import PathType, entrypoint_module, load_tuning_search from .utils import save_outputs_manifest, total_storage from .utils import redirect_std_streams from .utils import getenvs from .api import url_to_dir, print_url from .api import get_outputs, notify_qa_database, serialize_paths from .iterators import iter_inputs, iter_parameters from .config import config_has_error, ignore_config_errors from .config import project, project_root, subproject, config from .config import default_batches_files, get_default_database, default_batch_label, default_platform from .config import get_default_configuration, default_input_type from .config import commit_id, outputs_commit, artifacts_commit, root_qatools, artifacts_commit_root, outputs_commit_root from .config import user, is_ci, on_windows @click.group() @click.pass_context @click.option('--platform', default=default_platform) @click.option('--configuration', '--config', '-c', 'configurations', multiple=True, help="Will be passed to the run function") @click.option('--label', '-l', default=default_batch_label, help="Gives tuning experiments a name.") @click.option('--tuning', default=None, help="Extra parameters for tuning (JSON)") @click.option('--tuning-filepath', type=PathType(), default=None, help="File with extra parameters for tuning") @click.option('--dryrun', is_flag=True, help="Only show the commands that would be executed") @click.option('--share', is_flag=True, help="Show outputs in QA-Board, doesn't just save them locally.") @click.option('--database', type=PathType(), help="Input database location") @click.option('--type', 'input_type', default=default_input_type, help="How we define inputs") @click.option('--offline', is_flag=True, help="Do not notify QA-Board about run statuses.") def qa(ctx, platform, configurations, label, tuning, tuning_filepath, dryrun, share, database, input_type, offline): """Entrypoint to running your algo, launching batchs...""" # We want all paths to be relative to top-most qaboard.yaml # it should be located at the root of the git repository if config_has_error and not ignore_config_errors: click.secho('Please fix the error(s) above in qaboard.yaml', fg='red', err=True, bold=True) exit(1) # Click passes `ctx.obj` to downstream commands, we can use it as a scratchpad # http://click.pocoo.org/6/complex/ ctx.obj = {} will_show_help = '-h' in sys.argv or '--help' in sys.argv noop_command = 'get' in sys.argv or 'init' in sys.argv if root_qatools and root_qatools != Path().resolve() and not will_show_help and not noop_command: ctx.obj['previous_cwd'] = os.getcwd() click.echo(click.style("Working directory changed to: ", fg='blue') + click.style(str(root_qatools), fg='blue', bold=True), err=True) os.chdir(root_qatools) # We want open permissions on outputs and artifacts # it makes collaboration among mutliple users / automated tools so much easier... os.umask(0) ctx.obj['project'] = project ctx.obj['project_root'] = project_root ctx.obj['subproject'] = subproject ctx.obj['HOST'] = os.environ.get('HOST', os.environ.get('HOSTNAME')) ctx.obj['user'] = user ctx.obj['dryrun'] = dryrun ctx.obj['share'] = share ctx.obj['offline'] = offline ctx.obj['outputs_commit'] = outputs_commit ctx.obj['artifacts_commit'] = artifacts_commit # Note: to support multiple databases per project, # either use / as database, or somehow we need to hash the db in the output path. ctx.obj['raw_batch_label'] = label ctx.obj['batch_label'] = label if not share else f"@{user}| {label}" ctx.obj['platform'] = platform ctx.obj['input_type'] = input_type ctx.obj['inputs_settings'] = get_settings(input_type, config) ctx.obj['database'] = database if database else get_default_database(ctx.obj['inputs_settings']) # configuration singular is for backward compatibility to a time where there was a single str config ctx.obj['configuration'] = ':'.join(configurations) if configurations else get_default_configuration(ctx.obj['inputs_settings']) # we should refactor the str configuration away completly, and do a much simpler parsing, like # deserialize_config = lambda configurations: return [maybe_json_loads(c) for c in configurations] ctx.obj['configurations'] = deserialize_config(ctx.obj['configuration']) ctx.obj['extra_parameters'] = {} if tuning: ctx.obj['extra_parameters'] = json.loads(tuning) elif tuning_filepath: ctx.obj['tuning_filepath'] = tuning_filepath with tuning_filepath.open('r') as f: if tuning_filepath.suffix == '.yaml': ctx.obj['extra_parameters'] = yaml.load(f, Loader=yaml.SafeLoader) elif tuning_filepath.suffix == '.cde': from cde import Config ctx.obj['extra_parameters'] = Config.loads(f.read()).asdict() else: ctx.obj['extra_parameters'] = json.load(f) # batch runs will override this since batches may have different configurations ctx.obj['batch_conf_dir'] = make_batch_conf_dir(outputs_commit, ctx.obj['batch_label'], platform, ctx.obj['configurations'], ctx.obj['extra_parameters'], share) ctx.obj['batch_dir'] = make_batch_dir(outputs_commit, ctx.obj['batch_label'], platform, ctx.obj['configurations'], ctx.obj['extra_parameters'], share) # For convenience, we allow users to change environment variables using {ENV: {VAR: value}} # in configurations or tuning parameters environment_variables = {} for c in ctx.obj['configurations']: if not isinstance(c, dict): continue if 'ENV' in c: environment_variables.update(c['ENV']) if 'ENV' in ctx.obj['extra_parameters']: environment_variables.update(ctx.obj['extra_parameters']['ENV']) os.environ.update(environment_variables) # we manage stripping ansi color codes ourselfs since we redirect std streams # to both the original stream and a log file ctx.color = True # colors in log files colors will be interpreted in the UIs ctx.obj['color'] = is_ci or share @qa.command() @click.option('-i', '--input', 'input_path', type=PathType(), help='Path of the input/recording/test we should work on, relative to the database directory.') @click.option('-o', '--output', 'output_path', type=PathType(), default=None, help='Custom output directory path. If not provided, defaults to ctx.obj["batch_conf_dir"] / input_path.with_suffix('')') @click.argument('variable') @click.pass_context def get(ctx, input_path, output_path, variable): """Prints the value of the requested variable. Mostly useful for debug.""" try: output_directory = ctx.obj['batch_conf_dir'] / input_path.with_suffix('') if not output_path else output_path except: pass from .config import outputs_commit, commit_branch, artifacts_branch_root # backward compatibility if variable == "branch_ci_dir": variable = "artifacts_branch_root" if variable == "commit_ci_dir": variable = "outputs_commit" locals().update(globals()) locals().update(ctx.obj) if variable in locals(): print(locals().get(variable)) else: click.secho(f"Could not find {variable}", err=True, fg='red') exit(1) @qa.command(context_settings=dict( ignore_unknown_options=True, allow_interspersed_args=False, )) @click.pass_context @click.option('-i', '--input', 'input_path', required=True, type=PathType(), help='Path of the input/recording/test we should work on, relative to the database directory.') @click.option('-o', '--output', 'output_path', type=PathType(), default=None, help='Custom output directory path. If not provided, defaults to ctx.obj["batch_conf_dir"] / input_path.with_suffix('')') @click.option('--keep-previous', is_flag=True, help="Don't clean previous outputs before the run.") @click.option('--no-postprocess', is_flag=True, help="Don't do the postprocessing.") @click.option('--save-manifests-in-database', is_flag=True, help="Save the input and outputs manifests in the database.") @click.argument('forwarded_args', nargs=-1, type=click.UNPROCESSED) def run(ctx, input_path, output_path, keep_previous, no_postprocess, forwarded_args, save_manifests_in_database): """ Runs over a given input/recording/test and computes various success metrics and outputs. """ run_context = RunContext.from_click_run_context(ctx, config) # Usually we want to remove any files already present in the output directory. # It avoids issues with remaining state... This said, # In some cases users want to debug long, multi-stepped runs, for which they have their own caching if not keep_previous: import shutil shutil.rmtree(run_context.output_dir, ignore_errors=True) run_context.output_dir.mkdir(parents=True, exist_ok=True) with (run_context.output_dir / 'run.json').open('w') as f: json.dump({ # run_context.database is always made absolute, we keep it relative if given so "database": str(ctx.obj["database"]), "input_path": str(run_context.rel_input_path), "input_type": run_context.type, "configurations": run_context.configurations, "extra_parameters": run_context.extra_parameters, "platform": run_context.platform, }, f, sort_keys=True, indent=2, separators=(',', ': ')) # Without this, we can only log runs from `qa batch`, on linux, via LSF # this redirect is not 100% perfect, we don't get stdout from C calls # if not 'LSB_JOBID' in os.environ: # When using LSF, we usally already have incremental logs with redirect_std_streams(run_context.output_dir / 'log.txt', color=ctx.obj['color']): # Help reproduce qa runs with something copy-pastable in the logs if is_ci: from shlex import quote click.secho(' '.join(['qa', *map(quote, sys.argv[1:])]), fg='cyan', bold=True) click.echo(click.style("Outputs: ", fg='cyan') + click.style(str(run_context.output_dir), fg='cyan', bold=True), err=True) print_url(ctx) if not ctx.obj['offline']: notify_qa_database(**ctx.obj, is_pending=True, is_running=True) start = time.time() cwd = os.getcwd() try: runtime_metrics = entrypoint_module(config).run(run_context) except Exception as e: exc_type, exc_value, exc_traceback = sys.exc_info() click.secho(f'[ERROR] Your `run` function raised an exception: {e}', fg='red', bold=True) try: exc_type, exc_value, exc_traceback = sys.exc_info() click.secho(''.join(traceback.format_exception(exc_type, exc_value, exc_traceback)), fg='red') except Exception as e: # debug strange stale file errors, ideally remove this... print(f"ERROR: {e}") runtime_metrics = {'is_failed': True} if not runtime_metrics: click.secho('[WARNING] Your `run` function should return a dict with a least {"is_failed": False}', fg='yellow') runtime_metrics = {"is_failed": False} if not isinstance(runtime_metrics, dict): click.secho(f'[ERROR] Your `run` function did not return a dict, but {runtime_metrics}', fg='red', bold=True) runtime_metrics = {'is_failed': True} runtime_metrics['compute_time'] = time.time() - start # avoid issues if code in run() changes cwd if os.getcwd() != cwd: os.chdir(cwd) metrics = postprocess_(runtime_metrics, run_context, skip=no_postprocess or runtime_metrics['is_failed'], save_manifests_in_database=save_manifests_in_database) if not metrics: metrics = runtime_metrics if metrics['is_failed']: click.secho('[ERROR] The run has failed.', fg='red', err=True) click.secho(str(metrics), fg='red', bold=True) exit(1) else: click.secho(str(metrics), fg='green') def postprocess_(runtime_metrics, run_context, skip=False, save_manifests_in_database=False): """Computes computes various success metrics and outputs.""" from .utils import file_info try: if not skip: try: entrypoint_postprocess = entrypoint_module(config).postprocess except: metrics = runtime_metrics else: metrics = entrypoint_postprocess(runtime_metrics, run_context) else: metrics = runtime_metrics except: exc_type, exc_value, exc_traceback = sys.exc_info() # TODO: in case of import error because postprocess was not defined, just ignore it...? # TODO: we should provide a default postprocess function, that reads metrics.json and returns {**previous, **runtime_metrics} exc_type, exc_value, exc_traceback = sys.exc_info() click.secho(f'[ERROR] Your `postprocess` function raised an exception:', fg='red', bold=True) click.secho(''.join(traceback.format_exception(exc_type, exc_value, exc_traceback)), fg='red') metrics = {**runtime_metrics, 'is_failed': True} if 'is_failed' not in metrics: click.secho("[Warning] The result of the `postprocess` function misses a key `is_failed` (bool)", fg='yellow') metrics['is_failed'] = False if (run_context.output_dir / 'metrics.json').exists(): with (run_context.output_dir / 'metrics.json').open('r') as f: previous_metrics = json.load(f) metrics = { **previous_metrics, **metrics, } with (run_context.output_dir / 'metrics.json').open('w') as f: json.dump(metrics, f, sort_keys=True, indent=2, separators=(',', ': ')) # To help identify if input files change, we compute and save some metadata. if is_ci or save_manifests_in_database: manifest_inputs = run_context.obj.get('manifest-inputs', [run_context.input_path]) input_files = {} for manifest_input in manifest_inputs: manifest_input = Path(manifest_input) if manifest_input.is_dir(): for idx, path in enumerate(manifest_input.rglob('*')): if idx >= 200: break if not path.is_file(): continue input_files[path.as_posix()] = file_info(path, config=config) elif manifest_input.is_file(): input_files.update({manifest_input.as_posix(): file_info(manifest_input, config=config)}) with (run_context.output_dir / 'manifest.inputs.json').open('w') as f: json.dump(input_files, f, indent=2) outputs_manifest = save_outputs_manifest(run_context.output_dir, config=config) output_data = { 'storage': total_storage(outputs_manifest), } if save_manifests_in_database: if run_context.input_path.is_file(): click.secho('WARNING: saving the manifests in the database is only implemented for inputs that are *folders*.', fg='yellow', err=True) else: from .utils import copy copy(run_context.output_dir / 'manifest.inputs.json', run_context.input_path / 'manifest.inputs.json') copy(run_context.output_dir / 'manifest.outputs.json', run_context.input_path / 'manifest.outputs.json') if not run_context.obj.get('offline') and not run_context.obj.get('dryrun'): notify_qa_database(**run_context.obj, metrics=metrics, data=output_data, is_pending=False, is_running=False) return metrics @qa.command(context_settings=dict( ignore_unknown_options=True, )) @click.pass_context @click.option('-i', '--input', 'input_path', required=True, type=PathType(), help='Path of the input/recording/test we should work on, relative to the database directory.') @click.option('-o', '--output', 'output_path', type=PathType(), default=None, help='Custom output directory path. If not provided, defaults to ctx.obj["batch_conf_dir"] / input_path.with_suffix('')') @click.argument('forwarded_args', nargs=-1, type=click.UNPROCESSED) def postprocess(ctx, input_path, output_path, forwarded_args): """Run only the post-processing, assuming results already exist.""" run_context = RunContext.from_click_run_context(ctx, config) with redirect_std_streams(run_context.output_dir / 'log.txt', color=ctx.obj['color']): click.echo(click.style("Outputs: ", fg='cyan') + click.style(str(run_context.output_dir), fg='cyan', bold=True), err=True) print_url(ctx) metrics = postprocess_({}, run_context) if metrics['is_failed']: click.secho('[ERROR] The run has failed.', fg='red', err=True, bold=True) click.secho(str(metrics), fg='red') else: click.secho(str(metrics), fg='green') @qa.command(context_settings=dict( ignore_unknown_options=True, )) @click.pass_context @click.option('-i', '--input', 'input_path', required=True, type=PathType(), help='Path of the input/recording/test we should work on, relative to the database directory.') @click.option('-o', '--output', 'output_path', type=PathType(), default=None, help='Custom output directory path. If not provided, defaults to ctx.obj["batch_conf_dir"] / input_path.with_suffix('')') def sync(ctx, input_path, output_path): """Updates the database metrics using metrics.json""" run_context = RunContext.from_click_run_context(ctx, config) if (run_context.output_dir / 'metrics.json').exists(): with (run_context.output_dir / 'metrics.json').open('r') as f: metrics = json.load(f) notify_qa_database(**ctx.obj, metrics=metrics, is_pending=False, is_running=False) click.secho(str(metrics), fg='green') @qa.command(context_settings=dict( ignore_unknown_options=True, )) @click.pass_context @click.option('--output-id', 'output_id', help='Custom output directory path. If not provided, defaults to ctx.obj["batch_conf_dir"] / input_path.with_suffix('')') def wait(ctx, output_id): from .api import get_output while True: output = get_output(output_id) click.secho("...waiting") if output["is_pending"]: time.sleep(5) continue break exit(0 if not output["is_failed"] else 1) runners_config = config.get('runners', {}) if 'default' in runners_config: default_runner = runners_config['default'] else: task_runners = [r for r in runners_config if r not in ['default', 'local']] default_runner = task_runners[0] if task_runners else 'local' lsf_config = config['lsf'] if 'lsf' in config else config.get('runners', {}).get('lsf', {}) if 'lsf' in config: default_runner = 'lsf' if default_runner == 'lsf' and os.name=='nt': default_runner = 'local' local_config = config.get('runners', {}).get('local', {}) @qa.command(context_settings=dict( ignore_unknown_options=True, )) @click.option('--batch', '-b', 'batches', multiple=True, help="We run over all inputs+configs+database in those batches") @click.option('--batches-file', 'batches_files', type=PathType(), default=default_batches_files, multiple=True, help="YAML files listing batches of inputs+configs+database.") @click.option('--tuning-search', 'tuning_search_dict', help='string containing JSON describing the tuning parameters to explore') @click.option('--tuning-search-file', type=PathType(), default=None, help='tuning file describing the tuning parameters to explore') @click.option('--no-wait', is_flag=True, help="If true, returns as soon as the jobs are sent, otherwise waits for completion.") @click.option('--list', 'list_contexts', is_flag=True, help="Print as JSON details about each run we would do.") @click.option('--list-output-dirs', is_flag=True, help="Only print the prefixes for the results of each batch we run on.") @click.option('--list-inputs', is_flag=True, help="Print to stdout a JSON with a list of the inputs we would call qa run on.") @click.option('--runner', default=default_runner, help="Run runs locally or using a task queue like Celery, LSF...") @click.option('--local-concurrency', default=os.environ.get('QA_BATCH_CONCURRENCY', local_config.get('concurrency')), type=int, help="joblib's n_jobs: 0=unlimited, 2=2 at a time, -1=#cpu-1") @click.option('--lsf-threads', default=lsf_config.get('threads', 0), type=int, help="restrict number of lsf threads to use. 0=no restriction") @click.option('--lsf-memory', default=lsf_config.get('memory', 0), help="restrict memory (MB) to use. 0=no restriction") @click.option('--lsf-queue', default=lsf_config.get('queue'), help="LSF queue (-q)") @click.option('--lsf-fast-queue', default=lsf_config.get('fast_queue', lsf_config.get('queue')), help="Fast LSF queue, for interactive jobs") @click.option('--lsf-resources', default=lsf_config.get('resources', None), help="LSF resources restrictions (-R)") @click.option('--lsf-priority', default=lsf_config.get('priority', 0), type=int, help="LSF priority (-sp)") @click.option('--action-on-existing', default=config.get('outputs', {}).get('action_on_existing', "run"), help="When there are already finished successful runs, whether to do run / postprocess (only) / sync (re-use results) / skip") @click.option('--action-on-pending', default=config.get('outputs', {}).get('action_on_pending', "wait"), help="When there are already pending runs, whether to do wait (then run) / sync (use those runs' results) / skip (don't run) / continue (run as usual, can cause races)") @click.option('--prefix-outputs-path', type=PathType(), default=None, help='Custom prefix for the outputs; they will be at $prefix/$output_path') @click.argument('forwarded_args', nargs=-1, type=click.UNPROCESSED) @click.pass_context def batch(ctx, batches, batches_files, tuning_search_dict, tuning_search_file, no_wait, list_contexts, list_output_dirs, list_inputs, runner, local_concurrency, lsf_threads, lsf_memory, lsf_queue, lsf_fast_queue, lsf_resources, lsf_priority, action_on_existing, action_on_pending, prefix_outputs_path, forwarded_args): """Run on all the inputs/tests/recordings in a given batch using the LSF cluster.""" if not batches_files: click.secho(f'WARNING: Could not find how to identify input tests.', fg='red', err=True, bold=True) click.secho(f'Consider adding to qaboard.yaml somelike like:\n```\ninputs:\n batches: batches.yaml\n```', fg='red', err=True) click.secho(f'Where batches.yaml is formatted like in http://qa-docs/docs/batches-running-on-multiple-inputs', fg='red', err=True) return if not batches: if not len(forwarded_args): click.secho(f'ERROR: you must provide a batch', fg='red', err=True, bold=True) click.secho(f'Use either `qa batch BATCH`, or `qa batch --batch BATCH_2 --batch BATCH_2`', fg='red', err=True) exit(1) single_batch, *forwarded_args = forwarded_args batches = [single_batch] print_url(ctx) existing_outputs = get_outputs(ctx.obj) command_id = str(uuid.uuid4()) # unique IDs for triggered runs makes it easier to wait/cancel them os.environ['QA_BATCH']= 'true' # triggered runs will be less verbose than with just `qa run` os.environ['QA_BATCHES_FILES'] = json.dumps([str(b) for b in batches_files]) dryrun = ctx.obj['dryrun'] or list_output_dirs or list_inputs or list_contexts should_notify_qa_database = (is_ci or ctx.obj['share']) and not (dryrun or ctx.obj['offline']) if should_notify_qa_database: command_data = { "command_created_at_datetime": datetime.datetime.utcnow().isoformat(), "argv": sys.argv, "runner": runner, **ctx.obj, } job_url = getenvs(('BUILD_URL', 'CI_JOB_URL', 'CIRCLE_BUILD_URL', 'TRAVIS_BUILD_WEB_URL')) # jenkins, gitlabCI, cirlceCI, travisCI if job_url: command_data['job_url'] = job_url if not os.environ.get('QA_BATCH_COMMAND_HIDE_LOGS'): notify_qa_database(object_type='batch', command={command_id: command_data}, **ctx.obj) tuning_search, filetype = load_tuning_search(tuning_search_dict, tuning_search_file) default_runner_options = { "type": runner, "command_id": command_id, } # Each runner should add what it cares about... # TODO: Having --runner-X prefixes makes it all a mess, but still the help text is useful # TODO: It would be nice to generate the CLI help depending on the runner that's choosen, then we could use if runner == 'lsf': default_runner_options.update({ "project": lsf_config.get('project', str(project) if project else "qaboard"), "max_threads": lsf_threads, "max_memory": lsf_memory, 'resources': lsf_resources, "queue": lsf_queue, "fast_queue": lsf_fast_queue, "user": ctx.obj['user'], }) if runner == "local": default_runner_options["concurrency"] = local_concurrency if runner == 'local' or runner == 'celery': default_runner_options["cwd"] = ctx.obj['previous_cwd'] if 'previous_cwd' in ctx.obj else os.getcwd() jobs = JobGroup(job_options=default_runner_options) inputs_iter = iter_inputs(batches, batches_files, ctx.obj['database'], ctx.obj['configurations'], ctx.obj['platform'], default_runner_options, config, ctx.obj['inputs_settings']) for run_context in inputs_iter: input_configuration_str = serialize_config(run_context.configurations) for tuning_file, tuning_hash, tuning_params in iter_parameters(tuning_search, filetype=filetype, extra_parameters=ctx.obj['extra_parameters']): if not prefix_outputs_path: batch_conf_dir = make_batch_conf_dir( outputs_commit, ctx.obj["batch_label"], run_context.platform, run_context.configurations, tuning_params, ctx.obj['share'] ) else: batch_conf_dir = outputs_commit / prefix_outputs_path if tuning_file: batch_conf_dir = batch_conf_dir / Path(tuning_file).stem from qaboard.conventions import slugify_hash input_dir = run_context.rel_input_path.with_suffix('') if len(input_dir.as_posix()) > 90: input_dir = Path(slugify_hash(input_dir.as_posix(), maxlength=90)) run_context.output_dir = batch_conf_dir / input_dir if forwarded_args: run_forwarded_args = [a for a in forwarded_args if not a in ("--keep-previous", "--no-postprocess", "--save-manifests-in-database")] if run_forwarded_args: run_context.extra_parameters = {"forwarded_args": run_forwarded_args, **tuning_params} else: run_context.extra_parameters = tuning_params else: run_context.extra_parameters = tuning_params if list_output_dirs: print(run_context.output_dir) break if list_inputs: print(run_context.input_path) break matching_existing_outputs = [o for o in existing_outputs.values() if url_to_dir(o['output_dir_url']) == run_context.output_dir] matching_existing_output = matching_existing_outputs[0] if matching_existing_outputs else None # at most 1, garanteed by database constaints is_pending = matching_existing_output['is_pending'] if matching_existing_output else False is_failed = matching_existing_output['is_failed'] if matching_existing_output else run_context.is_failed() ran_before = True if matching_existing_output else run_context.ran() should_run = not is_pending and (action_on_existing=='run' or is_failed or not ran_before) if not should_run and action_on_existing=='skip': continue if is_pending and action_on_pending == 'skip': continue if not forwarded_args: forwarded_args_cli = None else: if not on_windows: # FIXME: we assume no single quotes... forwarded_args_cli = ' '.join(f"'{a}'" for a in forwarded_args) else: from .compat import escaped_for_cli forwarded_args_cli = ' '.join(escaped_for_cli(a) for a in forwarded_args) if input_configuration_str == get_default_configuration(ctx.obj['inputs_settings']): configuration_cli = None else: # We can't use --config, or "-c A -c B" until we ensure all clients updated a version supporting it if not on_windows: configuration = input_configuration_str.replace("'", "'\"'\"'") # support single-quotes configuration_cli = f"--configuration '{configuration}'" else: from .compat import escaped_for_cli configuration_cli = f'--configuration {escaped_for_cli(input_configuration_str)}' # We could serialize properly the run_context/runner_options, and e.g. call "qa --pickled-cli" and use the CLI command below just for logs... args = [ f"qa", f'--share' if ctx.obj["share"] else None, f'--offline' if ctx.obj['offline'] else None, f'--label "{ctx.obj["raw_batch_label"]}"' if ctx.obj["raw_batch_label"] != default_batch_label else None, f'--platform "{run_context.platform}"' if run_context.platform != default_platform else None, # TODO: make it customizable in batches f'--type "{run_context.type}"' if run_context.type != default_input_type else None, f'--database "{run_context.database.as_posix()}"' if run_context.database != get_default_database(ctx.obj['inputs_settings']) else None, configuration_cli, f'--tuning-filepath "{tuning_file}"' if tuning_params else None, 'run' if should_run else action_on_existing, f'--input "{run_context.rel_input_path}"', f'--output "{run_context.output_dir}"' if prefix_outputs_path else None, forwarded_args_cli if forwarded_args_cli else None, ] command = ' '.join([arg for arg in args if arg is not None]) click.secho(command, fg='cyan', err=True) click.secho(f" {run_context.output_dir if run_context.output_dir.is_absolute else run_context.output_dir.relative_to(subproject)}", fg='blue', err=True) import re if 'QA_TESTING' in os.environ: # we want to make sure we test the current code command = re.sub('^qa', 'python -m qaboard', command) if str(subproject) != '.': command = f"cd {subproject} && {command}" run_context.command = command run_context.job_options['command_id'] = command_id job = Job(run_context) if should_notify_qa_database and not is_pending: # TODO: accumulate and send all at once to avoid 100s of requests? db_output = notify_qa_database(**{ **ctx.obj, **run_context.obj, # for now we don't want to worry about backward compatibility, and input_path being abs vs relative... "is_pending": True, }) if db_output: # Note: the ID is already in the matching job above job.id = db_output["id"] if is_pending: wait_command = f"qa wait --output-id {matching_existing_output['id']}" if action_on_pending=="sync": job.id = matching_existing_output['id'] job.run_context.command = wait_command elif action_on_pending=="wait": job.run_context.command = f"{wait_command} || {job.run_context.command}" else: assert action_on_pending=="continue" jobs.append(job) if list_contexts: print(json.dumps([serialize_paths(j.run_context.asdict()) for j in jobs], indent=2)) return if not dryrun: is_failed = jobs.start( blocking=not no_wait, qa_context=ctx.obj, ) from .gitlab import gitlab_token, update_gitlab_status if gitlab_token and jobs and is_ci and 'QABOARD_TUNING' not in os.environ: update_gitlab_status(commit_id, 'failed' if is_failed else 'success', ctx.obj["batch_label"], f"{len(jobs)} results") if is_failed and not no_wait: del os.environ['QA_BATCH'] # restore verbosity print_url(ctx, status="failure") exit(1) @qa.command() # Do we want this? we could simply use groups not defined in qatools.yaml:artifacts as paths @click.option('--file', '-f', 'files', multiple=True, help="Save specific files instead of artifacts indicated by yaml file") @click.option('--exclude', 'excluded_groups', multiple=True, help="Exclude specific artifact groups") # Do we use this? yes in the API, but let's deprecate and remove for other uses... @click.option('--out', '-o', 'artifacts_path', default='', help="Path to save artifacts in case of specified files") @click.argument('groups', nargs=-1, type=click.UNPROCESSED, default=None) @click.pass_context def save_artifacts(ctx, files, excluded_groups, artifacts_path, groups): """Save the results at a standard location""" import filecmp from .config import is_in_git_repo, qatools_config_paths from .utils import copy, file_info from .compat import cased_path click.secho(f"Saving artifacts in: {artifacts_commit}", bold=True, underline=True) artifacts = {} if files: artifacts = {f"__{f}": {"glob": f} for f in files} else: if 'artifacts' not in config: config['artifacts'] = {} # We support both qaboard.yaml and qaboard.yaml for backward compatibility with SIRC's projects # Default artifacts config['artifacts']['__qaboard.yaml'] = {"glob": ['qaboard.yaml', 'qatools.yaml']} config['artifacts']['__qatools'] = {"glob": ['qatools/*', 'qa/*']} # Handle sub-projects config['artifacts']['__sub-qaboard.yaml'] = {"glob": [str(p.relative_to(root_qatools).parent / 'qaboard.yaml') for p in qatools_config_paths]} config['artifacts']['__sub-qaboard.yaml'] = {"glob": [str(p.relative_to(root_qatools).parent / 'qatools.yaml') for p in qatools_config_paths]} config['artifacts']['__metrics.yaml'] = {"glob": config.get('outputs', {}).get('metrics')} config['artifacts']['__batches.yaml'] = {"glob": default_batches_files} config['artifacts']['__envrc'] = {"glob": ['.envrc', '**/*.envrc']} if groups: if excluded_groups: groups = [g for g in groups if g not in excluded_groups] artifacts = {g: config['artifacts'][g] for g in groups if g in config['artifacts'].keys()} else: artifacts = config['artifacts'] if 'QA_VERBOSE_VERBOSE' in os.environ: print(artifacts) if not is_in_git_repo: click.secho( "You are not in a git repository, maybe in an artifacts folder. `save_artifacts` is unavailable.", fg='yellow', dim=True) exit(1) for artifact_name, artifact_config in artifacts.items(): click.secho(f'Saving artifacts: {artifact_name}', bold=True) manifest_path = artifacts_commit / 'manifests' / f'{artifact_name}.json' manifest_path.parent.mkdir(parents=True, exist_ok=True) if manifest_path.exists(): with manifest_path.open() as f: try: manifest = json.load(f) except: manifest = {} else: manifest = {} nb_files = 0 globs = artifact_config.get('glob') if not isinstance(globs, list): globs = [globs] for g in globs: if not g: continue for path in Path('.').glob(g): path = cased_path(path) if not path.is_file(): continue if artifacts_path: destination = artifacts_commit_root / artifacts_path / path else: destination = artifacts_commit_root / path if 'QA_VERBOSE_VERBOSE' in os.environ: print(destination) if destination.exists() and filecmp.cmp(str(path), str(destination), shallow=True): # when working on subprojects, the artifact might be copied already, # but manifests are saved per-subproject if path.as_posix() not in manifest: manifest[path.as_posix()] = file_info(path, config=config) continue if 'QA_VERBOSE' in os.environ or ctx.obj['dryrun']: click.secho(str(path), dim=True) if not ctx.obj['dryrun']: copy(path, destination) manifest[path.as_posix()] = file_info(path, config=config) if not ctx.obj['dryrun']: with manifest_path.open('w') as f: json.dump(manifest, f) if nb_files > 0: click.secho(f"{nb_files} files copied") if os.name == "nt" and not ctx.obj['dryrun']: # [Samsung-SIRC specific] print("... Fixing linux file permissions") try: # Windows does not set file permissions correctly on the shared storage, # it does not respect umask 0: files are not world-writable. # Trying to each_file.chmod(0o777) does not work either # The only option is to make the call from linux. # We could save a list of paths and chmod them with their parent directories... # but to make things faster to code, we just "ssh linux chmod everything" # from qaboard.compat import windows_to_linux_path # # We can assume SSH to be present on Windows10 # ssh = f"ssh -i \\\\networkdrive\\home\\{user}\\.ssh\\id_rsa -oStrictHostKeyChecking=no" # chmod = f'{ssh} {user}@{user}-srv \'chmod -R 777 "{windows_to_linux_path(artifacts_commit)}"\'' # print(chmod) # os.system(chmod) pass except Exception as e: print(f'WARNING: {e}') # if the commit was deleted, this notification will mark it as good again notify_qa_database(object_type='commit', **ctx.obj) @qa.command() @click.pass_context @click.option('--batch', '-b', 'batches', required=True, multiple=True, help="Only check bit-accuracy for this batch of inputs+configs+database.") @click.option('--batches-file', 'batches_files', type=PathType(), default=default_batches_files, multiple=True, help="YAML file listing batches of inputs+config+database selected from the database.") def check_bit_accuracy_manifest(ctx, batches, batches_files): """ Checks the bit accuracy of the results in the current ouput directory versus the latest commit on origin/develop. """ from .bit_accuracy import is_bit_accurate commit_dir = outputs_commit if is_ci else Path() all_bit_accurate = True nb_compared = 0 for run_context in iter_inputs(batches, batches_files, ctx.obj['database'], ctx.obj['configurations'], default_platform, {}, config, ctx.obj['inputs_settings']): nb_compared += 1 if run_context.input_path.is_file(): click.secho('ERROR: check_bit_accuracy_manifest only works for inputs that are folders', fg='red', err=True) # otherwise the manifest is at # * input_path.parent / 'manifest.json' in the database # * input_path.with_suffix('') / 'manifest.json' in the results # # reference_output_directory = run_context.input_path if run_context.input_path.is_folder() else run_context.input_path.parent exit(1) batch_conf_dir = make_batch_conf_dir(Path(), ctx.obj['batch_label'], ctx.obj["platform"], run_context.configurations, ctx.obj['extra_parameters'], ctx.obj['share']) input_is_bit_accurate = is_bit_accurate(commit_dir / batch_conf_dir, run_context.database, [run_context.rel_input_path]) all_bit_accurate = all_bit_accurate and input_is_bit_accurate if not all_bit_accurate: click.secho("\nError: you are not bit-accurate versus the manifest.", fg='red', underline=True, bold=True) click.secho("Reminder: the manifest lists the expected inputs/outputs for each test. It acts as an explicit gatekeeper against changes", fg='red', dim=True) if not run_context.database.is_absolute(): click.secho("If that's what you wanted, update and commit all manifests.", fg='red') # click.secho("If that's what you wanted, update all manifests using:", fg='red') # click.secho("$ qa batch * --save-manifests-in-database", fg='red') # click.secho("$ git add # your changes", fg='red') # click.secho("$ git commit # now retry your CI", fg='red') else: click.secho("To update the manifests for all tests, run:", fg='red') click.secho("$ qa batch --save-manifests --batch *", fg='red') exit(1) if not nb_compared: click.secho("\nWARNING: Nothing was compared! It's not likely to be what you expected...", fg='yellow', underline=True, bold=True) @qa.command() @click.pass_context @click.option( "--reference", default=config.get('project', {}).get('reference_branch', 'master'), help="Branch, tag or commit used as reference." ) @click.option('--batch', '-b', 'batches', multiple=True, help="Only check bit-accuracy for those batches of inputs+configs+database.") @click.option('--batches-file', 'batches_files', type=PathType(), default=default_batches_files, multiple=True, help="YAML file listing batches of inputs+config+database selected from the database.") @click.option('--reference-platform', help="Compare against a difference platform.") def check_bit_accuracy(ctx, reference, batches, batches_files, reference_platform): """ Checks the bit accuracy of the results in the current ouput directory versus the latest commit on origin/develop. """ from .config import is_in_git_repo, commit_branch, is_ci, outputs_project_root, repo_root from .bit_accuracy import is_bit_accurate from .gitlab import lastest_successful_ci_commit from .conventions import get_commit_dirs from .git import latest_commit, git_show, git_parents if not is_in_git_repo: click.secho("You are not in a git repository, maybe in an artifacts folder. `check_bit_accuracy` is unavailable.", fg='yellow', dim=True) exit(1) if is_ci and commit_branch == reference: click.secho(f'We are on branch {reference}', fg='cyan', bold=True, err=True) click.secho(f"Comparing bit-accuracy against this commit's ({commit_id[:8]}) parents.", fg='cyan', bold=True, err=True) # It will work until we try to rebase merge requests. # We really should use Gitlab' API (or our database) to ask about previous pipelines on the branch reference_commits = git_parents(commit_id) else: click.secho(f'Comparing bit-accuracy versus the latest remote commit of {reference}', fg='cyan', bold=True, err=True) reference_commits = [latest_commit(reference)] click.secho(f"{commit_id[:8]} versus {reference_commits}.", fg='cyan', err=True) # This where the new results are located commit_dir = outputs_commit_root if is_ci else Path() if not batches: output_directories = list(p.parent.relative_to(commit_dir) for p in (commit_dir / subproject / 'output').rglob('manifest.outputs.json')) else: output_directories = [] for run_context in iter_inputs(batches, batches_files, ctx.obj['database'], ctx.obj['configurations'], default_platform, {}, config, ctx.obj['inputs_settings']): batch_conf_dir = make_batch_conf_dir(subproject, ctx.obj['batch_label'], ctx.obj["platform"], run_context.configurations, ctx.obj["extra_parameters"], ctx.obj['share']) input_path = run_context.input_path.relative_to(run_context.database) output_directory = batch_conf_dir / input_path.with_suffix('') output_directories.append(output_directory) for reference_commit in reference_commits: # if the reference commit is pending or failed, we wait or maybe pick a parent reference_commit = lastest_successful_ci_commit(reference_commit) click.secho(f'Current directory : {commit_dir}', fg='cyan', bold=True, err=True) reference_rootproject_ci_dir = outputs_project_root / get_commit_dirs(reference_commit, repo_root) click.secho(f"Reference directory: {reference_rootproject_ci_dir}", fg='cyan', bold=True, err=True) all_bit_accurate = True for o in output_directories: all_bit_accurate = is_bit_accurate(commit_dir, reference_rootproject_ci_dir, [o], reference_platform) and all_bit_accurate if not all_bit_accurate: click.secho(f"\nERROR: results are not bit-accurate to {reference_commits}.", bg='red', bold=True) if is_ci: click.secho(f"\nTo investigate, go to", fg='red', underline=True) for reference_commit in reference_commits: click.secho(f"https://qa/{project.as_posix()}/commit/{commit_id}?reference={reference_commit}&selected_views=bit_accuracy", fg='red') exit(1) from .optimize import optimize qa.add_command(optimize) # TODO: split more... # from .bit_accuracy import check_bit_accuracy, check_bit_accuracy_manifest # qa.add_command(check_bit_accuracy) # qa.add_command(check_bit_accuracy_manifest) @qa.command() @click.pass_context def init(ctx): """Provide a sample qaboard.yaml configuration.""" from .init import qa_init qa_init(ctx) def main(): from .compat import ensure_cli_backward_compatibility ensure_cli_backward_compatibility() qa(obj={}, auto_envvar_prefix='QA') if __name__ == '__main__': main()
50.720455
318
0.703992
0
0
0
0
39,018
0.874177
0
0
17,249
0.386454
4b77f58f441974f14bdaad4bde4687feee866e3a
5,838
py
Python
20210220_simulation_sample/data_handler.py
3x3x3/Presentations
3c31b136ed4d9214bb3730fa41a4a575da38edc9
[ "MIT" ]
null
null
null
20210220_simulation_sample/data_handler.py
3x3x3/Presentations
3c31b136ed4d9214bb3730fa41a4a575da38edc9
[ "MIT" ]
null
null
null
20210220_simulation_sample/data_handler.py
3x3x3/Presentations
3c31b136ed4d9214bb3730fa41a4a575da38edc9
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import threading import time import global_def as gd from db_reader import DbReaderDef, DbReaer from queue import Queue, Empty class DataHandlerThd(threading.Thread): def __init__(self, req_queue: Queue, rcv_queue: Queue, db_host: str, db_port: int, db_user: str, db_pw: str, db_name: str, db_char_set: str = 'utf8'): threading.Thread.__init__(self) self._db_host = db_host self._db_port = db_port self._db_user = db_user self._db_pw = db_pw self._db_name = db_name self._db_char_set = db_char_set self._req_queue = req_queue self._rcv_queue = rcv_queue self.is_run = False def _send_err_msg(self, msg: str) -> None: self._rcv_queue.put({ gd.KEY_NM_EVT: gd.EVT_TYPE_ERR, gd.KEY_NM_MSG: msg }) def _read_db(self, req: dict) -> bool: req_date = int(req.get(gd.KEY_NM_DATE, 0)) tbl_infos = req.get(gd.KEY_NM_TBL_INFOS, None) if 19900101 > req_date or 30000101 < req_date: self._send_err_msg('Invalid Date') return False if list != type(tbl_infos) or 0 == len(tbl_infos): self._send_err_msg('Invalid Table Infos1') return False db_readers = [] for reader_idx, tbl_info in enumerate(tbl_infos): tbl_nm = tbl_info.get(gd.KEY_NM_TBL_NM, None) col_nms = tbl_info.get(gd.KEY_NM_COL_NMS, []) if tbl_nm is None or 0 == len(col_nms): self._send_err_msg('Invalid Table Infos2') return False db_reader = DbReaer(reader_idx, req_date, tbl_nm, col_nms, self._db_host, self._db_port, self._db_user, self._db_pw, self._db_name, self._db_char_set) db_readers.append(db_reader) for db_reader in db_readers: db_reader.read_thd.start() is_st_read = False is_error = False while not is_st_read: for db_reader in db_readers: thd_state: int = db_reader.get_thd_state() if DbReaderDef.STATE_ERROR == thd_state: is_st_read = True is_error = True break elif DbReaderDef.STATE_READY == thd_state: break else: is_st_read = True time.sleep(0.5) if is_error: for db_reader in db_readers: db_reader.set_stop_thd() time.sleep(1) self._send_err_msg('Error in DbReaderThd1') return False # 처음에 하나씩 데이터를 읽는다 empty_reader_idxs = [] for reader_idx, db_reader in enumerate(db_readers): if not db_reader.read_next_data(): empty_reader_idxs.append(reader_idx) # 텅빈 Reader들을 목록에서 제거 for reader_idx in empty_reader_idxs: del db_readers[reader_idx] reader_cnt = len(db_readers) fin_readers = [] while 0 < reader_cnt: min_rtime_idx = -1 min_rtime = 9999999999999 find_min_ts = False is_exist_fin_readers = False for idx, db_reader in enumerate(db_readers): row: list = db_reader.last_data # 마지막 데이터가 비었을때 if row is None: thd_state = db_reader.get_thd_state() if DbReaderDef.STATE_WORKING == thd_state: time.sleep(0.5) db_reader.read_next_data() find_min_ts = False break elif DbReaderDef.STATE_FINISHED == thd_state: fin_readers.append(idx) is_exist_fin_readers = True continue elif DbReaderDef.STATE_ERROR == thd_state: self._send_err_msg('Error in DbReaderThd2') fin_readers.append(idx) is_exist_fin_readers = True continue pk_rtime = row[0] if min_rtime > pk_rtime: min_rtime = pk_rtime min_rtime_idx = idx find_min_ts = True # 가장 과거의 값을 찾았다면 if find_min_ts: target_reader: DbReaer = db_readers[min_rtime_idx] self._rcv_queue.put({ gd.KEY_NM_EVT: gd.EVT_TYPE_READ_DB, gd.KEY_NM_IDX: target_reader.reader_idx, gd.KEY_NM_DATA: target_reader.last_data }) target_reader.read_next_data() # 종료된 Reader가 생겼다면 if is_exist_fin_readers: fin_readers.sort(reverse=True) for fin_reader_idx in fin_readers: del db_readers[fin_reader_idx] reader_cnt = len(db_readers) fin_readers.clear() self._rcv_queue.put({ gd.KEY_NM_EVT: gd.EVT_TYPE_FIN }) return True def run(self): self.is_run = True while self.is_run: try: req = self._req_queue.get(True, 1) evt_type = req.get(gd.KEY_NM_EVT) if gd.EVT_TYPE_READ_DB == evt_type: print(f'Read DB Start!, data: {req}') self._read_db(req) print(f'Read DB End!, data: {req}') elif gd.EVT_TYPE_FIN == evt_type: break except Empty as em: pass except Exception as e: self.is_run = False break
32.797753
162
0.52381
5,790
0.974092
0
0
0
0
0
0
385
0.064771
4b786431aa3dbf51672c3a6c4d1ccbdb01c1f809
7,865
py
Python
todo/views.py
Azarn/mytodo
599b5017b9a952100f05a6180dba5bca0823ad70
[ "Apache-2.0" ]
null
null
null
todo/views.py
Azarn/mytodo
599b5017b9a952100f05a6180dba5bca0823ad70
[ "Apache-2.0" ]
null
null
null
todo/views.py
Azarn/mytodo
599b5017b9a952100f05a6180dba5bca0823ad70
[ "Apache-2.0" ]
null
null
null
import logging from rest_framework import mixins, generics, permissions, exceptions from django.conf import settings from django.utils import timezone from .serializers import CategorySerializer, TagSerializer, TodoSerializer from .models import Category, Tag, Todo logger = logging.getLogger(__name__) class MyGenericApiView(generics.GenericAPIView): # Disabling "options" method metadata_class = None def initial(self, request, *args, **kwargs): super().initial(request, *args, **kwargs) timezone.activate(request.user.profile.timezone) @staticmethod def _raise_invalid_param(param_name): raise exceptions.ParseError('parameter `{0}` is invalid'.format(param_name)) def parse_get_int(self, param_name, default=None): param = self.request.query_params.get(param_name, default) if param != default: try: param = int(param) except ValueError: self._raise_invalid_param(param_name) return param def parse_get_bool(self, param_name, default=None): param = self.parse_get_int(param_name, default) if param != default: if param not in (0, 1): self._raise_invalid_param(param_name) param = bool(param) return param # Hiding "options" from available methods @property def allowed_methods(self): methods = super().allowed_methods methods.remove('OPTIONS') return methods class CategoryList(mixins.ListModelMixin, mixins.CreateModelMixin, MyGenericApiView): serializer_class = CategorySerializer permission_classes = (permissions.IsAuthenticated,) def get_queryset(self): return Category.objects.filter(user=self.request.user) def perform_create(self, serializer): serializer.save(user=self.request.user) def get(self, request, *args, **kwargs): return self.list(request, args, kwargs) def post(self, request, *args, **kwargs): return self.create(request, *args, **kwargs) class CategoryDetail(mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, MyGenericApiView): serializer_class = CategorySerializer permission_classes = (permissions.IsAuthenticated,) def get_queryset(self): return Category.objects.filter(user=self.request.user) def get(self, request, *args, **kwargs): return self.retrieve(request, *args, **kwargs) def put(self, request, *args, **kwargs): return self.update(request, *args, partial=True, **kwargs) def delete(self, request, *args, **kwargs): return self.destroy(request, *args, **kwargs) class TagList(mixins.ListModelMixin, mixins.CreateModelMixin, MyGenericApiView): serializer_class = TagSerializer permission_classes = (permissions.IsAuthenticated,) def get_queryset(self): return Tag.objects.filter(user=self.request.user) def perform_create(self, serializer): serializer.save(user=self.request.user) def get(self, request, *args, **kwargs): return self.list(request, args, kwargs) def post(self, request, *args, **kwargs): return self.create(request, *args, **kwargs) class TagDetail(mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, MyGenericApiView): serializer_class = TagSerializer permission_classes = (permissions.IsAuthenticated,) def get_queryset(self): return Tag.objects.filter(user=self.request.user) def get(self, request, *args, **kwargs): return self.retrieve(request, *args, **kwargs) def put(self, request, *args, **kwargs): return self.update(request, *args, partial=True, **kwargs) def delete(self, request, *args, **kwargs): return self.destroy(request, *args, **kwargs) class TodoList(mixins.ListModelMixin, mixins.CreateModelMixin, MyGenericApiView): serializer_class = TodoSerializer permission_classes = (permissions.IsAuthenticated,) def get_queryset(self): """ Gets query according to GET params Available GET params: only_done: if specified, todos will be filtered by `todo.is_done` = only_done category: if specified todos will be filtered by this category tags: if specified todos will be filtered by this tags list only_one_day: if specified changes behaviour of by_date(see below) to show todos only for one day by_date: if specified todos will be filtered by this date, if it is equal to `None`, filters todos without deadline :return: queryset """ q = Todo.objects.filter(user=self.request.user) only_done = self.parse_get_bool('only_done') only_one_day = self.parse_get_bool('only_one_day', False) category = self.request.query_params.get('category') tags = self.request.query_params.getlist('tags') by_date = self.request.query_params.get('by_date') if only_done is not None: if only_done: q = q.filter(is_done=True) else: q = q.filter(is_done=False) if category is not None: try: category = int(category) except ValueError: raise exceptions.ParseError('parameter `category` is invalid') else: q = q.filter(category__pk=category) if tags: try: tags = list(map(int, tags)) except ValueError: raise exceptions.ParseError('parameter `tags` is invalid') else: for t in tags: q = q.filter(tags__pk=t) if by_date is not None: if by_date in ('today', 'tomorrow', 'week', 'none'): date = timezone.localtime(timezone.now()) else: try: date = timezone.datetime.strptime(by_date, settings.DATE_FORMAT) except TypeError: raise exceptions.ParseError('parameter `by_date` is invalid') date = timezone.make_aware(timezone.datetime.combine(date, timezone.datetime.max.time())) if by_date == 'tomorrow': date += timezone.timedelta(days=1) elif by_date == 'week': date += timezone.timedelta(days=6) logger.warn(str(date)) if by_date == 'none': q = q.filter(deadline__isnull=True) elif only_one_day: q = q.filter(deadline__date=date) else: q = q.filter(deadline__lte=date) return q.prefetch_related('tags') def perform_create(self, serializer): serializer.save(user=self.request.user) def get(self, request, *args, **kwargs): return self.list(request, args, kwargs) def post(self, request, *args, **kwargs): return self.create(request, *args, **kwargs) class TodoDetail(mixins.RetrieveModelMixin, mixins.UpdateModelMixin, mixins.DestroyModelMixin, MyGenericApiView): serializer_class = TodoSerializer permission_classes = (permissions.IsAuthenticated,) def get_queryset(self): return Todo.objects.filter(user=self.request.user) def get(self, request, *args, **kwargs): return self.retrieve(request, *args, **kwargs) def put(self, request, *args, **kwargs): return self.update(request, *args, partial=True, **kwargs) def delete(self, request, *args, **kwargs): return self.destroy(request, *args, **kwargs)
34.047619
105
0.628608
7,537
0.958296
0
0
279
0.035474
0
0
885
0.112524
4b7a04ca06d8701872be7f11c6588abbce31dce4
16,294
py
Python
hypothesis/_settings.py
EnjoyLifeFund/macHighSierra-py36-pkgs
5668b5785296b314ea1321057420bcd077dba9ea
[ "BSD-3-Clause", "BSD-2-Clause", "MIT" ]
null
null
null
hypothesis/_settings.py
EnjoyLifeFund/macHighSierra-py36-pkgs
5668b5785296b314ea1321057420bcd077dba9ea
[ "BSD-3-Clause", "BSD-2-Clause", "MIT" ]
null
null
null
hypothesis/_settings.py
EnjoyLifeFund/macHighSierra-py36-pkgs
5668b5785296b314ea1321057420bcd077dba9ea
[ "BSD-3-Clause", "BSD-2-Clause", "MIT" ]
null
null
null
# coding=utf-8 # # This file is part of Hypothesis (https://github.com/DRMacIver/hypothesis) # # Most of this work is copyright (C) 2013-2015 David R. MacIver # ([email protected]), but it contains contributions by others. See # https://github.com/DRMacIver/hypothesis/blob/master/CONTRIBUTING.rst for a # full list of people who may hold copyright, and consult the git log if you # need to determine who owns an individual contribution. # # This Source Code Form is subject to the terms of the Mozilla Public License, # v. 2.0. If a copy of the MPL was not distributed with this file, You can # obtain one at http://mozilla.org/MPL/2.0/. # # END HEADER """A module controlling settings for Hypothesis to use in falsification. Either an explicit settings object can be used or the default object on this module can be modified. """ from __future__ import division, print_function, absolute_import import os import inspect import warnings import threading from collections import namedtuple from hypothesis.errors import InvalidArgument, HypothesisDeprecationWarning from hypothesis.configuration import hypothesis_home_dir from hypothesis.utils.conventions import not_set from hypothesis.utils.dynamicvariables import DynamicVariable __all__ = [ 'settings', ] all_settings = {} _db_cache = {} class SettingsProperty(object): def __init__(self, name): self.name = name def __get__(self, obj, type=None): if obj is None: return self else: try: return obj.__dict__[self.name] except KeyError: raise AttributeError(self.name) def __set__(self, obj, value): obj.__dict__[self.name] = value def __delete__(self, obj): try: del obj.__dict__[self.name] except KeyError: raise AttributeError(self.name) @property def __doc__(self): return '\n'.join(( all_settings[self.name].description, 'default value: %r' % (getattr(settings.default, self.name),) )) default_variable = DynamicVariable(None) class SettingsMeta(type): def __init__(self, *args, **kwargs): super(SettingsMeta, self).__init__(*args, **kwargs) @property def default(self): return default_variable.value @default.setter def default(self, value): if default_variable.value is not None: raise AttributeError('Cannot assign settings.default') self._assign_default_internal(value) def _assign_default_internal(self, value): default_variable.value = value class settings(SettingsMeta('settings', (object,), {})): """A settings object controls a variety of parameters that are used in falsification. These may control both the falsification strategy and the details of the data that is generated. Default values are picked up from the settings.default object and changes made there will be picked up in newly created settings. """ _WHITELISTED_REAL_PROPERTIES = [ '_database', '_construction_complete', 'storage' ] __definitions_are_locked = False _profiles = {} def __getattr__(self, name): if name in all_settings: d = all_settings[name].default if inspect.isfunction(d): d = d() return d else: raise AttributeError('settings has no attribute %s' % (name,)) def __init__( self, parent=None, **kwargs ): self._construction_complete = False self._database = kwargs.pop('database', not_set) explicit_kwargs = list(kwargs) defaults = parent or settings.default if defaults is not None: for setting in all_settings.values(): if kwargs.get(setting.name, not_set) is not_set: kwargs[setting.name] = getattr(defaults, setting.name) if self._database is not_set: self._database = defaults.database for name, value in kwargs.items(): if name not in all_settings: raise InvalidArgument( 'Invalid argument %s' % (name,)) setattr(self, name, value) self.storage = threading.local() self._construction_complete = True for k in explicit_kwargs: deprecation = all_settings[k].deprecation if deprecation: note_deprecation(deprecation, self) def defaults_stack(self): try: return self.storage.defaults_stack except AttributeError: self.storage.defaults_stack = [] return self.storage.defaults_stack def __call__(self, test): test._hypothesis_internal_use_settings = self return test @classmethod def define_setting( cls, name, description, default, options=None, deprecation=None, ): """Add a new setting. - name is the name of the property that will be used to access the setting. This must be a valid python identifier. - description will appear in the property's docstring - default is the default value. This may be a zero argument function in which case it is evaluated and its result is stored the first time it is accessed on any given settings object. """ if settings.__definitions_are_locked: from hypothesis.errors import InvalidState raise InvalidState( 'Settings have been locked and may no longer be defined.' ) if options is not None: options = tuple(options) if default not in options: raise InvalidArgument( 'Default value %r is not in options %r' % ( default, options ) ) all_settings[name] = Setting( name, description.strip(), default, options, deprecation) setattr(settings, name, SettingsProperty(name)) @classmethod def lock_further_definitions(cls): settings.__definitions_are_locked = True def __setattr__(self, name, value): if name in settings._WHITELISTED_REAL_PROPERTIES: return object.__setattr__(self, name, value) elif name == 'database': if self._construction_complete: raise AttributeError( 'Settings objects are immutable and may not be assigned to' ' after construction.' ) else: return object.__setattr__(self, '_database', value) elif name in all_settings: if self._construction_complete: raise AttributeError( 'Settings objects are immutable and may not be assigned to' ' after construction.' ) else: setting = all_settings[name] if ( setting.options is not None and value not in setting.options ): raise InvalidArgument( 'Invalid %s, %r. Valid options: %r' % ( name, value, setting.options ) ) return object.__setattr__(self, name, value) else: raise AttributeError('No such setting %s' % (name,)) def __repr__(self): bits = [] for name in all_settings: value = getattr(self, name) bits.append('%s=%r' % (name, value)) bits.sort() return 'settings(%s)' % ', '.join(bits) @property def database(self): """An ExampleDatabase instance to use for storage of examples. May be None. If this was explicitly set at settings instantiation then that value will be used (even if it was None). If not and the database_file setting is not None this will be lazily loaded as an SQLite backed ExampleDatabase using that file the first time this property is accessed on a particular thread. """ try: if self._database is not_set and self.database_file is not None: from hypothesis.database import ExampleDatabase from hypothesis.database.backend import SQLiteBackend if self.database_file not in _db_cache: _db_cache[self.database_file] = ( ExampleDatabase( backend=SQLiteBackend(self.database_file))) return _db_cache[self.database_file] if self._database is not_set: self._database = None return self._database except AttributeError: import traceback traceback.print_exc() assert False def __enter__(self): default_context_manager = default_variable.with_value(self) self.defaults_stack().append(default_context_manager) default_context_manager.__enter__() return self def __exit__(self, *args, **kwargs): default_context_manager = self.defaults_stack().pop() return default_context_manager.__exit__(*args, **kwargs) @staticmethod def register_profile(name, settings): """registers a collection of values to be used as a settings profile. These settings can be loaded in by name. Enable different defaults for different settings. - settings is a settings object """ settings._profiles[name] = settings @staticmethod def get_profile(name): """Return the profile with the given name. - name is a string representing the name of the profile to load A InvalidArgument exception will be thrown if the profile does not exist """ try: return settings._profiles[name] except KeyError: raise InvalidArgument( "Profile '{0}' has not been registered".format( name ) ) @staticmethod def load_profile(name): """Loads in the settings defined in the profile provided If the profile does not exist an InvalidArgument will be thrown. Any setting not defined in the profile will be the library defined default for that setting """ settings._assign_default_internal(settings.get_profile(name)) Setting = namedtuple( 'Setting', ( 'name', 'description', 'default', 'options', 'deprecation')) settings.define_setting( 'min_satisfying_examples', default=5, description=""" Raise Unsatisfiable for any tests which do not produce at least this many values that pass all assume() calls and which have not exhaustively covered the search space. """ ) settings.define_setting( 'max_examples', default=200, description=""" Once this many satisfying examples have been considered without finding any counter-example, falsification will terminate. """ ) settings.define_setting( 'max_iterations', default=1000, description=""" Once this many iterations of the example loop have run, including ones which failed to satisfy assumptions and ones which produced duplicates, falsification will terminate. """ ) settings.define_setting( 'max_shrinks', default=500, description=""" Once this many successful shrinks have been performed, Hypothesis will assume something has gone a bit wrong and give up rather than continuing to try to shrink the example. """ ) settings.define_setting( 'timeout', default=60, description=""" Once this many seconds have passed, falsify will terminate even if it has not found many examples. This is a soft rather than a hard limit - Hypothesis won't e.g. interrupt execution of the called function to stop it. If this value is <= 0 then no timeout will be applied. """ ) settings.define_setting( 'derandomize', default=False, description=""" If this is True then hypothesis will run in deterministic mode where each falsification uses a random number generator that is seeded based on the hypothesis to falsify, which will be consistent across multiple runs. This has the advantage that it will eliminate any randomness from your tests, which may be preferable for some situations . It does have the disadvantage of making your tests less likely to find novel breakages. """ ) settings.define_setting( 'strict', default=os.getenv('HYPOTHESIS_STRICT_MODE') == 'true', description=""" If set to True, anything that would cause Hypothesis to issue a warning will instead raise an error. Note that new warnings may be added at any time, so running with strict set to True means that new Hypothesis releases may validly break your code. You can enable this setting temporarily by setting the HYPOTHESIS_STRICT_MODE environment variable to the string 'true'. """ ) settings.define_setting( 'database_file', default=lambda: ( os.getenv('HYPOTHESIS_DATABASE_FILE') or os.path.join(hypothesis_home_dir(), 'examples.db') ), description=""" database: An instance of hypothesis.database.ExampleDatabase that will be used to save examples to and load previous examples from. May be None in which case no storage will be used. """ ) class Verbosity(object): def __repr__(self): return 'Verbosity.%s' % (self.name,) def __init__(self, name, level): self.name = name self.level = level def __eq__(self, other): return isinstance(other, Verbosity) and ( self.level == other.level ) def __ne__(self, other): return not self.__eq__(other) def __hash__(self): return self.level def __lt__(self, other): return self.level < other.level def __le__(self, other): return self.level <= other.level def __gt__(self, other): return self.level > other.level def __ge__(self, other): return self.level >= other.level @classmethod def by_name(cls, key): result = getattr(cls, key, None) if isinstance(result, Verbosity): return result raise InvalidArgument('No such verbosity level %r' % (key,)) Verbosity.quiet = Verbosity('quiet', 0) Verbosity.normal = Verbosity('normal', 1) Verbosity.verbose = Verbosity('verbose', 2) Verbosity.debug = Verbosity('debug', 3) Verbosity.all = [ Verbosity.quiet, Verbosity.normal, Verbosity.verbose, Verbosity.debug ] ENVIRONMENT_VERBOSITY_OVERRIDE = os.getenv('HYPOTHESIS_VERBOSITY_LEVEL') if ENVIRONMENT_VERBOSITY_OVERRIDE: DEFAULT_VERBOSITY = Verbosity.by_name(ENVIRONMENT_VERBOSITY_OVERRIDE) else: DEFAULT_VERBOSITY = Verbosity.normal settings.define_setting( 'verbosity', options=Verbosity.all, default=DEFAULT_VERBOSITY, description='Control the verbosity level of Hypothesis messages', ) settings.define_setting( name='stateful_step_count', default=50, description=""" Number of steps to run a stateful program for before giving up on it breaking. """ ) settings.define_setting( 'perform_health_check', default=True, description=u""" If set to True, Hypothesis will run a preliminary health check before attempting to actually execute your test. """ ) settings.lock_further_definitions() settings.register_profile('default', settings()) settings.load_profile('default') assert settings.default is not None def note_deprecation(message, s=None): # If *either* self or the current default are non-strict # then this should be an error. This is to handle e.g. the case # where defining a new setting while non-strict updates a # profile which is strict. This should not be an error, but # using the profile here would cause it to be one. if s is None: s = settings.default assert s is not None strict = settings.default.strict and s.strict verbosity = s.verbosity warning = HypothesisDeprecationWarning(message) if strict: raise warning elif verbosity > Verbosity.quiet: warnings.warn(warning, stacklevel=3)
31.334615
79
0.650669
10,162
0.623665
0
0
4,463
0.273905
0
0
6,233
0.382533
4b7b56b22f9f50b0ab5dcb31b7bb4cdc39078ed0
2,014
py
Python
2_writeups/4_other/pdf/pipeline.py
araujorayza/robot_hacking_manual
d11feecc8931b1449b0ab30a51a55f71f51dd965
[ "Apache-2.0" ]
141
2021-11-14T15:27:04.000Z
2022-03-30T00:44:48.000Z
2_writeups/4_other/pdf/pipeline.py
araujorayza/robot_hacking_manual
d11feecc8931b1449b0ab30a51a55f71f51dd965
[ "Apache-2.0" ]
1
2021-11-17T06:38:44.000Z
2021-11-17T06:38:45.000Z
2_writeups/4_other/pdf/pipeline.py
araujorayza/robot_hacking_manual
d11feecc8931b1449b0ab30a51a55f71f51dd965
[ "Apache-2.0" ]
18
2021-11-15T09:55:48.000Z
2022-03-08T10:25:58.000Z
""" Script to generate a security pipeline for PDF files. It does the following: - Adds specified meta-data - Encrypts file Run: python3 pipeline.py """ from PyPDF2 import PdfFileWriter, PdfFileReader from PyPDF2.generic import NameObject, createStringObject def encrypt(input_pdf, output_pdf, password): pdf_writer = PdfFileWriter() pdf_reader = PdfFileReader(input_pdf) for page in range(pdf_reader.getNumPages()): pdf_writer.addPage(pdf_reader.getPage(page)) pdf_writer.encrypt(user_pwd=password, owner_pwd=None, use_128bit=True) with open(output_pdf, 'wb') as fh: pdf_writer.write(fh) def meta(input_pdf, output_pdf, value): pdf_writer = PdfFileWriter() pdf_reader = PdfFileReader(input_pdf) for page in range(pdf_reader.getNumPages()): pdf_writer.addPage(pdf_reader.getPage(page)) # pdf_writer.encrypt(user_pwd=password, owner_pwd=None, # use_128bit=True) infoDict = pdf_writer._info.getObject() infoDict.update({NameObject('/Version'): createStringObject(u'234ds2')}) info = pdf_reader.documentInfo for key in info: infoDict.update({NameObject(key): createStringObject(info[key])}) # add the grade # infoDict.update({NameObject('/Grade'): createStringObject(u'A+')}) # infoDict.update({NameObject('/Grade2'): createStringObject(u'A+')}) infoDict.update({NameObject('/Key'): createStringObject(value)}) with open(output_pdf, 'wb') as fh: pdf_writer.write(fh) if __name__ == '__main__': # path for the file to process filepath = "/Users/victor/Desktop/Apex.AI_Threat_Model_AliasRobotics.pdf" # meta-data-value meta_value = u'HitachiVentures' meta(input_pdf=filepath, output_pdf=filepath+"underNDA.pdf", value=meta_value) encrypt(input_pdf=filepath+"underNDA.pdf", output_pdf=filepath+"underNDA_encrypted.pdf", password='4l14srobotics')
28.366197
77
0.682224
0
0
0
0
0
0
0
0
639
0.317279
4b7c945d6b1d560f6d85d5ab876aed99787d4072
1,989
py
Python
code/MergeTrack/print_max_reid_distance.py
MTonyM/PReMVOS
3d01f0c6156628083a4c8441b4b57622c500e04e
[ "MIT" ]
140
2018-10-25T11:58:34.000Z
2022-01-18T15:29:38.000Z
code/MergeTrack/print_max_reid_distance.py
MTonyM/PReMVOS
3d01f0c6156628083a4c8441b4b57622c500e04e
[ "MIT" ]
18
2018-11-21T04:48:03.000Z
2020-09-14T09:30:56.000Z
code/MergeTrack/print_max_reid_distance.py
MTonyM/PReMVOS
3d01f0c6156628083a4c8441b4b57622c500e04e
[ "MIT" ]
32
2018-10-25T11:58:57.000Z
2021-12-27T06:13:45.000Z
import glob from numpy.linalg import norm import numpy as np from copy import deepcopy as copy from MergeTrack.merge_functions import read_ann,read_props from MergeTrack.ReID_net_functions import ReID_net_init, add_ReID input_images = "DAVIS/val17/" input_proposals = "DAVIS/ReID_props/" first_frame_anns = "DAVIS/val17-ff/" output_images = "DAVIS/final_results/" output_proposals = "DAVIS/final_props/" ReID_net = ReID_net_init() dataset_max_distances = [] for video_fn in sorted(glob.glob(input_images+"*/")): video_proposals = [] templates = [] for image_fn in sorted(glob.glob(video_fn+"*")): ann_fn = image_fn.replace(input_images,first_frame_anns).replace('.jpg','.png') if glob.glob(ann_fn): new_templates = read_ann(ann_fn) new_templates = add_ReID(new_templates, image_fn, ReID_net) # import json # ff_fn = image_fn.replace(input_images, "DAVIS/ff_test/").replace('.jpg', '.json') # with open(ff_fn, "r") as f: # new_templates = json.load(f) # for id, templ in enumerate(new_templates): # templ['ReID'] = np.array(templ['ReID']) # templ['id'] = id templates = templates + new_templates prop_fn = image_fn.replace(input_images,input_proposals).replace('.jpg','.json') proposals = read_props(prop_fn) video_proposals.append(proposals) ReIDs = [[prop['ReID'] for prop in props] for props in video_proposals] template_ReIDs = [templ['ReID'] for templ in templates] all_reid_distances = [np.array([[norm(c_reid - gt_reid) for c_reid in curr] for gt_reid in template_ReIDs]) for curr in ReIDs] all_reid_distances_no_inf = copy(all_reid_distances) for mat in all_reid_distances_no_inf: mat[np.isinf(mat)] = 0 max_distances = np.array([mat.max(axis=1) if mat.shape[1]>0 else np.zeros((mat.shape[0])) for mat in all_reid_distances_no_inf]).max(axis=0) print(max_distances) dataset_max_distances.append(max_distances.max()) print(np.array(dataset_max_distances).max())
38.25
142
0.723479
0
0
0
0
0
0
0
0
400
0.201106
4b7d11da5ac6e1b0ebc4170e7d035cb4092ec2fa
1,377
py
Python
algorithms/tests/test_string_matching.py
t3rm1n4l/python-algorithms
0fbcb38b26d8690028cd5a676743950fdf3a060f
[ "MIT" ]
1
2018-05-02T07:37:43.000Z
2018-05-02T07:37:43.000Z
algorithms/tests/test_string_matching.py
t3rm1n4l/python-algorithms
0fbcb38b26d8690028cd5a676743950fdf3a060f
[ "MIT" ]
null
null
null
algorithms/tests/test_string_matching.py
t3rm1n4l/python-algorithms
0fbcb38b26d8690028cd5a676743950fdf3a060f
[ "MIT" ]
null
null
null
import unittest import string_matching class StringMatchingTest(unittest.TestCase): def test_string_matching_naive(self): t = 'ababbababa' s = 'aba' self.assertEquals(string_matching.string_matching_naive(t, s), [0, 5, 7]) t = 'ababbababa' s = 'abbb' self.assertEquals(string_matching.string_matching_naive(t, s), []) def test_string_matching_rabin_karp(self): t = 'ababbababa' s = 'aba' self.assertEquals(string_matching.string_matching_rabin_karp(t, s), [0, 5, 7]) t = 'ababbababa' s = 'abbb' self.assertEquals(string_matching.string_matching_rabin_karp(t, s), []) def test_string_matching_knuth_morris_pratt(self): t = 'ababbababa' s = 'aba' self.assertEquals(string_matching.string_matching_knuth_morris_pratt(t, s), [0, 5, 7]) t = 'ababbababa' s = 'abbb' self.assertEquals(string_matching.string_matching_knuth_morris_pratt(t, s), []) def test_string_matching_boyer_moore_horspool(self): t = 'ababbababa' s = 'aba' self.assertEquals(string_matching.string_matching_boyer_moore_horspool(t, s), [0, 5, 7]) t = 'ababbababa' s = 'abbb' self.assertEquals(string_matching.string_matching_boyer_moore_horspool(t, s), []) if __name__ == '__main__': unittest.main()
33.585366
96
0.655773
1,287
0.934641
0
0
0
0
0
0
150
0.108932
4b7d6c918015930582e1fb1d514d24f1d777be05
1,411
py
Python
molecool_test/tests/test_molecule.py
radifar/molecool_test
9e0027656d6f68d2efd9cdf8f24872b4bcea6cb9
[ "BSD-3-Clause" ]
null
null
null
molecool_test/tests/test_molecule.py
radifar/molecool_test
9e0027656d6f68d2efd9cdf8f24872b4bcea6cb9
[ "BSD-3-Clause" ]
null
null
null
molecool_test/tests/test_molecule.py
radifar/molecool_test
9e0027656d6f68d2efd9cdf8f24872b4bcea6cb9
[ "BSD-3-Clause" ]
null
null
null
import numpy as np import pytest import molecool_test @pytest.fixture def methane_molecule(): symbols = np.array(['C', 'H', 'H', 'H', 'H']) coordinates = np.array([ [1, 1, 1], [2.4, 1, 1], [-0.4, 1, 1], [1, 1, 2.4], [1, 1, -0.4], ]) return symbols, coordinates def test_move_methane(methane_molecule): symbols, coordinates = methane_molecule coordinates[0] += 5 def test_build_bond_list(methane_molecule): symbols, coordinates = methane_molecule bonds = molecool_test.build_bond_list(coordinates) assert len(bonds) == 4 for bond_length in bonds.values(): assert bond_length == 1.4 def test_build_bond_failure(methane_molecule): symbols, coordinates = methane_molecule with pytest.raises(ValueError): bonds = molecool_test.build_bond_list(coordinates, min_bond=-1) def test_molecular_mass(methane_molecule): symbols, coordinates = methane_molecule calculated_mass = molecool_test.calculate_molecular_mass(symbols) actual_mass = 16.04 assert pytest.approx(actual_mass, abs=1e-2) == calculated_mass def test_center_of_mass(methane_molecule): symbols, coordinates = methane_molecule center_of_mass = molecool_test.calculate_center_of_mass(symbols, coordinates) expected_center = np.array([1,1,1]) assert np.array_equal(center_of_mass, expected_center)
21.707692
81
0.697378
0
0
0
0
263
0.186393
0
0
15
0.010631
4b7e597bab0f3442569b2c0f944ee9a51ebdc5c8
5,004
py
Python
tests/unit/html/test_search_page.py
tttgm/basketball_reference_web_scraper
2dbd9d7bacbcfee17f08bcf8629bd7d50893761d
[ "MIT" ]
325
2015-10-27T03:15:49.000Z
2022-03-16T06:49:12.000Z
tests/unit/html/test_search_page.py
tttgm/basketball_reference_web_scraper
2dbd9d7bacbcfee17f08bcf8629bd7d50893761d
[ "MIT" ]
173
2018-10-16T04:11:05.000Z
2022-03-29T17:52:08.000Z
tests/unit/html/test_search_page.py
tttgm/basketball_reference_web_scraper
2dbd9d7bacbcfee17f08bcf8629bd7d50893761d
[ "MIT" ]
97
2016-04-09T19:11:28.000Z
2022-03-21T09:57:50.000Z
from unittest import TestCase from unittest.mock import patch, MagicMock, PropertyMock from basketball_reference_web_scraper.html import SearchPage, PlayerSearchResult class TestSearchPage(TestCase): def test_nba_aba_baa_players_content_query(self): self.assertEqual( SearchPage(html=MagicMock()).nba_aba_baa_players_content_query, '//div[@id="searches"]/div[@id="players"]', ) @patch.object(SearchPage, 'nba_aba_baa_players_content_query', new_callable=PropertyMock) def test_nba_aba_baa_players_pagination_links_query(self, mocked_query): mocked_query.return_value = "some query" self.assertEqual( SearchPage(html=MagicMock()).nba_aba_baa_players_pagination_links_query, 'some query/div[@class="search-pagination"]/a', ) @patch.object(SearchPage, 'nba_aba_baa_players_content_query', new_callable=PropertyMock) def test_nba_aba_baa_player_search_items_query(self, mocked_query): mocked_query.return_value = "some query" self.assertEqual( SearchPage(html=MagicMock()).nba_aba_baa_player_search_items_query, 'some query/div[@class="search-item"]', ) @patch.object(SearchPage, 'nba_aba_baa_players_pagination_links_query', new_callable=PropertyMock) def test_nba_aba_baa_players_pagination_links(self, mocked_query): mocked_query.return_value = "some query" html = MagicMock() links = [MagicMock(return_value="some"), MagicMock(return_value="links")] html.xpath = MagicMock(return_value=links) self.assertEqual( SearchPage(html=html).nba_aba_baa_players_pagination_links, links, ) html.xpath.asset_called_once_with("some query") @patch.object(SearchPage, 'nba_aba_baa_players_pagination_links', new_callable=PropertyMock) def test_nba_aba_baa_players_pagination_url_is_none_when_no_pagination_links(self, mocked_links): mocked_links.return_value = [] self.assertIsNone(SearchPage(html=MagicMock()).nba_aba_baa_players_pagination_url) @patch.object(SearchPage, 'nba_aba_baa_players_pagination_links', new_callable=PropertyMock) def test_nba_aba_baa_players_pagination_url_is_first_link_href_attrib_when_single_link_is_not_at_end_of_results( self, mocked_links ): link = MagicMock() link.text_content = MagicMock(return_value="jaebaebae") link.attrib = MagicMock() link.attrib.__getitem__ = MagicMock(return_value="some text content") mocked_links.return_value = [link] self.assertEqual( SearchPage(html=MagicMock()).nba_aba_baa_players_pagination_url, "some text content", ) link.attrib.__getitem__.assert_called_once_with("href") @patch.object(SearchPage, 'nba_aba_baa_players_pagination_links', new_callable=PropertyMock) def test_nba_aba_baa_players_pagination_url_is_none_when_single_link_is_at_end_of_results( self, mocked_links ): link = MagicMock() link.text_content = MagicMock(return_value="Previous 100 Results") mocked_links.return_value = [link] self.assertIsNone(SearchPage(html=MagicMock()).nba_aba_baa_players_pagination_url) link.text_content.assert_called_once_with() @patch.object(SearchPage, 'nba_aba_baa_players_pagination_links', new_callable=PropertyMock) def test_nba_aba_baa_players_pagination_url_is_second_link_href_attrib_when_multiple_links( self, mocked_links ): first_link = MagicMock() first_link.attrib = MagicMock() first_link.attrib.__getitem__ = MagicMock(return_value="some text content") second_link = MagicMock() second_link.attrib = MagicMock() second_link.attrib.__getitem__ = MagicMock(return_value="some other text content") mocked_links.return_value = [first_link, second_link] self.assertEqual( SearchPage(html=MagicMock()).nba_aba_baa_players_pagination_url, "some other text content", ) second_link.attrib.__getitem__.assert_called_once_with("href") @patch.object(SearchPage, 'nba_aba_baa_player_search_items_query', new_callable=PropertyMock) def test_nba_aba_baa_players(self, mocked_query): mocked_query.return_value = "some query" first_result = MagicMock(name="first html result") second_result = MagicMock(name="second html result") third_result = MagicMock(name="third html result") html = MagicMock() html.xpath = MagicMock(return_value=[first_result, second_result, third_result]) self.assertEqual( SearchPage(html=html).nba_aba_baa_players, [ PlayerSearchResult(html=first_result), PlayerSearchResult(html=second_result), PlayerSearchResult(html=third_result), ] )
42.40678
116
0.711631
4,832
0.965627
0
0
4,531
0.905476
0
0
714
0.142686
4b7fad07fb9954bb150ff9b9a3fc6a0e8f2cf560
19,891
py
Python
cave/com.raytheon.viz.gfe/localization/gfe/userPython/smartTools/WindGustFromAlgorithm.py
srcarter3/awips2
37f31f5e88516b9fd576eaa49d43bfb762e1d174
[ "Apache-2.0" ]
null
null
null
cave/com.raytheon.viz.gfe/localization/gfe/userPython/smartTools/WindGustFromAlgorithm.py
srcarter3/awips2
37f31f5e88516b9fd576eaa49d43bfb762e1d174
[ "Apache-2.0" ]
null
null
null
cave/com.raytheon.viz.gfe/localization/gfe/userPython/smartTools/WindGustFromAlgorithm.py
srcarter3/awips2
37f31f5e88516b9fd576eaa49d43bfb762e1d174
[ "Apache-2.0" ]
1
2021-10-30T00:03:05.000Z
2021-10-30T00:03:05.000Z
## # This software was developed and / or modified by Raytheon Company, # pursuant to Contract DG133W-05-CQ-1067 with the US Government. # # U.S. EXPORT CONTROLLED TECHNICAL DATA # This software product contains export-restricted data whose # export/transfer/disclosure is restricted by U.S. law. Dissemination # to non-U.S. persons whether in the United States or abroad requires # an export license or other authorization. # # Contractor Name: Raytheon Company # Contractor Address: 6825 Pine Street, Suite 340 # Mail Stop B8 # Omaha, NE 68106 # 402.291.0100 # # See the AWIPS II Master Rights File ("Master Rights File.pdf") for # further licensing information. ## # ---------------------------------------------------------------------------- # This software is in the public domain, furnished "as is", without technical # support, and with no warranty, express or implied, as to its usefulness for # any purpose. # # New_WindGust_Tool # # Authors: Tom Mazza NWS Charleston, WV Created: 04/25/03 # Matthew H. Belk NWS Taunton, MA Last Modified: 06/16/03 # Mathewson FSL Modified: 3/30/04 # -change in model names to OB3 names #---------------------------------------------------------------------------- # # SOFTWARE HISTORY # # Date Ticket# Engineer Description # ------------ ---------- ----------- -------------------------- # 02/10/2016 5283 nabowle Remove NGM support. # ---------------------------------------------------------------------------- ## # This is an absolute override file, indicating that a higher priority version # of the file will completely replace a lower priority version of the file. ## ToolType = "numeric" WeatherElementEdited = "WindGust" from numpy import * # without this, the builtin max() is used from numpy import max import LogStream # You can screen the elements for which your tool will appear by using # a ScreenList. For example: #ScreenList = ["MixHgt","WindGust", "TransWind"] # Set up variables to be solicited from the user: VariableList = [ ("Momentum algorithm:", "RUC", "radio", ["RUC", "Power"]), ("Use BL Winds:", "No", "radio", ["Yes", "No"]), ("Model:", "NAM12", "radio", ["GFS80", "NAM12", "gfsLR", "RAP40"]) ] #Set up Class import SmartScript ## For available commands, see SmartScript toolName = 'WindGustFromAlgorithm' class Tool (SmartScript.SmartScript): def __init__(self, dbss): SmartScript.SmartScript.__init__(self, dbss) # Define your site ID self._SITEID = "BOX" # Required Method: Execute # Called once for each grid # Fill in the arguments you want to use -- WeatherElement1, WeatherElement2... def execute(self, Wind, MixHgt, Topo, GridTimeRange): "Determines WindGust using one of two algorithms, one from the RUC or a power relationship. This tool assumes your mixing height has already been adjusted for your surface temperatures." sounding = self.makeNumericSounding(self._model, "wind", self._modelCube, GridTimeRange, noDataError=0) ######################################################################## # If we don't have a model sounding at this point in time, or the # size of the grids do not match if sounding is None: # or sounding[0].shape != Topo.shape: LogStream.logProblem(toolName, ': cannot obtain a Wind sounding') return None # leaves current WindGust grid alone ######################################################################## # If we made it this far, split up the sounding into its component # cubes of height and wind (gh_Cube, wind_Cube) = sounding if gh_Cube is None: LogStream.logProblem(toolName, 'gh_Cube is None') return None if wind_Cube is None: LogStream.logProblem(toolName, 'wind_Cube is None') return None ######################################################################## # Convert topography from feet to meters self._topo = self.ftToM(Topo) ######################################################################## # Initialize a cube to hold BL wind grids bl_WindCube = {} ######################################################################## # Cycle through all the BL levels we have for this model for lvl in self._blCube: #################################################################### # Initialize BL wind grid for this level grid = None #################################################################### # If this is the NAM40/20 model if self._model.find('NAM40') != -1: ################################################################ # Get BL winds from other NAM40/NAM20 file tempModel = self._model.replace('NAM40', 'NAM20') ################################################################ # Try to get model BL winds for this time grid = self.getGrids(tempModel, "wind", lvl, GridTimeRange, noDataError=0) #################################################################### # Otherwise else: ################################################################ # Try to get model BL winds for this time grid = self.getGrids(self._model, "Wind", lvl, GridTimeRange, noDataError=0) #################################################################### # Add this grid to the BL wind cube - if it is valid if grid != None: ################################################################ # Store the wind speeds at this BL level bl_WindCube[lvl] = grid[0] #################################################################### # Otherwise else: ################################################################ # Store a placeholder bl_WindCube[lvl] = None ######################################################################## # Convert mixing height from ft ASL to m ASL mixHgt_m = self.ftToM(MixHgt) ######################################################################## # Make a 3D mask where the model sounding level is ABOVE the ground, # but below the Mixing Height self._mixedLayer = (gh_Cube >= self._topo) & (gh_Cube <= mixHgt_m) ######################################################################## # Method to compute WindGust using a version of the RUC technique # adapted by Matthew H. Belk (BOX). ######################################################################## # Initialize WindGust using current 10m Wind speeds - (mag, dir) WindGust = Wind[0] ######################################################################## # Move vertically through the model BL cube for lvl in self._blCube: #################################################################### # Make a mask where this BL surface is at or below the MixHgt blMask = MixHgt <= self._blHgt[lvl] #################################################################### # If there are any points in the mixed layer at this surface, and # there actually is a wind grid if any(blMask) and bl_WindCube[lvl] != None: ################################################################ # Get wind magnitude at current level - remember model winds # are in m/s and need to be in kts for comparison curMag = self.mpsToKt(bl_WindCube[lvl]) ################################################################ # Compute difference between wind at this level and SFC wind # where points are in the mixed layer deltaSpd = curMag - Wind[0] ################################################################ # Get the depth of the mixed layer to this point (m AGL) deltaZ = self._blHgt[lvl] ################################################################ # Adjust change in wind speed by a coefficient - using the # lesser of 0.5 or (deltaZ / 2000) # First get the factor, which will range from 0.5 to 1.0, # higher closer to the ground delta = max(1.0 - deltaZ/2000.0, 0.5) ################################################################ # Employ the power relationship if selected: it focuses in on # how much lower than one this factor will be (it ranges from # no less than 1 just above the surface to 0.5 lower than 1 # 1000 or more feet from the surface). The power relationship # takes this small number (between 0 and 0.5) to the second # power, which makes it smaller still. It actually first # doubles it, then squares it, then halves it again. This # causes a difference of 0 to stay 0, a difference of 0.5 to # stay at 0.5, but a difference of 0.25 will become 0.125. # This difference is then subtracted from one, to get a new, # equal or larger factor by which to multiply the potential # wind gust, to arrive at a gust potential that decreases more # slowly at first with height, then more rapidly later on, to # arrive at the same factor up at 1000 m and more above the # surface. The resulting wind gust is always equal to or # greater than using the RUC algorthm straight up. if self._algorithm == 'Power': delta = 1 - (pow((2 * (1 - delta)), 2)) / 2 ################################################################ # Adjust wind speed difference by chosen coefficient deltaSpd *= delta gustV = Wind[0] + deltaSpd ################################################################ # Make a mask where this WindGust is > current WindGust newGust = gustV > WindGust ################################################################ # Assign new WindGust where new WindGust is greater and the # surface is still within the mixed layer WindGustMask = newGust & blMask WindGust[WindGustMask] = gustV[WindGustMask] ######################################################################## # Move vertically through the model cube for i in xrange(gh_Cube.shape[0]): #################################################################### # If there are any points in the mixed layer at this surface if any(self._mixedLayer[i]): ################################################################ # Get wind magnitude at current level - remember model winds # are in m/s and need to be in kts for comparison curMag = self.mpsToKt(wind_Cube[0][i]) ################################################################ # Compute difference between wind at this level and SFC wind # where points are in the mixed layer deltaSpd = curMag - Wind[0] ################################################################ # Get the depth of the mixed layer to this point (m AGL) deltaZ = gh_Cube[i] - self._topo ################################################################ # Adjust change in wind speed by a coefficient - using the # lesser of 0.5 or (deltaZ / 2000) # First get the factor, which will range from 0.5 to 1.0, # higher closer to the ground delta = max(1.0-deltaZ/2000.0,0.5) ################################################################ # Employ the power relationship if selected: it focuses in on # how much lower than one this factor will be (it ranges from # no less than 1 just above the surface to 0.5 lower than 1 # 1000 or more feet from the surface). The power relationship # takes this small number (between 0 and 0.5) to the second # power, which makes it smaller still. It actually first # doubles it, then squares it, then halves it again. This # causes a difference of 0 to stay 0, a difference of 0.5 to # stay at 0.5, but a difference of 0.25 will become 0.125. # This difference is then subtracted from one, to get a new, # equal or larger factor by which to multiply the potential # wind gust, to arrive at a gust potential that decreases more # slowly at first with height, then more rapidly later on, to # arrive at the same factor up at 1000 feet and more above the # surface. The resulting wind gust is always equal to or # greater than using the RUC algorthm straight up. if self._algorithm == 'Power': delta = 1 - (pow((2 * (1 - delta)), 2)) / 2 ################################################################ # Adjust wind speed difference by chosen coefficient deltaSpd *= delta gustV = Wind[0] + deltaSpd ################################################################ # Make a mask where this WindGust is > current WindGust newGust = gustV > WindGust ################################################################ # Assign new WindGust where new WindGust is greater and the # surface is still within the mixed layer WindGustMask = newGust & self._mixedLayer[i] WindGust[WindGustMask] = gustV[WindGustMask] ######################################################################## # Return the computed WindGust return WindGust # Optional Methods # These methods can have the additional argument: # ToolTimeRange -- selected time range over which we are running the tool def preProcessTool(self, varDict): # Called once at beginning of Tool # Cannot have WeatherElement or Grid arguments ######################################################################## # Get site ID try: siteID=self.mutableID().siteID() except: siteID=self._SITEID ######################################################################## # Get name of chosen model - and fix it up so we can use it later on. # This will grab the latest version of the chosen model from the D2D # netCDF files. self._model = "%s_D2D_%s" % (siteID, varDict["Model:"]) ######################################################################## # Get chosen algorithm self._algorithm = varDict["Momentum algorithm:"] ######################################################################## # Get answer if we should use BL winds useBLwinds = varDict["Use BL Winds:"] ######################################################################## # Initialize a list of model levels self._modelCube = [] ######################################################################## # Determine model levels available for each model if self._model.find( 'GFS80') != -1 or \ self._model.find( 'GFS') != -1: self._modelCube = ["MB850", "MB700", "MB500", "MB400", "MB300"] self._blCube = [] elif self._model.find( 'NAM12') != -1: self._modelCube = ["MB1000", "MB950", "MB900", "MB850", "MB800", "MB750", "MB700", "MB650", "MB600", "MB550", "MB500", "MB450", "MB400", "MB350"] self._blCube = ["BL030", "BL03060", "BL6090", "BL90120", "BL12015"] elif self._model.find( 'NAM40') != -1 or \ self._model.find( 'NAM20') != -1: self._modelCube = ["MB975", "MB950", "MB925", "MB900", "MB875", "MB850", "MB825", "MB800", "MB775", "MB750", "MB725", "MB700", "MB675", "MB650", "MB625", "MB600", "MB550", "MB500", "MB450", "MB400", "MB350", "MB300"] self._blCube = ["BL030", "BL03060", "BL6090", "BL90120", "BL120150"] elif self._model.find( 'gfsLR') != -1: self._modelCube = ["MB1000", "MB850", "MB700", "MB500", "MB300"] self._blCube = [] elif self._model.find( 'RAP40') != -1: self._modelCube = ["MB1000", "MB950", "MB900", "MB850", "MB800", "MB750", "MB700", "MB650", "MB600", "MB550", "MB500", "MB450", "MB400", "MB350", "MB300"] self._blCube = ["BL030", "BL6090", "BL15018"] ######################################################################## # If we should not use the BL winds if useBLwinds is 'No': #################################################################### # Reset the levels in the BL cube so we don't do anything self._blCube = [] ######################################################################## # Determine height of all possible BL levels available for each model. # If level is not at a fixed height AGL, use the hydrostatic equation. # Assume the density of the air is 1 kg/m3 and gravity is 9.80 m/s^2. # The height will be in m AGL at the center of the layer. Remember # there are 100 Pa per 1 mb. self._blHgt = {'BL030' : (15.0 * 100.0/ 9.8), 'BL3060' : (45.0 * 100.0 / 9.8), 'BL03060' : (45.0 * 100.0 / 9.8), 'BL6090' : (75.0 * 100.0 / 9.8), 'BL90120' : (105.0 * 100.0 / 9.8), 'BL12015' : (135.0 * 100.0 / 9.8), 'BL120150': (135.0 * 100.0 / 9.8), 'BL15018' : (165.0 * 100.0 / 9.8), 'FH1829' : 1829.0, 'FH2743' : 2743.0, 'FH3658' : 3658.0 } LogStream.logDebug(toolName, ': preProcessTool complete.')
48.045894
195
0.42934
17,304
0.869941
0
0
0
0
0
0
12,495
0.628174
4b7fc93c2e30ca54b02519e2a781a191d7e736a1
6,705
py
Python
pochta/tracking.py
john-phonk/fs-pochta-api
c3b7df4ecdbfc45fb482cedd8ab6c2927e0a1c9d
[ "MIT" ]
16
2019-05-13T01:12:10.000Z
2022-01-17T06:21:35.000Z
pochta/tracking.py
john-phonk/fs-pochta-api
c3b7df4ecdbfc45fb482cedd8ab6c2927e0a1c9d
[ "MIT" ]
4
2020-03-06T06:46:35.000Z
2020-11-22T04:24:34.000Z
pochta/tracking.py
john-phonk/fs-pochta-api
c3b7df4ecdbfc45fb482cedd8ab6c2927e0a1c9d
[ "MIT" ]
6
2019-08-10T13:18:21.000Z
2021-11-25T08:57:30.000Z
from abc import ABC from typing import List from zeep import CachingClient, Client, Settings from .exceptions import APIError class _BaseClient(ABC): """API клиент сервиса отслеживания посылок. https://tracking.pochta.ru/specification """ WSDL = '' def __init__(self, login: str, password: str, caching=True): """Инициализация API клиента сервиса отслеживания посылок. :param login: Логин от системы трекинга :param password: Пароль от системы трекинга :param caching: Флаг, позволяющий отключить кэширование в zeep """ self._login = login self._password = password zeep_client = CachingClient if caching else Client self._client = zeep_client( self.WSDL, settings=Settings(strict=False), ) class SingleTracker(_BaseClient): """Клиент для взаимодеействия с API единичной обработки запросов.""" WSDL = 'https://tracking.russianpost.ru/rtm34?wsdl' def get_history(self, barcode: str) -> dict: """ История операций над отправлением. Метод getOperationHistory используется для получения информации о конкретном отправлении. Метод возвращает подробную информацию по всем операциям, совершенным над отправлением. https://tracking.pochta.ru/specification#getOperationHistory :param barcode: Идентификатор регистрируемого почтового отправления в одном из форматов: - внутрироссийский, состоящий из 14 символов (цифровой) - международный, состоящий из 13 символов (буквенно-цифровой) в формате S10. :return: Ответ метода getOperationHistory содержит список элементов historyRecord. Каждый из них содержит информацию об одной операции над отправлением. Если над отправлением еще не зарегистрировано ни одной операции, то возвращается пустой список элементов historyRecord. """ return self._client.service.getOperationHistory( OperationHistoryRequest={ 'Barcode': barcode, 'MessageType': '0' }, AuthorizationHeader={ 'login': self._login, 'password': self._password, }, ) def get_order_events_for_mail(self, barcode: str) -> dict: """ История операций с наложенным платежом. Метод PostalOrderEventsForMail позволяет получить информацию об операциях с наложенным платежом, который связан с конкретным почтовым отправлением. https://tracking.pochta.ru/specification#PostalOrderEventsForMail :param barcode: Идентификатор регистрируемого почтового отправления в одном из форматов: - внутрироссийский, состоящий из 14 символов (цифровой); - международный, состоящий из 13 символов (буквенно-цифровой) в формате S10. :return: Список событий """ return self._client.service.PostalOrderEventsForMail( PostalOrderEventsForMailInput={ 'Barcode': barcode, }, AuthorizationHeader={ 'login': self._login, 'password': self._password, }, ) class BatchTracker(_BaseClient): """Клиент для взаимодеействия с API пакетной обработки запросов.""" WSDL = 'https://tracking.russianpost.ru/fc?wsdl' def get_ticket(self, barcodes: List[str]) -> str: """Получения билета на подготовку информации по списку идентификаторов отправлений. Метод getTicket используется для получения билета на подготовку информации по списку идентификаторов отправлений. В запросе передается список идентификаторов отправлений. При успешном вызове метод возвращает идентификатор билета. Ограничения и рекомендации по использованию: - Количество идентификаторов отправлений в одном запросе не должно превышать *3000*. - Рекомендуется выполнять первое обращение за ответом по билету не ранее, чем через 15 минут от момента выдачи билета. - В случае неготовности результата повторные обращения по тому же билету следует выполнять не чаще, чем 1 раз в 15 минут - Время хранения ответа по билету в Сервисе отслеживания составляет 32 часа. По истечении этого периода ответ удаляется. https://tracking.pochta.ru/specification раздел "Пакетная обработка" п.3 :param barcodes: Идентификаторы регистрируемых почтовогых отправлений в одном из форматов: - внутрироссийский, состоящий из 14 символов (цифровой) - международный, состоящий из 13 символов (буквенно-цифровой) в формате S10. :return: Ответ метода getTicket содержит информацию о выданном билете в объекте ticketResponse в случае успешного запроса, функция возвращает номер созданного ticket, полученного из ticketResponse.value """ # По умолчанию zeep генерирует Request старой версии, # где запрос отправляется в виде файла с метаданными # Поэтому, вручную создаём объект Request и убираем аттрибуты, относящиеся к файлу request = self._client.get_type('{http://fclient.russianpost.org}file') request.attributes.clear() items = [{'Barcode': barcode} for barcode in barcodes] response = self._client.service.getTicket( request=request(Item=items), login=self._login, password=self._password, language='RUS', ) if response['error'] is not None: raise APIError(f'Response body contains error: {response["error"]}') return response['value'] def get_response_by_ticket(self, ticket: str) -> List[dict]: """Метод используется для получения информации об отправлениях по ранее полученному билету. Вызывает метод answerByTicketRequest используемый для получения информации об отправлениях по ранее полученному билету. https://tracking.pochta.ru/specification раздел "Пакетная обработка" п.4 :param ticket: Строка, содержащая номер ticket, полученного ранее при вызове getTicket :return: Результаты пакетной обработки в виде списка словарей, содержащих результаты выполнения запроса на пакетную обработку """ response = self._client.service.getResponseByTicket( ticket=ticket, login=self._login, password=self._password, ) if response['error'] is not None: raise APIError(f'Response body contains error: {response["error"]}') return response['value']['Item']
40.149701
99
0.670097
9,091
0.985154
0
0
0
0
0
0
7,109
0.770373
4b7fd5f816b4e255d1e40adf591dc8b3e21efaa2
2,291
py
Python
CH04_Iterators_and_Generators/4.4.Implementing_the_iterator_protocol.py
Chang-Liu-TAMU/Python-Cookbook-reading
7b974c32f77b4b3d7cfeed30d1671081057c566f
[ "MIT" ]
null
null
null
CH04_Iterators_and_Generators/4.4.Implementing_the_iterator_protocol.py
Chang-Liu-TAMU/Python-Cookbook-reading
7b974c32f77b4b3d7cfeed30d1671081057c566f
[ "MIT" ]
null
null
null
CH04_Iterators_and_Generators/4.4.Implementing_the_iterator_protocol.py
Chang-Liu-TAMU/Python-Cookbook-reading
7b974c32f77b4b3d7cfeed30d1671081057c566f
[ "MIT" ]
null
null
null
# @Time: 2022/4/12 20:50 # @Author: chang liu # @Email: [email protected] # @File:4.4.Implementing_the_iterator_protocol.py ################ clean version ######################### # class Node: # def __init__(self, val): # self._value = val # self._children = [] # # def __repr__(self): # return "Node({!r})".format(self._value) # # def add_child(self, node): # self._children.append(node) # # def __iter__(self): # return iter(self._children) # # def depth_first(self): # yield self # for c in self: # yield from c.depth_first() ############# some messy version #################### class Node: def __init__(self, value): self._value = value self._children = [] def __repr__(self): return "Node({!r})".format(self._value) def add_child(self, node): self._children.append(node) def __iter__(self): return iter(self._children) # def iter(self): # return iter(self._children) def depth_first(self): return DepthFirstIterator(self) # def __iter__(self): # return DepthFirstIterator(self) class DepthFirstIterator: ''' DFS traversal ''' def __init__(self, start_node): self._node = start_node self._children_iter = None self._child_iter = None def __iter__(self): return self def __next__(self): if self._children_iter == None: self._children_iter = iter(self._node) # self._children_iter = self._node.iter() return self._node elif self._child_iter: try: following = next(self._child_iter) return following except StopIteration: self._child_iter = None return next(self) else: self._child_iter = next(self._children_iter).depth_first() return next(self) # return next(self._child_iter) root = Node(0) left = Node(1) right = Node(2) left.add_child(Node(3)) left.add_child(Node(4)) right.add_child(Node(5)) right.add_child(Node(6)) root.add_child(left) root.add_child(right) for i in root.depth_first(): print(i) # for i in root: # print(i)
22.91
70
0.572676
1,340
0.584897
0
0
0
0
0
0
907
0.395897
4b800dc76b871db39c746e292171f32b25ee44ff
29,762
py
Python
FGPVAE_model.py
metodj/FGP-VAE
607559ab465b29878f10a5d95b8e3c6ec8d94e0c
[ "MIT" ]
3
2021-01-27T14:06:01.000Z
2021-09-09T12:10:34.000Z
FGPVAE_model.py
metodj/FGP-VAE
607559ab465b29878f10a5d95b8e3c6ec8d94e0c
[ "MIT" ]
null
null
null
FGPVAE_model.py
metodj/FGP-VAE
607559ab465b29878f10a5d95b8e3c6ec8d94e0c
[ "MIT" ]
null
null
null
import tensorflow as tf import tensorflow_probability as tfp import numpy as np import pickle import random from utils import gauss_cross_entropy tfk = tfp.math.psd_kernels def _add_diagonal_jitter(matrix, jitter=1e-6): return tf.linalg.set_diag(matrix, tf.linalg.diag_part(matrix) + jitter) class FGP: dtype = np.float64 def __init__(self, init_amplitude, init_length_scale, GP_joint, L_w, object_vectors_init=None, object_prior_corr=False, K_obj_normalize=False): """ GP class for FGPVAE. :param init_amplitude: :param init_length_scale: :param GP_joint: :param L_w: number of local latent channels :param object_vectors_init: initizalition for object vectors (GP-LVM) :param object_prior_corr: whether or not correlated object priors are used :param K_obj_normalize: whether or not to normalize object kernel (linear kernel) """ self.object_prior_corr = object_prior_corr self.K_obj_normalize = K_obj_normalize if GP_joint: self.amplitude = tf.Variable(initial_value=init_amplitude, name="GP_amplitude", trainable=True, dtype=self.dtype) self.length_scale = tf.Variable(initial_value=init_length_scale, name="GP_length_scale", trainable=True, dtype=self.dtype) else: self.amplitude = tf.constant(init_amplitude, dtype=self.dtype) self.length_scale = tf.constant(init_length_scale, dtype=self.dtype) # kernels self.kernel_local = tfk.ExpSinSquared(amplitude=self.amplitude, length_scale=self.length_scale, period=2*np.pi) self.kernel_global = tfk.Linear() # GP-LVM, object vectors if object_vectors_init is not None: self.object_vectors = tf.Variable(initial_value=object_vectors_init, name="GP_LVM_object_vectors", dtype=self.dtype) else: self.object_vectors = None # number of local (views/angles) channels self.L_w = L_w def build_1d_gp_local(self, X, Y, varY, X_test): """ Fits GP for local latent channels. Takes input-output dataset and returns post mean, var, marginal lhood. This is standard GP regression with heteroscedastic noise. :param X: inputs tensor (batch, npoints) :param Y: outputs tensor (batch, npoints) :param varY: outputs tensor (batch, npoints) :param X_test: (batch, ns) input points to compute post mean + var Returns: p_m: (batch, ns) post mean at X_test p_v: (batch, ns) post var at X_test logZ: (batch) marginal lhood of each dataset in batch """ # Prepare all constants batch = tf.shape(X)[0] n = tf.shape(X)[1] ns = tf.shape(X_test)[1] # K_x + \sigma_x^* K = self.kernel_local.matrix(tf.expand_dims(X, 2), tf.expand_dims(X, 2)) # (batch, n n) K = K + tf.matrix_diag(varY) # (batch, n, n) chol_K = tf.linalg.cholesky(K) # (batch, n, n) # lhood term 1/3 lhood_pi_term = tf.cast(n, dtype=self.dtype) * np.log(2 * np.pi) # lhood term 2/3 lhood_logdet_term = 2 * tf.reduce_sum(tf.log(tf.matrix_diag_part(chol_K)), 1) # (batch) # lhood term 3/3 Y = tf.expand_dims(Y, 2) iKY = tf.cholesky_solve(chol_K, Y) # (batch, n, 1) lh_quad_term = tf.matmul(tf.transpose(Y, (0,2,1)), iKY) # (batch, 1, 1) lh_quad_term = tf.reshape(lh_quad_term, [batch]) # log P(Y|X) = -1/2 * ( n log(2 pi) + Y inv(K+noise) Y + log det(K+noise)) gp_lhood = -0.5 * (lhood_pi_term + lh_quad_term + lhood_logdet_term) # Compute posterior mean and variances Ks = self.kernel_local.matrix(tf.expand_dims(X, 2), tf.expand_dims(X_test, 2)) # (batch, n, ns) Ks_t = tf.transpose(Ks, (0, 2, 1)) # (batch, ns, n) # posterior mean p_m = tf.matmul(Ks_t, iKY) p_m = tf.reshape(p_m, (batch, ns)) # posterior variance iK_Ks = tf.cholesky_solve(chol_K, Ks) # (batch, n, ns) Ks_iK_Ks = tf.reduce_sum(Ks * iK_Ks, axis=1) # (batch, ns) p_v = 1 - Ks_iK_Ks # (batch, ns) p_v = tf.reshape(p_v, (batch, ns)) return p_m, p_v, gp_lhood, K def build_1d_gp_global(self, means, vars): """ Fits GP for global latent channels. :param Y: encoder means (batch, npoints) :param varY: encoder vars (batch, npoints) Returns: p_m: (batch) posterior means p_v: (batch) post vars logZ: (batch) product of Gaussians terms """ n = tf.shape(means)[1] sigma_squared_bar = 1 / (tf.reduce_sum(tf.math.reciprocal_no_nan(vars), axis=1) + 1) mu_bar = sigma_squared_bar * tf.reduce_sum(means * tf.math.reciprocal_no_nan(vars), axis=1) lhood = tf.log(tf.sqrt(sigma_squared_bar)) + 0.5*tf.math.reciprocal_no_nan(sigma_squared_bar)*mu_bar**2 - \ 0.5*tf.cast(n, dtype=self.dtype)*tf.log(2.0*tf.cast(np.pi, dtype=self.dtype)) - \ tf.reduce_sum(tf.log(tf.sqrt(vars)), axis=1) - 0.5*tf.reduce_sum(tf.math.reciprocal_no_nan(vars)*means**2) return mu_bar, sigma_squared_bar, lhood @staticmethod def preprocess_1d_gp_global_correlated_object_priors(means, vars): """ Product of Gaussians for each global latent channel. See 2.9 in FGPVAE.tex N = nr. of digits N_t = nr. of angles for digit t :param means: (N, N_t) :param vars: (N, N_t) Returns: bar_means: \Bar{\mu} (1, N,) bar_vars: \Bar{\sigma}^2 (1, N,) C_tilde: \Tilde{C} (1, N,) """ N_t = tf.shape(means)[1] N_t = tf.cast(N_t, dtype=tf.float64) alpha = tf.reduce_sum(tf.math.reciprocal_no_nan(vars), axis=1) beta = tf.reduce_sum(means / vars, axis=1) bar_means = tf.expand_dims(beta / alpha, 0) # expand_dims to make it compatible with batching latter on bar_vars = tf.expand_dims(1 / alpha, 0) # expand_dims to make it compatible with batching latter on # C_1 = (2.0 * np.pi)**(-0.5 * N_t) * tf.reduce_prod(vars**(-0.5), axis=1) C_1 = (2.0 * np.pi) ** (-0.5 * N_t) * tf.reduce_prod(tf.sqrt(tf.math.reciprocal_no_nan(vars)), axis=1) C_2 = tf.exp(-0.5*tf.reduce_sum(means**2/vars, axis=1)) C_3 = tf.exp(0.5*beta**2 / alpha) C_4 = tf.sqrt(2*np.pi/alpha) C_tilde = tf.expand_dims(C_1*C_2*C_3*C_4, 0) # expand_dims to make it compatible with batching latter on # C_tilde = tf.clip_by_value(C_tilde, 1e-90, 100) bar_vars = tf.clip_by_value(bar_vars, 1e-3, 100) return bar_means, bar_vars, C_tilde def kernel_matrix_correlated_object_priors(self, x, y): """ Computes object kernel matrix in case correlated object priors are used. See 2.9 in FGPVAE.tex :param x: (1, N, 10) :param y: (1, N, 10) :param K_obj_normalized: whether or not to normalize (between -1 and 1) object kernel matrix (linear kernel) :return: object kernel matrix (1, N, N) """ # unpack auxiliary data if self.object_vectors is None: x_object, y_object =x[:, :, 2:], y[:, :, 2:] else: x_object = tf.gather(self.object_vectors, tf.cast(x[:, :, 0], dtype=tf.int64)) y_object = tf.gather(self.object_vectors, tf.cast(y[:, :, 0], dtype=tf.int64)) # compute kernel matrix object_matrix = self.kernel_global.matrix(x_object, y_object) if self.K_obj_normalize: # normalize object matrix obj_norm = 1 / tf.matmul(tf.math.reduce_euclidean_norm(x_object, axis=2, keepdims=True), tf.transpose(tf.math.reduce_euclidean_norm(y_object, axis=2, keepdims=True), perm=[0, 2, 1])) object_matrix = object_matrix * obj_norm return object_matrix def X_matrix(self, x): """ Computes X matrix. We need this function (instead of working directly with X) in order to support GP-LVM vectors joint optimization. :param x: (1, N, 10) :param normalized: whether or not to normalize object vectors (so that every object vector has norm 1) :return: """ # unpack auxiliary data if self.object_vectors is None: x_object = x[:, :, 2:] else: x_object = tf.gather(self.object_vectors, tf.cast(x[:, :, 0], dtype=tf.int64)) if self.K_obj_normalize: x_object = x_object / tf.math.reduce_euclidean_norm(x_object, axis=2, keepdims=True) return x_object def build_1d_gp_global_correlated_object_priors(self, X, Y, varY, X_test, C_tilde, omit_C_tilde, bayesian_reg_view, EPSILON=1e-6): """ See 2.9 in FGPVAE.tex Since using build_1d_gp_global_correlated_object_priors leads to numerical issues, we add support for fitting global GP using Bayesian linear regression view. :param X: auxiliary data, train points of GP (1, N, 10) :param Y: encoded and processed means for train points (1, N) :param varY: encoded and processed vars for train points (1, N) :param X_test: auxiliary data, test points of GP (1, N_s, 10) :param C_tilde: (1, N) :param omit_C_tilde: omit C_tilde from derivation and modify cross-entropy term instead :param bayesian_reg_view: whether or not to use Bayesian regression view to fit global GP. :param EPSILON: for numerical stability in log() :return: """ if bayesian_reg_view: p = 8 # dimension of object vectors N = tf.shape(X)[1] # get (and normalize) X and X_test X = self.X_matrix(X) # (1, N, p) X_T = tf.transpose(X, (0, 2, 1)) # (1, p, N) X_test = self.X_matrix(X_test) # (1, N_s, p) X_test_T = tf.transpose(X_test, (0, 2, 1)) # (1, p, N_s) # posterior params A = tf.matmul(X_T, tf.matmul(tf.linalg.diag(tf.math.reciprocal_no_nan(varY)), X)) + \ tf.expand_dims(tf.eye(p, dtype=tf.float64), 0) # (1, p, p) A_inv = tf.linalg.inv(_add_diagonal_jitter(A)) # (1, p, p) w_bar = tf.linalg.matvec(A_inv, tf.linalg.matvec(X_T, tf.math.reciprocal_no_nan(varY) * Y)) # (1, p) p_m = tf.linalg.matvec(X_test, w_bar) # (1, N) p_v = tf.linalg.diag_part(tf.matmul(X_test, tf.matmul(A_inv, X_test_T))) # (1, N) p_v = tf.clip_by_value(p_v, 1e-6, 100) # log GPML (marginal likelihood) lhood_pi_term = tf.cast(N, dtype=tf.float64) * np.log(2 * np.pi) # () mid_mat = tf.linalg.diag(varY) - tf.matmul(X, tf.matmul(A_inv, X_T)) # (1, N, N) Y_tilde = tf.math.reciprocal_no_nan(varY) * Y # (1, N) lhood_quad_term = tf.reduce_sum(Y_tilde * tf.linalg.matvec(mid_mat, Y_tilde), axis=1) # (1, ) A_chol = tf.linalg.cholesky(_add_diagonal_jitter(A)) # (1, p, p) lhood_logdet_term = tf.reduce_sum(tf.math.log(tf.math.sqrt(varY)), axis=1) + \ 2 * tf.reduce_sum(tf.log(tf.matrix_diag_part(A_chol)), axis=1) # (1, ) gp_lhood = -0.5 * (lhood_pi_term + lhood_quad_term + lhood_logdet_term) # (1, ) # add C_tilde terms if not omit_C_tilde: gp_lhood = gp_lhood + tf.reduce_sum(tf.log(C_tilde + EPSILON)) # (1, ) else: # Prepare all constants batch = tf.shape(X)[0] n = tf.shape(X)[1] ns = tf.shape(X_test)[1] # K_x + \sigma_x^* K = self.kernel_matrix_correlated_object_priors(X, X) # (batch, n n) K = K + tf.matrix_diag(varY) # (batch, n, n) chol_K = tf.linalg.cholesky(K) # (batch, n, n) # no cholesky_solve implementation # inv_K = tf.linalg.inv(_add_diagonal_jitter(K, 1e-2)) # lhood term 1/3 lhood_pi_term = tf.cast(n, dtype=self.dtype) * np.log(2 * np.pi) # lhood term 2/3 lhood_logdet_term = 2 * tf.reduce_sum(tf.log(tf.matrix_diag_part(chol_K)), 1) # (batch) # lhood term 3/3 Y = tf.expand_dims(Y, 2) # (batch, n, 1) iKY = tf.cholesky_solve(_add_diagonal_jitter(chol_K), Y) # (batch, n, 1) lh_quad_term = tf.matmul(tf.transpose(Y, (0, 2, 1)), iKY) # (batch, 1, 1) lh_quad_term = tf.reshape(lh_quad_term, [batch]) # no cholesky_solve implementation # iKY = tf.linalg.matvec(inv_K, Y) # lh_quad_term = tf.matmul(iKY, tf.transpose(Y, (1, 0))) # (batch, 1, 1) # lh_quad_term = tf.reshape(lh_quad_term, [batch]) # log P(Y|X) = -1/2 * ( n log(2 pi) + Y inv(K+noise) Y + log det(K+noise)) gp_lhood = -0.5 * (lhood_pi_term + lh_quad_term + lhood_logdet_term) # add C_tilde terms if not omit_C_tilde: gp_lhood = gp_lhood + tf.reduce_sum(tf.log(C_tilde + EPSILON)) # Compute posterior mean and variances Ks = self.kernel_matrix_correlated_object_priors(X, X_test) # (batch, n, ns) Ks_t = tf.transpose(Ks, (0, 2, 1)) # (batch, ns, n) # posterior mean p_m = tf.matmul(Ks_t, iKY) # no cholesky_solve implementation # p_m = tf.matmul(Ks_t, tf.expand_dims(iKY, 2)) p_m = tf.reshape(p_m, (batch, ns)) # posterior variance iK_Ks = tf.cholesky_solve(_add_diagonal_jitter(chol_K), Ks) # (batch, n, ns) Ks_iK_Ks = tf.reduce_sum(Ks * iK_Ks, axis=1) # (batch, ns) # no cholesky_solve implementation # Ks_iK_Ks = 1 - tf.linalg.diag_part(tf.matmul(Ks, tf.matmul(inv_K, Ks))) p_v = 1 - Ks_iK_Ks # (batch, ns) p_v = tf.reshape(p_v, (batch, ns)) p_v = tf.clip_by_value(p_v, 1e-6, 100) # drop first axis p_m = tf.squeeze(p_m) p_v = tf.squeeze(p_v) gp_lhood = tf.squeeze(gp_lhood) return p_m, p_v, gp_lhood def forward_pass_FGPVAE_rotated_mnist(data_batch, beta, vae, GP, N_t, clipping_qs, bayes_reg_view, omit_C_tilde, C_ma, lagrange_mult, alpha, kappa, GECO=False): """ :param data_batch: :param beta: :param vae: :param GP: :param N_t: :param clipping_qs: :param bayes_reg_view: whether or not to use Bayesian regresion view for linear kernel in global channels :param omit_C_tilde: omit C_tilde from derivation and modify cross-entropy term instead :param C_ma: average constraint from t-1 step (GECO) :param lagrange_mult: lambda from t-1 step (GECO) :param kappa: reconstruction level parameter for GECO :param alpha: moving average parameter for GECO :param GECO: whether or not to use GECO algorithm for training :return: """ images, aux_data = data_batch aux_data = tf.reshape(aux_data, (-1, N_t, 10)) L = vae.L L_w = GP.L_w w = tf.shape(images)[1] h = tf.shape(images)[2] K = tf.cast(w, dtype=tf.float64) * tf.cast(h, dtype=tf.float64) b = tf.cast(tf.shape(images)[0], dtype=tf.float64) # batch_size # ENCODER NETWORK qnet_mu, qnet_var = vae.encode(images) qnet_mu = tf.reshape(qnet_mu, (-1, N_t, L)) qnet_var = tf.reshape(qnet_var, (-1, N_t, L)) # clipping of VAE posterior variance if clipping_qs: qnet_var = tf.clip_by_value(qnet_var, 1e-3, 100) # GP p_m, p_v, lhoods_local, lhoods_global = [], [], [], [] for i in range(L_w): # fit local GPs p_m_i, p_v_i, lhood_i, K_local = GP.build_1d_gp_local(X=aux_data[:, :, 1], Y=qnet_mu[:, :, i], varY=qnet_var[:, :, i], X_test=aux_data[:, :, 1]) p_m.append(p_m_i) p_v.append(p_v_i) lhoods_local.append(lhood_i) ce_global_arr = [] for i in range(L_w, L): # fit global GPs if GP.object_prior_corr: object_aux_data_filtered = tf.transpose(aux_data[:, ::N_t, :], perm=[1, 0, 2]) bar_means, bar_vars, C_tilde = GP.preprocess_1d_gp_global_correlated_object_priors(qnet_mu[:, :, i], qnet_var[:, :, i]) p_m_i, p_v_i, lhood_i = GP.build_1d_gp_global_correlated_object_priors(object_aux_data_filtered, bar_means, bar_vars, object_aux_data_filtered, C_tilde, bayesian_reg_view=bayes_reg_view, omit_C_tilde=omit_C_tilde) if omit_C_tilde: ce_global_i = gauss_cross_entropy(p_m_i, p_v_i, bar_means, bar_vars) ce_global_arr.append(ce_global_i) else: p_m_i, p_v_i, lhood_i = GP.build_1d_gp_global(means=qnet_mu[:, :, i], vars=qnet_var[:, :, i]) # repeat p_m_i and p_v_i N_t times, since those are shared across all images within one object dataset D_t p_m_i = tf.tile(tf.expand_dims(p_m_i, 1), [1, N_t]) p_v_i = tf.tile(tf.expand_dims(p_v_i, 1), [1, N_t]) p_m.append(p_m_i) p_v.append(p_v_i) lhoods_global.append(lhood_i) p_m = tf.stack(p_m, axis=2) p_v = tf.stack(p_v, axis=2) if GP.object_prior_corr: # for local channels sum over latent channels and over digits' datasets # for global channels we only sum over latent channels (as there is only one global GP per channel) lhoods = tf.reduce_sum(lhoods_local, axis=(0, 1)) + tf.reduce_sum(lhoods_global, axis=0) # CE (cross-entropy) if omit_C_tilde: ce_global = tf.reduce_sum(ce_global_arr) ce_local = gauss_cross_entropy(p_m[:, :, :L_w], p_v[:, :, :L_w], qnet_mu[:, :, :L_w], qnet_var[:, :, :L_w]) ce_local = tf.reduce_sum(ce_local, (0, 1, 2)) # sum also over digits' datasets ce_term = ce_global + ce_local else: ce_term = gauss_cross_entropy(p_m, p_v, qnet_mu, qnet_var) ce_term = tf.reduce_sum(ce_term, (0, 1, 2)) # sum also over digits' datasets # KL part elbo_kl_part = lhoods - ce_term else: lhoods = lhoods_global + lhoods_local lhoods = tf.reduce_sum(lhoods, axis=0) # CE (cross-entropy) ce_term = gauss_cross_entropy(p_m, p_v, qnet_mu, qnet_var) ce_term = tf.reduce_sum(ce_term, (1, 2)) # KL part elbo_kl_part = lhoods - ce_term # SAMPLE epsilon = tf.random.normal(shape=tf.shape(p_m), dtype=tf.float64) latent_samples = p_m + epsilon * tf.sqrt(p_v) # DECODER NETWORK (Gaussian observational likelihood, MSE) recon_images = vae.decode(tf.reshape(latent_samples, (-1, L))) if GP.object_prior_corr: if GECO: recon_loss = tf.reduce_sum((tf.reshape(images, (-1, N_t, w, h)) - tf.reshape(recon_images, (-1, N_t, w, h))) ** 2, axis=[2, 3]) recon_loss = tf.reduce_sum(recon_loss/K - kappa**2) C_ma = alpha * C_ma + (1 - alpha) * recon_loss / b # elbo = - (1/L) * KL_term + lagrange_mult * C_ma # elbo = - (1/b) * KL_term + lagrange_mult * C_ma # elbo = - KL_term + lagrange_mult * C_ma elbo = - elbo_kl_part + lagrange_mult * (recon_loss / b + tf.stop_gradient(C_ma - recon_loss / b)) lagrange_mult = lagrange_mult * tf.exp(C_ma) else: recon_loss = tf.reduce_sum((tf.reshape(images, (-1, N_t, w, h)) - tf.reshape(recon_images, (-1, N_t, w, h))) ** 2, axis=[1, 2, 3]) recon_loss = tf.reduce_sum(recon_loss) / K elbo = -recon_loss + (beta / L) * elbo_kl_part else: if GECO: recon_loss = tf.reduce_mean((tf.reshape(images, (-1, N_t, w, h)) - tf.reshape(recon_images, (-1, N_t, w, h))) ** 2, axis=[2, 3]) N_t = tf.cast(N_t, dtype=tf.float64) C_ma = alpha * C_ma + (1 - alpha) * tf.reduce_mean(recon_loss - kappa ** 2) recon_loss = tf.reduce_sum(recon_loss - kappa ** 2) # elbo = - (1/L) * elbo_kl_part + lagrange_mult * C_ma # elbo = - (1/b) * elbo_kl_part + lagrange_mult * C_ma # elbo = - elbo_kl_part + lagrange_mult * C_ma elbo = - elbo_kl_part + lagrange_mult * (recon_loss / N_t + tf.stop_gradient(C_ma - recon_loss / N_t)) lagrange_mult = lagrange_mult * tf.exp(C_ma) else: recon_loss = tf.reduce_sum((tf.reshape(images, (-1, N_t, w, h)) - tf.reshape(recon_images, (-1, N_t, w, h))) ** 2, axis=[1, 2, 3]) # ELBO # beta plays role of sigma_gaussian_decoder here (\lambda(\sigma_y) in Casale paper) # K and L are not part of ELBO. They are used in loss objective to account for the fact that magnitudes of # reconstruction and KL terms depend on number of pixels (K) and number of latent GPs used (L), respectively recon_loss = recon_loss / K elbo = -recon_loss + (beta/L) * elbo_kl_part # average across object datasets elbo = tf.reduce_sum(elbo) elbo_kl_part = tf.reduce_sum(elbo_kl_part) recon_loss = tf.reduce_sum(recon_loss) return elbo, recon_loss, elbo_kl_part, p_m, p_v, qnet_mu, qnet_var, recon_images, latent_samples, C_ma, lagrange_mult def predict_FGPVAE_rotated_mnist(test_images, test_aux_data, train_images, train_aux_data, vae, GP, bayes_reg_view, omit_C_tilde, N_t=15, clipping_qs=False): """ Get FGPVAE predictions for rotated MNIST test data. :param test_data_batch: :param train_images: :param train_aux_data: :param vae: :param GP: :param N_t: :param clipping_qs: :return: """ L = vae.L L_w = GP.L_w w = tf.shape(train_images)[1] h = tf.shape(train_images)[2] train_aux_data = tf.reshape(train_aux_data, (-1, N_t, 10)) test_aux_data = tf.expand_dims(test_aux_data, 1) # encode train images qnet_mu, qnet_var = vae.encode(train_images) qnet_mu = tf.reshape(qnet_mu, (-1, N_t, L)) qnet_var = tf.reshape(qnet_var, (-1, N_t, L)) # clipping of VAE posterior variance if clipping_qs: qnet_var = tf.clip_by_value(qnet_var, 1e-3, 100) # GP, get latent embeddings for test images p_m, p_v = [], [] for i in range(L_w): # fit local GPs p_m_i, p_v_i, _ , _= GP.build_1d_gp_local(X=train_aux_data[:, :, 1], Y=qnet_mu[:, :, i], varY=qnet_var[:, :, i], X_test=test_aux_data[:, :, 1]) p_m.append(p_m_i) p_v.append(p_v_i) for i in range(L_w, L): # fit global GPs if GP.object_prior_corr: object_aux_data_filtered = tf.transpose(train_aux_data[:, ::N_t, :], perm=[1, 0, 2]) bar_means, bar_vars, C_tilde = GP.preprocess_1d_gp_global_correlated_object_priors(qnet_mu[:, :, i], qnet_var[:, :, i]) p_m_i, p_v_i, _ = GP.build_1d_gp_global_correlated_object_priors(object_aux_data_filtered, bar_means, bar_vars, object_aux_data_filtered, C_tilde, omit_C_tilde=omit_C_tilde, bayesian_reg_view=bayes_reg_view) else: p_m_i, p_v_i, _ = GP.build_1d_gp_global(means=qnet_mu[:, :, i], vars=qnet_var[:, :, i]) p_m.append(tf.expand_dims(p_m_i, 1)) p_v.append(tf.expand_dims(p_v_i, 1)) p_m = tf.stack(p_m, axis=2) p_v = tf.stack(p_v, axis=2) # SAMPLE epsilon = tf.random.normal(shape=tf.shape(p_m), dtype=tf.float64) latent_samples = p_m + epsilon * tf.sqrt(p_v) # decode, calculate error (Gaussian observational likelihood, MSE) recon_images = vae.decode(tf.reshape(latent_samples, (-1, L))) recon_loss = tf.reduce_mean((test_images - recon_images) ** 2) return recon_images, recon_loss def extrapolate_experiment_eval_data(mnist_path, digit, N_t, pred_angle_id=7, nr_angles=16): """ Prepare validation dataset for the extrapolate experiment. :param mnist_path: :param digit: :param N_t: how many angles do we observe for each image in test set :param pred_angle_id: which angle to leave out for prediction :param nr_angles: size of object dataset :return: """ eval_data_dict = pickle.load(open(mnist_path + 'eval_data{}_not_shuffled.p'.format(digit), 'rb')) eval_images, eval_aux_data = eval_data_dict["images"], eval_data_dict["aux_data"] pred_angle_mask = [pred_angle_id + i * nr_angles for i in range(int(len(eval_aux_data) / nr_angles))] not_pred_angle_mask = [i for i in range(len(eval_images)) if i not in pred_angle_mask] observed_images = eval_images[not_pred_angle_mask] observed_aux_data = eval_aux_data[not_pred_angle_mask] # randomly drop some observed angles if N_t < 15: digit_mask = [True]*N_t + [False]*(15-N_t) mask = [random.sample(digit_mask, len(digit_mask)) for _ in range(int(len(eval_aux_data)/nr_angles))] flatten = lambda l: [item for sublist in l for item in sublist] mask = flatten(mask) observed_images = observed_images[mask] observed_aux_data = observed_aux_data[mask] test_images = eval_images[pred_angle_mask] test_aux_data = eval_aux_data[pred_angle_mask] return observed_images, observed_aux_data, test_images, test_aux_data def latent_samples_FGPVAE(train_images, train_aux_data, vae, GP, N_t, clipping_qs=False): """ Get latent samples for training data. For t-SNE plots :) :param train_images: :param train_aux_data: :param vae: :param GP: :param clipping_qs: :return: """ train_aux_data = tf.reshape(train_aux_data, (-1, N_t, 10)) L = vae.L L_w = GP.L_w # ENCODER NETWORK qnet_mu, qnet_var = vae.encode(train_images) qnet_mu = tf.reshape(qnet_mu, (-1, N_t, L)) qnet_var = tf.reshape(qnet_var, (-1, N_t, L)) # clipping of VAE posterior variance if clipping_qs: qnet_var = tf.clip_by_value(qnet_var, 1e-3, 100) # GP p_m, p_v = [], [] for i in range(L_w): # fit local GPs p_m_i, p_v_i, _, _ = GP.build_1d_gp_local(X=train_aux_data[:, :, 1], Y=qnet_mu[:, :, i], varY=qnet_var[:, :, i], X_test=train_aux_data[:, :, 1]) p_m.append(p_m_i) p_v.append(p_v_i) for i in range(L_w, L): # fit global GPs p_m_i, p_v_i, lhood_i = GP.build_1d_gp_global(means=qnet_mu[:, :, i], vars=qnet_var[:, :, i]) # repeat p_m_i and p_v_i N_t times, since those are shared across all images within one object dataset D_t p_m_i = tf.tile(tf.expand_dims(p_m_i, 1), [1, N_t]) p_v_i = tf.tile(tf.expand_dims(p_v_i, 1), [1, N_t]) p_m.append(p_m_i) p_v.append(p_v_i) p_m = tf.stack(p_m, axis=2) p_v = tf.stack(p_v, axis=2) # SAMPLE epsilon = tf.random.normal(shape=tf.shape(p_m), dtype=tf.float64) latent_samples = p_m + epsilon * tf.sqrt(p_v) return latent_samples
42.456491
123
0.556145
14,600
0.490558
0
0
1,543
0.051845
0
0
8,632
0.290034
4b81b5fe4aceb22e7a99ad217502d745e5a0019f
1,069
py
Python
spotty/commands/abstract_command.py
Inculus/spotty
56863012668a6c13ad13c2a04f900047e229fbe6
[ "MIT" ]
246
2018-09-03T09:09:48.000Z
2020-07-18T21:07:15.000Z
spotty/commands/abstract_command.py
Inculus/spotty
56863012668a6c13ad13c2a04f900047e229fbe6
[ "MIT" ]
42
2018-10-09T19:41:56.000Z
2020-06-15T22:55:58.000Z
spotty/commands/abstract_command.py
Inculus/spotty
56863012668a6c13ad13c2a04f900047e229fbe6
[ "MIT" ]
27
2018-10-09T22:16:40.000Z
2020-06-08T22:26:00.000Z
from abc import ABC, abstractmethod from argparse import Namespace, ArgumentParser from spotty.commands.writers.abstract_output_writrer import AbstractOutputWriter class AbstractCommand(ABC): """Abstract class for a Spotty sub-command.""" @property @abstractmethod def name(self) -> str: """The sub-command name.""" raise NotImplementedError @property def description(self) -> str: """The sub-command description. It will be displayed in the help text.""" return '' def configure(self, parser: ArgumentParser): """Adds arguments for the sub-command.""" parser.add_argument('-d', '--debug', action='store_true', help='Show debug messages') @abstractmethod def run(self, args: Namespace, output: AbstractOutputWriter): """Runs the sub-command. Args: args: Arguments provided by argparse. output: Output writer. Raises: ValueError: If command's arguments can't be processed. """ raise NotImplementedError
30.542857
93
0.654818
902
0.843779
0
0
612
0.572498
0
0
454
0.424696
4b849b209996da99ee667a5b45419939d4653d3a
9,495
py
Python
tests/test_protocols/test_generator.py
cyenyxe/agents-aea
c2aec9127028ae13def3f69fbc80d35400de1565
[ "Apache-2.0" ]
null
null
null
tests/test_protocols/test_generator.py
cyenyxe/agents-aea
c2aec9127028ae13def3f69fbc80d35400de1565
[ "Apache-2.0" ]
1
2020-02-21T14:28:13.000Z
2020-03-05T14:53:53.000Z
tests/test_protocols/test_generator.py
cyenyxe/agents-aea
c2aec9127028ae13def3f69fbc80d35400de1565
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- # ------------------------------------------------------------------------------ # # Copyright 2018-2019 Fetch.AI Limited # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # ------------------------------------------------------------------------------ """This module contains the tests for the protocol generator.""" import inspect import os import shutil import tempfile import yaml from aea.configurations.base import ProtocolSpecification from aea.configurations.loader import ConfigLoader from aea.protocols.generator import ProtocolGenerator CUR_PATH = os.path.dirname(inspect.getfile(inspect.currentframe())) # type: ignore class TestGenerateProtocol: """Test that the generating a protocol works correctly in correct preconditions.""" @classmethod def setup_class(cls): """Set the test up.""" # Specification cls.protocol_name = "two_party_negotiation" cls.specification_file_name = "spec.yaml" correct_specification = { "name": cls.protocol_name, "author": "fetchai", "version": "0.1.0", "license": "Apache-2.0", "description": "A protocol for negotiation over a fixed set of resources involving two parties.", "speech_acts": { "cfp": {"query": "DataModel"}, "propose": {"query": "DataModel", "price": "float"}, "accept": {}, "decline": {}, "match_accept": {}, }, } # Dump the config cls.cwd = os.getcwd() # os.mkdir(os.path.join(CUR_PATH, "temp")) cls.t = tempfile.mkdtemp() os.chdir(cls.t) # cls.path_to_specification = os.path.join(".", cls.specification_file_name) cls.path_to_specification = os.path.join(cls.t, cls.specification_file_name) yaml.safe_dump(correct_specification, open(cls.path_to_specification, "w")) # Load the config cls.config_loader = ConfigLoader( "protocol-specification_schema.json", ProtocolSpecification ) cls.protocol_specification = cls.config_loader.load( open(cls.path_to_specification) ) # Generate the protocol cls.protocol_generator = ProtocolGenerator(cls.protocol_specification, cls.t) cls.protocol_generator.generate() # Add as module # dotted_path = "packages.fetchai.protocols." + cls.protocol_name # import pdb;pdb.set_trace() # module_object = load_module(dotted_path, Path(os.path.join(cls.t, cls.protocol_name))) # import_module(dotted_path, module_object) # sys.modules[dotted_path] = module_object # def test_exit_code_equal_to_0(self): # """Test that the exit code is equal to 0.""" # from packages.fetchai.protocols.two_party_negotiation.message import TwoPartyNegotiationMessage # # from two_party_negotiation.serialization import TwoPartyNegotiationSerializer # # from two_party_negotiation.message import DataModel # assert 0 == 0 @classmethod def teardown_class(cls): """Tear the test down.""" os.chdir(cls.cwd) try: shutil.rmtree(cls.t) # os.remove(os.path.join(cls.t, cls.protocol_name)) except (OSError, IOError): pass # class TestCases(TestCase): # """Test class for the light protocol generator.""" # # def test_all_custom_data_types(self): # """Test all custom data types.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "all_custom.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # test_protocol_template.load() # test_protocol_generator = ProtocolGenerator(test_protocol_template, 'tests') # test_protocol_generator.generate() # # from two_party_negotiation_protocol.message import TwoPartyNegotiationMessage # from two_party_negotiation_protocol.serialization import TwoPartyNegotiationSerializer # from two_party_negotiation_protocol.message import DataModel # from two_party_negotiation_protocol.message import Signature # # data_model = DataModel() # signature = Signature() # content_list = [data_model, signature] # # message = TwoPartyNegotiationMessage(message_id=5, target=4, performative="propose", contents=content_list) # print(str.format("message is {}", message)) # message.check_consistency() # serialized_message = TwoPartyNegotiationSerializer().encode(msg=message) # print(str.format("serialized message is {}", serialized_message)) # deserialised_message = TwoPartyNegotiationSerializer().decode(obj=serialized_message) # print(str.format("deserialized message is {}", deserialised_message)) # # assert message == deserialised_message, "Failure" # # def test_correct_functionality(self): # """End to end test of functionality.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "correct_spec.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # test_protocol_template.load() # test_protocol_generator = ProtocolGenerator(test_protocol_template, 'tests') # test_protocol_generator.generate() # # from two_party_negotiation_protocol.message import TwoPartyNegotiationMessage # from two_party_negotiation_protocol.serialization import TwoPartyNegotiationSerializer # from two_party_negotiation_protocol.message import DataModel # # data_model = DataModel() # content_list = [data_model, 10.5] # # message = TwoPartyNegotiationMessage(message_id=5, target=4, performative="propose", contents=content_list) # print(str.format("message is {}", message)) # message.check_consistency() # serialized_message = TwoPartyNegotiationSerializer().encode(msg=message) # print(str.format("serialized message is {}", serialized_message)) # deserialised_message = TwoPartyNegotiationSerializer().decode(obj=serialized_message) # print(str.format("deserialized message is {}", deserialised_message)) # # assert message == deserialised_message, "Failure" # # def test_missing_name(self): # """Test missing name handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "missing_name.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # self.assertRaises(ProtocolSpecificationParseError, test_protocol_template.load) # # def test_missing_description(self): # """Test missing description handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "missing_description.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # assert test_protocol_template.load(), "Failure" # # def test_missing_speech_acts(self): # """Test missing speech acts handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "missing_speech_acts.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # self.assertRaises(ProtocolSpecificationParseError, test_protocol_template.load) # # def test_extra_fields(self): # """Test extra fields handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "extra_fields.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # assert test_protocol_template.load(), "Failure" # # def test_one_document(self): # """Test one document handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "one_document.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # self.assertRaises(ProtocolSpecificationParseError, test_protocol_template.load) # # def test_wrong_speech_act_type_sequence_performatives(self): # """Test wrong speech act handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "wrong_speech_act_type_sequence_performatives.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # self.assertRaises(ProtocolSpecificationParseError, test_protocol_template.load) # # def test_wrong_speech_act_type_dictionary_contents(self): # """Test wrong speech act dictionary contents handling.""" # test_protocol_specification_path = os.path.join(CUR_PATH, "data", "wrong_speech_act_type_dictionary_contents.yaml") # test_protocol_template = ProtocolTemplate(test_protocol_specification_path) # # self.assertRaises(ProtocolSpecificationParseError, test_protocol_template.load)
45.214286
128
0.685519
2,705
0.284887
0
0
2,195
0.231174
0
0
7,697
0.810637
4b86ef7acd08f81f39f9fde4c5d2779a3995da3e
6,981
py
Python
tabfkioskgoogledrive/MyGDTest3.py
isalan06/myflaskapiserver
2922f62c9b9ede2b6cba2db774e924b226a120f7
[ "MIT" ]
null
null
null
tabfkioskgoogledrive/MyGDTest3.py
isalan06/myflaskapiserver
2922f62c9b9ede2b6cba2db774e924b226a120f7
[ "MIT" ]
null
null
null
tabfkioskgoogledrive/MyGDTest3.py
isalan06/myflaskapiserver
2922f62c9b9ede2b6cba2db774e924b226a120f7
[ "MIT" ]
null
null
null
import os.path import os from googleapiclient.discovery import build from google_auth_oauthlib.flow import InstalledAppFlow from google.auth.transport.requests import Request from google.oauth2.credentials import Credentials from datetime import datetime # If modifying these scopes, delete the file token.json. SCOPES = ['https://www.googleapis.com/auth/drive'] def main(): """Shows basic usage of the Drive v3 API. Prints the names and ids of the first 10 files the user has access to. """ creds = None # The file token.json stores the user's access and refresh tokens, and is # created automatically when the authorization flow completes for the first # time. if os.path.exists('token.json'): creds = Credentials.from_authorized_user_file('token.json', SCOPES) # If there are no (valid) credentials available, let the user log in. if not creds or not creds.valid: if creds and creds.expired and creds.refresh_token: print("Refresh Creds") creds.refresh(Request()) else: flow = InstalledAppFlow.from_client_secrets_file( 'client_secrets.json', SCOPES) creds = flow.run_local_server(port=0) # Save the credentials for the next run with open('token.json', 'w') as token: token.write(creds.to_json()) service = build('drive', 'v3', credentials=creds) # Call the Drive v3 API results = service.files().list( q="mimeType = 'application/vnd.google-apps.folder' and '0ALNhV0hP-QYDUk9PVA' in parents", pageSize=100, fields="nextPageToken, files(id, name, parents)").execute() items = results.get('files', []) pic_id = '' if not items: print('No files found.') else: print('1st Files:') for item in items: if item['name']=='KIOSK Picture': pic_id = item['id'] print(u'{0} ({1}) - {2}'.format(item['name'], item['id'], item['parents'])) #print(pic_id) # Check Machine ID q_str = "mimeType = 'application/vnd.google-apps.folder' and '" + str(pic_id) +"' in parents" #print(q_str) results = service.files().list( q=q_str, #"mimeType = 'application/vnd.google-apps.folder' and '" + str(pic_id) +"' in parents", pageSize=10, fields="nextPageToken, files(id, name, parents)").execute() items = results.get('files', []) bHasBaseFolder = False sMachineID = 'Test_MachineID' sMachineID_ID = '' if not items: print('No files found.') else: print('2nd Files:') for item in items: if item['name']==sMachineID: bHasBaseFolder = True sMachineID_ID = item['id'] print(u'{0} ({1}) - {2}'.format(item['name'], item['id'], item['parents'])) if bHasBaseFolder == False: file_metadata = { 'name': sMachineID, 'mimeType': 'application/vnd.google-apps.folder', 'parents': [str(pic_id)] } file = service.files().create(body=file_metadata, fields='id').execute() sMachineID_ID = str(file.get('id')) print('Folder ID: %s' % file.get('id')) #print(sMachineID_ID) # Check Date Folder sTodayDateString = datetime.now().strftime("%Y%m%d") sTodayDate_ID = '' bHasBaseFolder = False q_str = "mimeType = 'application/vnd.google-apps.folder' and '" + str(sMachineID_ID) +"' in parents" results = service.files().list( q=q_str, pageSize=10, fields="nextPageToken, files(id, name, parents)").execute() items = results.get('files', []) if not items: print('No files found.') else: print('3nd Files:') for item in items: if item['name']==sTodayDateString: bHasBaseFolder = True sTodayDate_ID = item['id'] print(u'{0} ({1}) - {2}'.format(item['name'], item['id'], item['parents'])) if bHasBaseFolder == False: file_metadata = { 'name': sTodayDateString, 'mimeType': 'application/vnd.google-apps.folder', 'parents': [str(sMachineID_ID)] } file = service.files().create(body=file_metadata, fields='id').execute() sTodayDate_ID = str(file.get('id')) print('Folder ID: %s' % file.get('id')) #Check Test Location sTestLocation='我是測試考場(真的是測試用)' sTestLocation_ID = '' bHasBaseFolder = False q_str = "mimeType = 'application/vnd.google-apps.folder' and '" + str(sTodayDate_ID) +"' in parents" results = service.files().list( q=q_str, pageSize=10, fields="nextPageToken, files(id, name, parents)").execute() items = results.get('files', []) if not items: print('No files found.') else: print('4nd Files:') for item in items: if item['name']==sTestLocation: bHasBaseFolder = True sTestLocation_ID = item['id'] print(u'{0} ({1}) - {2}'.format(item['name'], item['id'], item['parents'])) if bHasBaseFolder == False: file_metadata = { 'name': sTestLocation, 'mimeType': 'application/vnd.google-apps.folder', 'parents': [str(sTodayDate_ID)] } file = service.files().create(body=file_metadata, fields='id').execute() sTestLocation_ID = str(file.get('id')) print('Folder ID: %s' % file.get('id')) sTestLocation_ID = CreateGoogleDriveFolder(service, sTestLocation, sTodayDate_ID) print('Check Function') print(sTestLocation_ID) def CreateGoogleDriveFolder(service, titlestring, folderid): returnfolderid='' bHasFolder = False q_str = "mimeType = 'application/vnd.google-apps.folder' and '" + str(folderid) +"' in parents" results = service.files().list( q=q_str, pageSize=10, fields="nextPageToken, files(id, name, parents)").execute() items = results.get('files', []) if not items: print('No files found.') else: for item in items: if item['name']==titlestring: bHasFolder = True returnfolderid = item['id'] print(u'{0} ({1}) - {2}'.format(item['name'], item['id'], item['parents'])) if bHasFolder == False: try: file_metadata = { 'name': titlestring, 'mimeType': 'application/vnd.google-apps.folder', 'parents': [str(folderid)] } file = service.files().create(body=file_metadata, fields='id').execute() returnfolderid = str(file.get('id')) print('Folder ID: %s' % file.get('id')) except Exception as ex: print(ex) return returnfolderid if __name__ == '__main__': main()
37.331551
104
0.578284
0
0
0
0
0
0
0
0
2,228
0.318059
4b8822ba30ebfce8e71617e36c3ea7b8cc487de1
46
py
Python
qarithmetic/__init__.py
daniil-lyakhov/QArithmetic
7a5df9504e17c1979107c119bbaf0c5b2750a619
[ "Apache-2.0" ]
null
null
null
qarithmetic/__init__.py
daniil-lyakhov/QArithmetic
7a5df9504e17c1979107c119bbaf0c5b2750a619
[ "Apache-2.0" ]
null
null
null
qarithmetic/__init__.py
daniil-lyakhov/QArithmetic
7a5df9504e17c1979107c119bbaf0c5b2750a619
[ "Apache-2.0" ]
null
null
null
from .QArithmetic import * from .qft import *
15.333333
26
0.73913
0
0
0
0
0
0
0
0
0
0
4b88bb3938cbed6bd9ddf6e52090c0d588399179
2,631
py
Python
clustering/conditional_probs.py
griffij/QuakeRates
70069bb271a1987e72fcbdf3aa0c0a8a79591580
[ "Apache-2.0" ]
null
null
null
clustering/conditional_probs.py
griffij/QuakeRates
70069bb271a1987e72fcbdf3aa0c0a8a79591580
[ "Apache-2.0" ]
null
null
null
clustering/conditional_probs.py
griffij/QuakeRates
70069bb271a1987e72fcbdf3aa0c0a8a79591580
[ "Apache-2.0" ]
null
null
null
"""Calculate conditional probability of a short interevent time being followed by another short interevent time, compared with the unconditional probability. This is used to test whether fault records have memory """ import os, sys from glob import glob import numpy as np import matplotlib.pyplot as plt from QuakeRates.dataman.parse_params import parse_param_file, \ get_event_sets # Define parameter files filepath = '../params' param_file_list = glob(os.path.join(filepath, '*.txt')) n_samples = 500 # Number of Monte Carlo samples of the eq chronologies half_n = int(n_samples/2) plot_dir = './plots_conditional_probs' if not os.path.exists(plot_dir): os.makedirs(plot_dir) # Define subset to take #faulting_styles = ['Reverse'] #faulting_styles = ['Normal'] #faulting_styles = ['Strike_slip'] faulting_styles = ['all'] tectonic_regions = ['all'] #tectonic_regions = ['Plate_boundary_master', 'Plate_boundary_network'] min_number_events = 10 names, event_sets, event_certainties, num_events = \ get_event_sets(param_file_list, tectonic_regions, faulting_styles, min_number_events) # Now loop over paleo-earthquake records for i, event_set in enumerate(event_sets): # Generate some chronologies event_set.gen_chronologies(n_samples, observation_end=2019, min_separation=1) print(num_events[i]) event_set.calculate_cov() # Calculate interevent times and mean as part of this # Lists to store results uncond_probs = [] cond_probs = [] for j, sample in enumerate(event_set.interevent_times.T): num_less_mean = len(np.argwhere(sample < event_set.means[j])) uncond_prob_less_mean = num_less_mean/event_set.num_events count_short = 0 for k, ie_time in enumerate(sample): if k==0: ie_time_0 = ie_time else: if ie_time < event_set.means[i] and \ ie_time_0 < event_set.means[i]: count_short += 1 ie_time_0 = ie_time cond_prob_less_mean = count_short/num_less_mean uncond_probs.append(uncond_prob_less_mean) cond_probs.append(cond_prob_less_mean) print(uncond_probs) print(cond_probs) uncond_probs = np.array(uncond_probs) cond_probs = np.array(cond_probs) probs_ratio = cond_probs/uncond_probs print(probs_ratio) plt.clf() plt.hist(probs_ratio, bins = 10, facecolor='0.6', edgecolor='0.2', density=True) figname = 'conditional_prob_ratio_histogram_%s.png' % names[i] fig_filename = os.path.join(plot_dir, figname) plt.savefig(fig_filename)
36.541667
84
0.707336
0
0
0
0
0
0
0
0
734
0.278981
4b8d4301cde6fb6de24f8efb96ff5081761e33de
643
py
Python
Computer Science/Development/GUI/robo_controls.py
zbendt/ECE-Capstone-Project
1bafc61f896191ccd5a843980500fb4b8bbeb8bd
[ "MIT" ]
null
null
null
Computer Science/Development/GUI/robo_controls.py
zbendt/ECE-Capstone-Project
1bafc61f896191ccd5a843980500fb4b8bbeb8bd
[ "MIT" ]
null
null
null
Computer Science/Development/GUI/robo_controls.py
zbendt/ECE-Capstone-Project
1bafc61f896191ccd5a843980500fb4b8bbeb8bd
[ "MIT" ]
null
null
null
import time #Controls functions for the delta sleep_time = 0.5 def turn_on_vacuum(): print("Turning on vacuum pump") def pickup(): print("Picking up sample...") time.sleep(sleep_time) def drop(): print("Dropping sample...") time.sleep(sleep_time) def move_to_start(): print("Moving to Start...") time.sleep(sleep_time) def move_to_camera(): print("Moving to Camera...") time.sleep(sleep_time) def move_to_passed(): print("Moving to Pass Stack...") time.sleep(sleep_time) def move_to_failed(): print("Moving to Fail Stack...") time.sleep(sleep_time)
20.09375
37
0.634526
0
0
0
0
0
0
0
0
190
0.29549
4b8e4f10e68bbf6b6e9801bf943ec3cb8b4d1bf7
3,120
py
Python
src/dynamic_programming/basic_scripts/value_iteration.py
johannesharmse/move_37_course
a2060129cbc6fb651113aa18f1a6ea2673845182
[ "MIT" ]
1
2019-03-13T06:29:54.000Z
2019-03-13T06:29:54.000Z
src/dynamic_programming/basic_scripts/value_iteration.py
johannesharmse/move_37_course
a2060129cbc6fb651113aa18f1a6ea2673845182
[ "MIT" ]
null
null
null
src/dynamic_programming/basic_scripts/value_iteration.py
johannesharmse/move_37_course
a2060129cbc6fb651113aa18f1a6ea2673845182
[ "MIT" ]
null
null
null
# From The School of AI's Move 37 Course https://www.theschool.ai/courses/move-37-course/ # Coding demo by Colin Skow # Forked from https://github.com/lazyprogrammer/machine_learning_examples/tree/master/rl # Credit goes to LazyProgrammer from __future__ import print_function, division from builtins import range # Note: you may need to update your version of future # sudo pip install -U future import numpy as np from grid_world import standard_grid from utils import print_values, print_policy # SMALL_ENOUGH is referred to by the mathematical symbol theta in equations SMALL_ENOUGH = 1e-3 GAMMA = 0.9 ALL_POSSIBLE_ACTIONS = ('U', 'D', 'L', 'R') def best_action_value(grid, V, s): # finds the highest value action (max_a) from state s, returns the action and value best_a = None best_value = float('-inf') grid.set_state(s) # loop through all possible actions to find the best current action for a in ALL_POSSIBLE_ACTIONS: transititions = grid.get_transition_probs(a) expected_v = 0 expected_r = 0 for (prob, r, state_prime) in transititions: expected_r += prob * r expected_v += prob * V[state_prime] v = expected_r + GAMMA * expected_v if v > best_value: best_value = v best_a = a return best_a, best_value def calculate_values(grid): # initialize V(s) V = {} states = grid.all_states() for s in states: V[s] = 0 # repeat until convergence # V[s] = max[a]{ sum[s',r] { p(s',r|s,a)[r + gamma*V[s']] } } while True: # biggest_change is referred to by the mathematical symbol delta in equations biggest_change = 0 for s in grid.non_terminal_states(): old_v = V[s] _, new_v = best_action_value(grid, V, s) V[s] = new_v biggest_change = max(biggest_change, np.abs(old_v - new_v)) if biggest_change < SMALL_ENOUGH: break return V def initialize_random_policy(): # policy is a lookup table for state -> action # we'll randomly choose an action and update as we learn policy = {} for s in grid.non_terminal_states(): policy[s] = np.random.choice(ALL_POSSIBLE_ACTIONS) return policy def calculate_greedy_policy(grid, V): policy = initialize_random_policy() # find a policy that leads to optimal value function for s in policy.keys(): grid.set_state(s) # loop through all possible actions to find the best current action best_a, _ = best_action_value(grid, V, s) policy[s] = best_a return policy if __name__ == '__main__': # this grid gives you a reward of -0.1 for every non-terminal state # we want to see if this will encourage finding a shorter path to the goal grid = standard_grid(obey_prob=0.8, step_cost=-0.5) # print rewards print("rewards:") print_values(grid.rewards, grid) # calculate accurate values for each square V = calculate_values(grid) # calculate the optimum policy based on our values policy = calculate_greedy_policy(grid, V) # our goal here is to verify that we get the same answer as with policy iteration print("values:") print_values(V, grid) print("policy:") print_policy(policy, grid)
31.515152
89
0.709936
0
0
0
0
0
0
0
0
1,329
0.425962
4b8f66af4fc844e8c289287b2a2bc4ba119f529e
19,238
py
Python
photoplace/addons/CSVImport/GTKcsvimport.py
jriguera/photoplace
93674ef8531d0e5b8f26de9ba568ed8e115b27e1
[ "Apache-2.0" ]
10
2015-02-20T19:01:19.000Z
2021-12-13T23:07:19.000Z
photoplace/addons/CSVImport/GTKcsvimport.py
jriguera/photoplace
93674ef8531d0e5b8f26de9ba568ed8e115b27e1
[ "Apache-2.0" ]
1
2020-06-16T13:23:05.000Z
2021-02-13T14:14:57.000Z
photoplace/addons/CSVImport/GTKcsvimport.py
jriguera/photoplace
93674ef8531d0e5b8f26de9ba568ed8e115b27e1
[ "Apache-2.0" ]
4
2017-03-28T23:06:14.000Z
2019-09-25T07:59:36.000Z
#!/usr/bin/env python # -*- coding: utf-8 -*- # # GTKcsvimport.py # # Copyright 2010-2015 Jose Riguera Lopez <[email protected]> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Parse a CSV to add variables or geolocate photos. GTK User Interface. """ __program__ = "photoplace.csvimport" __author__ = "Jose Riguera Lopez <[email protected]>" __version__ = "0.1.2" __date__ = "Dec 2014" __license__ = "Apache 2.0" __copyright__ ="(c) Jose Riguera" import os.path import csv import sys import codecs import warnings import gettext import locale warnings.filterwarnings('ignore', module='gtk') try: import pygtk pygtk.require("2.0") import gtk import gobject except Exception as e: warnings.resetwarnings() print("Warning: %s" % str(e)) print("You don't have the PyGTK 2.0 module installed") raise warnings.resetwarnings() from csvimport import * # I18N gettext support __GETTEXT_DOMAIN__ = __program__ __PACKAGE_DIR__ = os.path.abspath(os.path.dirname(__file__)) __LOCALE_DIR__ = os.path.join(__PACKAGE_DIR__, u"locale") try: if not os.path.isdir(__LOCALE_DIR__): print ("Error: Cannot locate default locale dir: '%s'." % (__LOCALE_DIR__)) __LOCALE_DIR__ = None locale.setlocale(locale.LC_ALL,"") #gettext.bindtextdomain(__GETTEXT_DOMAIN__, __LOCALE_DIR__) t = gettext.translation(__GETTEXT_DOMAIN__, __LOCALE_DIR__, fallback=False) _ = t.ugettext except Exception as e: print ("Error setting up the translations: %s" % (str(e))) _ = lambda s: unicode(s) class GTKCSVImport(object): def __init__(self, plugin, gui, userfacade, logger): object.__init__(self) self.plugin = gtk.VBox(False) self.logger = logger self.options = None self.userfacade = userfacade # 1st line hbox = gtk.HBox(False) label_name = gtk.Label() align = gtk.Alignment(0.01, 0.5, 0, 0) label_name.set_markup(_("CSV file:")) align.add(label_name) hbox.pack_start(align, False, False, 5) self.button_addfile = gtk.Button() image = gtk.Image() image.set_from_stock(gtk.STOCK_ADD, gtk.ICON_SIZE_BUTTON) self.button_addfile.set_image(image) self.button_addfile.set_tooltip_text(_("Select a CSV file to load photo's variables")) self.button_addfile.set_label(_("Select file")) self.button_addfile.connect('clicked', self._load_csv) align = gtk.Alignment(0.01, 0.5, 0, 0) align.add(self.button_addfile) hbox.pack_start(align, False, False, 5) self.plugin.pack_start(hbox, False, False, 5) # 3rd line hbox_headers = gtk.HBox(False) label_headers = gtk.Label() label_headers.set_markup(_("Headers:")) hbox_headers.pack_start(label_headers, False, False, 5) self.entry_headers = gtk.Entry(max=256) self.entry_headers.connect('focus-out-event', self._out_entry) self.entry_headers.connect('activate', self._set_entry) self.entry_headers.set_tooltip_text(_("List of column headers of the CSV file. Each header will be a variable for each photo")) self.entry_headers.set_sensitive(False) hbox_headers.pack_start(self.entry_headers, True, True, 2) label_headerid = gtk.Label() label_headerid.set_markup(_("where photo name is:")) hbox_headers.pack_start(label_headerid, False, False, 0) self.cb_headerid = gtk.combo_box_new_text() self.cb_headerid.connect("changed", self._combo_id) self.cb_headerid.set_tooltip_text(_("Name of the column to match with each photo file name. It must be one of the Headers")) self.cb_headerid.set_sensitive(False) self.cb_headerid.append_text(' ') hbox_headers.pack_start(self.cb_headerid, False, False, 5) self.plugin.pack_start(hbox_headers, False, False, 5) # 4st line self.checkbutton_geolocate = gtk.CheckButton(_("Geolocate photos using CSV headers")) self.checkbutton_geolocate.set_tooltip_text(_("It is active, it will assign the following headers to each photo. It will geotag the photos by using those headers, but, warning: GPX data will take precedence!")) self.checkbutton_geolocate.connect('toggled', self._geolocate) self.checkbutton_geolocate.set_sensitive(False) # Headers Variables self.frame = gtk.Frame() self.frame.set_label_widget(self.checkbutton_geolocate) table = gtk.Table(2, 4, True) label_lat = gtk.Label() label_lat.set_markup(_("Latitude:")) align = gtk.Alignment(1.00, 0.5, 0, 0) align.add(label_lat) table.attach(align, 0, 1, 0, 1, gtk.FILL) self.cb_lat = gtk.combo_box_new_text() self.cb_lat.connect("changed", self._combo_geolocate, CSVImport_CONFKEY_HEADER_LAT) self.cb_lat.set_tooltip_text(_("Latitude header name")) table.attach(self.cb_lat, 1, 2, 0, 1) label_lon = gtk.Label() label_lon.set_markup(_("Longitude:")) align = gtk.Alignment(1.00, 0.5, 0, 0) align.add(label_lon) table.attach(align, 2, 3, 0, 1, gtk.FILL) self.cb_lon = gtk.combo_box_new_text() self.cb_lon.connect("changed", self._combo_geolocate, CSVImport_CONFKEY_HEADER_LON) self.cb_lon.set_tooltip_text(_("Longitude header name")) table.attach(self.cb_lon, 3, 4, 0, 1) label_date = gtk.Label() label_date.set_markup(_("Time-Date:")) align = gtk.Alignment(1.00, 0.5, 0, 0) align.add(label_date) table.attach(align, 0, 1, 1, 2) self.cb_date = gtk.combo_box_new_text() self.cb_date.connect("changed", self._combo_geolocate, CSVImport_CONFKEY_HEADER_DATE) table.attach(self.cb_date, 1, 2, 1, 2) label_ele = gtk.Label() label_ele.set_markup(_("Elevation:")) align = gtk.Alignment(1.00, 0.5, 0, 0) align.add(label_ele) table.attach(align, 2, 3, 1, 2) self.cb_ele = gtk.combo_box_new_text() self.cb_ele.connect("changed", self._combo_geolocate, CSVImport_CONFKEY_HEADER_ELE) self.cb_ele.set_tooltip_text(_("Elevation header name")) table.attach(self.cb_ele, 3, 4, 1, 2) table.set_border_width(20) table.set_row_spacings(5) self.frame.add(table) self.frame.set_border_width(5) self.frame.set_sensitive(False) self.plugin.pack_start(self.frame, False, False, 5) # Button self.button_process = gtk.Button() self.button_process.set_label(_("Process")) image = gtk.Image() image.set_from_stock(gtk.STOCK_EXECUTE, gtk.ICON_SIZE_BUTTON) self.button_process.set_image(image) self.button_process.connect('clicked', self.process) align = gtk.Alignment(0.50, 0.5, 0.1, 0) align.add(self.button_process) self.plugin.pack_start(align, False, False, 0) self.button_process.set_sensitive(False) # Attributes self.rootplugin = plugin self.rootgui = gui self.window = gui.builder.get_object("window") self.events = True def _load_csv(self, widget): dialog = gtk.FileChooserDialog(title=_("Select CSV file ..."), parent=self.window, action=gtk.FILE_CHOOSER_ACTION_OPEN, buttons=(gtk.STOCK_CANCEL, gtk.RESPONSE_CANCEL, gtk.STOCK_OPEN, gtk.RESPONSE_OK)) ffilter = gtk.FileFilter() ffilter.set_name(_("Comma Separated Values (CSV)")) ffilter.add_pattern("*.csv") dialog.add_filter(ffilter) ffilter = gtk.FileFilter() ffilter.set_name(_("All files")) ffilter.add_pattern("*") dialog.add_filter(ffilter) filename = None if dialog.run() == gtk.RESPONSE_OK: filename = dialog.get_filename() dialog.destroy() self._set_csv(filename) def _set_csv(self, filename): if filename != None and os.path.isfile(filename): if not isinstance(filename, unicode): try: filename = unicode(filename, 'UTF-8') except: pass shortfilename = " " + os.path.basename(filename) + " " if len(shortfilename) > 150: shortfilename = shortfilename[0:150] + " ... " image = self.button_addfile.get_image() image.clear() self.button_addfile.set_label(shortfilename) self.frame.set_sensitive(True) self.checkbutton_geolocate.set_sensitive(True) self.cb_headerid.set_sensitive(True) self.entry_headers.set_sensitive(True) self.button_process.set_sensitive(True) self._init_csv(filename) self._geolocate() else: self.reset() def _init_csv(self, filename): dgettext = dict() dgettext['file'] = filename try: fd = open(filename, 'rb') except Exception as e: dgettext['error'] = str(e) msg = _("Cannot read file '%(file)s': %(error)s") self.logger.error(msg % dgettext) self.rootgui.show_dialog(_("Error"), msg, _('Please check file permisions')) else: dialect = 'excel' headers = self.options[CSVImport_CONFKEY_HEADERS] delimiter = self.options[CSVImport_CONFKEY_DELIMITER] quotechar = self.options[CSVImport_CONFKEY_QUOTECHAR] if not delimiter or not quotechar: dialect = csv.Sniffer().sniff(fd.read(10240)) delimiter = dialect.delimiter quotechar = dialect.quotechar fd.seek(0) else: dgettext['delimiter'] = delimiter dgettext['quotechar'] = quotechar has_header = csv.Sniffer().has_header(fd.read(10240)) fd.seek(0) headers_defined = False if headers: headers_defined = True else: reader = csv.DictReader(fd, dialect=dialect, delimiter=delimiter, quotechar=quotechar) if has_header: reader.next() headers = reader.fieldnames headers_defined = True self.options[CSVImport_CONFKEY_FILENAME] = filename if not headers_defined: msg = _("File has no headers") tip = _('You have to define the name of the headers.') self.rootgui.show_dialog(_("Warning"), msg, tip, gtk.MESSAGE_WARNING) else: self.entry_headers.set_text(', '.join(headers)) try: index = headers.index(self.options[CSVImport_CONFKEY_HEADER_ID]) except: index = 0 self._set_combo(self.cb_headerid, headers, CSVImport_CONFKEY_HEADER_ID, index) self.rootplugin.update_headers(headers) fd.close() def process(self, widget=None): filename = self.options[CSVImport_CONFKEY_FILENAME] if filename != None: self.rootplugin.init_csv(filename) counter = self.rootplugin.process_csv(self.userfacade.state.geophotos) self.rootplugin.logger.info(_("%d photos processed with CSV data.") % counter) self.rootplugin.end_csv() self.rootgui.reload_treeviewgeophotos() def _geolocate(self, widget=None): self.events = False value = self.checkbutton_geolocate.get_active() self.cb_date.set_sensitive(value) self.cb_ele.set_sensitive(value) self.cb_lon.set_sensitive(value) self.cb_lat.set_sensitive(value) self.options[CSVImport_CONFKEY_GEOLOCATE] = value if not value: self.options[CSVImport_CONFKEY_HEADER_LAT] = '' self.options[CSVImport_CONFKEY_HEADER_LON] = '' self.options[CSVImport_CONFKEY_HEADER_ELE] = '' self.options[CSVImport_CONFKEY_HEADER_DATE] = '' self._empty_combo(self.cb_date) self._empty_combo(self.cb_ele) self._empty_combo(self.cb_lon) self._empty_combo(self.cb_lat) else: headers = [" "] + self.options[CSVImport_CONFKEY_HEADERS] try: headers.remove(self.cb_headerid.get_active_text()) except: pass self._set_combo(self.cb_date, headers) self._set_combo(self.cb_ele, headers) self._set_combo(self.cb_lon, headers) self._set_combo(self.cb_lat, headers) counter = 0 for i in headers: item = i.lower() if i == self.options[CSVImport_CONFKEY_HEADER_LAT]: self.cb_lat.set_active(counter) elif i == self.options[CSVImport_CONFKEY_HEADER_LON]: self.cb_lon.set_active(counter) elif i == self.options[CSVImport_CONFKEY_HEADER_ELE]: self.cb_ele.set_active(counter) elif i == self.options[CSVImport_CONFKEY_HEADER_DATE]: self.cb_date.set_active(counter) elif 'lat' in item: self.cb_lat.set_active(counter) self.options[CSVImport_CONFKEY_HEADER_LAT] = i elif 'lon' in item: self.cb_lon.set_active(counter) self.options[CSVImport_CONFKEY_HEADER_LON] = i elif 'ele' in item: self.cb_ele.set_active(counter) self.options[CSVImport_CONFKEY_HEADER_ELE] = i elif 'date' in item or 'time' in item: self.cb_date.set_active(counter) self.options[CSVImport_CONFKEY_HEADER_DATE] = i counter += 1 self.events = True def _out_entry(self, widget, e): widget.set_text(', '.join(self.options[CSVImport_CONFKEY_HEADERS])) return False def _set_entry(self, widget): value = widget.get_text() items = [] try: char = None for c in [',', ';', '|', '#']: if c in value: char = c break else: raise Exception for item in value.split(char): items.append(item.strip()) if len(items) < 2: raise Exception except: msg = _("Cannot set headers") tip = _("Please, define the name of the headers to be used as variables.") self.rootgui.show_dialog(_("Error"), msg, tip) else: try: index = items.index(self.options[CSVImport_CONFKEY_HEADER_ID]) except: index = 0 self._set_combo(self.cb_headerid, items, CSVImport_CONFKEY_HEADER_ID, index) self.rootplugin.update_headers(items) self._geolocate() def _set_combo(self, cb, items=[], key=None, active=None): self.events = False cb.get_model().clear() for item in items: cb.append_text(item) if active != None: self.options[key] = items[active] cb.set_active(active) self.events = True def _empty_combo(self, cb): cb.get_model().clear() def _combo_geolocate(self, widget, key): if self.events: header = widget.get_active_text() if header in self.options[CSVImport_CONFKEY_HEADERS]: self.options[key] = header else: self.options[key] = '' def _activate_combo(self, cb, key, value, no): counter = 0 for row in cb.get_model(): if value == row[0]: if row[0] == no: cb.set_active(0) self.options[key] = '' else: cb.set_active(counter) self.options[key] = row[0] break counter += 1 def _combo_id(self, widget): if self.events: header = widget.get_active_text() self.options[CSVImport_CONFKEY_HEADER_ID] = header header_lat = self.cb_lat.get_active_text() header_lon = self.cb_lon.get_active_text() header_ele = self.cb_ele.get_active_text() header_date = self.cb_date.get_active_text() self._geolocate() self._activate_combo(self.cb_lat, CSVImport_CONFKEY_HEADER_LAT, header_lat, header) self._activate_combo(self.cb_lon, CSVImport_CONFKEY_HEADER_LON, header_lon, header) self._activate_combo(self.cb_ele, CSVImport_CONFKEY_HEADER_ELE, header_ele, header) self._activate_combo(self.cb_date, CSVImport_CONFKEY_HEADER_DATE, header_date, header) def show(self, widget=None, options=None): if widget: widget.add(self.plugin) if options: self.setup(options) self.plugin.show_all() def hide(self, reset=False): self.plugin.hide_all() if reset: self.reset() def reset(self): self.button_process.set_sensitive(False) self.checkbutton_geolocate.set_sensitive(False) self.frame.set_sensitive(False) self._empty_combo(self.cb_headerid) self.cb_headerid.set_sensitive(False) self.options[CSVImport_CONFKEY_HEADER_ID] = '' self.entry_headers.set_sensitive(False) self.entry_headers.set_text('') image = self.button_addfile.get_image() image.set_from_stock(gtk.STOCK_ADD, gtk.ICON_SIZE_BUTTON) self.button_addfile.set_image(image) self.button_addfile.set_label(_("Select file")) self.checkbutton_geolocate.set_active(False) self.options[CSVImport_CONFKEY_FILENAME] = '' self.rootplugin.update_headers() self.userfacade.state.photovariables = self.rootplugin.photovariables_old self._geolocate() self.rootgui.reload_treeviewgeophotos() self.events = True def setup(self, options): self.options = options self.cb_date.set_tooltip_text(_("Date header name. Format should be: ") + self.options[CSVImport_CONFKEY_DATE_PARSE]) if options[CSVImport_CONFKEY_GEOLOCATE]: self.checkbutton_geolocate.set_active(True) filename = options[CSVImport_CONFKEY_FILENAME] if filename: self._set_csv(filename) #self.entry_headers.set_text(', '.join(options[CSVImport_CONFKEY_HEADERS])) #EOF
39.995842
218
0.615501
17,159
0.891933
0
0
0
0
0
0
2,542
0.132134
4b8fcf8f0fe4212ea52ae11e77f6cd66ebb3437f
9,024
py
Python
src/opt_utils.py
mateuszz0000/POSA
1295065251dd22c89d923fbff7d8bf4c53339d95
[ "CNRI-Python", "Xnet", "Info-ZIP", "X11" ]
71
2021-05-02T21:40:29.000Z
2022-03-30T03:52:01.000Z
src/opt_utils.py
mateuszz0000/POSA
1295065251dd22c89d923fbff7d8bf4c53339d95
[ "CNRI-Python", "Xnet", "Info-ZIP", "X11" ]
4
2021-06-18T06:31:29.000Z
2021-12-07T07:29:21.000Z
src/opt_utils.py
mateuszz0000/POSA
1295065251dd22c89d923fbff7d8bf4c53339d95
[ "CNRI-Python", "Xnet", "Info-ZIP", "X11" ]
10
2021-05-08T08:16:31.000Z
2022-02-17T04:40:30.000Z
# -*- coding: utf-8 -*- # Max-Planck-Gesellschaft zur Förderung der Wissenschaften e.V. (MPG) is # holder of all proprietary rights on this computer program. # You can only use this computer program if you have closed # a license agreement with MPG or you get the right to use the computer # program from someone who is authorized to grant you that right. # Any use of the computer program without a valid license is prohibited and # liable to prosecution. # # Copyright©2020 Max-Planck-Gesellschaft zur Förderung # der Wissenschaften e.V. (MPG). acting on behalf of its Max Planck Institute # for Intelligent Systems. All rights reserved. # # Contact: [email protected] import torch import torch.nn.functional as F import numpy as np import torchgeometry as tgm from src import misc_utils, eulerangles from tqdm import tqdm def compute_afford_loss(vertices=None, scene_data=None, gen_batch=None, pen_w=0.0, no_obj_classes=None, use_semantics=False, semantics_w=0.0, **kwargs): contact_ids = gen_batch[:, :, 0] > 0.5 x = misc_utils.read_sdf(vertices, scene_data['sdf'], scene_data['grid_dim'], scene_data['grid_min'], scene_data['grid_max'], mode="bilinear").squeeze() batch_size = vertices.shape[0] device = vertices.device contact_loss = torch.sum(x[contact_ids.flatten()] ** 2) pen_loss = torch.tensor(0.0) if pen_w > 0: mask = x.lt(0).flatten().int() + (~contact_ids.flatten()).int() x_neg = torch.abs(x[mask == 2]) if len(x_neg) == 0: pen_loss = torch.tensor(0.0) else: pen_loss = pen_w * x_neg.sum() semantics_loss = torch.tensor(0.0) if use_semantics: # Read semantics x_semantics = misc_utils.read_sdf(vertices, scene_data['semantics'], scene_data['grid_dim'], scene_data['grid_min'], scene_data['grid_max'], mode="bilinear").squeeze() x_semantics = contact_ids.flatten().float() * x_semantics.unsqueeze(0) x_semantics = torch.zeros(x_semantics.shape[0], x_semantics.shape[1], no_obj_classes, device=device).scatter_( -1, x_semantics.unsqueeze(-1).type(torch.long), 1.) # Compute loss targets = gen_batch[:, :, 1:].argmax(dim=-1).type(torch.long).reshape(batch_size, -1) semantics_loss = semantics_w * F.cross_entropy(x_semantics.permute(0, 2, 1), targets, reduction='sum') return contact_loss, pen_loss, semantics_loss def eval_init_points(init_pos=None, init_ang=None, vertices=None, scene_data=None, gen_batch=None, **kwargs): with torch.no_grad(): losses = [] init_pos_batches = init_pos.split(1) for i in tqdm(range(len(init_pos_batches))): curr_init_pos = init_pos_batches[i] rot_aa = torch.cat((torch.zeros((1, 2), device=vertices.device), init_ang[i].reshape(1, 1)), dim=1) rot_mat = tgm.angle_axis_to_rotation_matrix(rot_aa.reshape(-1, 3))[:, :3, :3] curr_vertices = torch.bmm(rot_mat, vertices.permute(0, 2, 1)).permute(0, 2, 1) curr_vertices = curr_vertices + curr_init_pos contact_loss, pen_loss, semantics_loss = compute_afford_loss(vertices=curr_vertices, scene_data=scene_data, gen_batch=gen_batch, **kwargs) loss = contact_loss + pen_loss + semantics_loss losses.append(loss.item()) # Sort initial positions and orientations from best to wrost losses = np.array(losses) ids = np.argsort(losses) losses = losses[ids] init_pos = init_pos[ids] init_ang = init_ang[ids] return losses, init_pos, init_ang def init_points_culling(init_pos=None, vertices=None, scene_data=None, gen_batch=None, max_init_points=50, **kwargs): init_ang = [] angles = torch.arange(0, 2 * np.pi, np.pi / 2, device=vertices.device) angles[0] = 1e-9 for ang in angles: init_ang.append(ang * torch.ones(init_pos.shape[0], 1, device=vertices.device)) init_ang = torch.cat(init_ang).to(init_pos.device) init_pos = init_pos.repeat(angles.shape[0], 1, 1) # Shuffle rnd_ids = np.random.choice(init_pos.shape[0], init_pos.shape[0], replace=False) init_pos = init_pos[rnd_ids, :] init_ang = init_ang[rnd_ids, :] losses, init_pos, init_ang = eval_init_points(init_pos=init_pos, init_ang=init_ang, vertices=vertices.unsqueeze(0), scene_data=scene_data, gen_batch=gen_batch, **kwargs) # Select only a subset from initial points for optimization if init_pos.shape[0] > max_init_points: init_pos = init_pos[:max_init_points] init_ang = init_ang[:max_init_points] return init_pos, init_ang class opt_wrapper(object): def __init__(self, vertices=None, vertices_can=None, pelvis=None, scene_data=None, down_sample_fn=None, down_sample_fn2=None, device=None, dtype=None, pen_w=None, use_semantics=None, no_obj_classes=None, nv=None, optimizer=None, gen_batch=None, body_model=None, opt_pose=False, semantics_w=None, init_body_pose=None, pose_w=None, **kwargs): self.optimizer = optimizer self.vertices = vertices self.vertices_can = vertices_can self.pelvis = pelvis self.scene_data = scene_data self.down_sample_fn = down_sample_fn self.down_sample_fn2 = down_sample_fn2 self.device = device self.dtype = dtype self.pen_w = pen_w self.pose_w = pose_w self.semantics_w = semantics_w self.use_semantics = use_semantics self.no_obj_classes = no_obj_classes self.nv = nv self.gen_batch = gen_batch self.opt_pose = opt_pose self.body_model = body_model self.init_body_pose = init_body_pose self.R_smpl2scene = torch.tensor(eulerangles.euler2mat(np.pi / 2, 0, 0, 'sxyz'), dtype=dtype, device=device) def compute_vertices(self, t_free, y_ang, vertices=None, down_sample=True): curr_batch_size = self.vertices.shape[0] rot_aa = torch.cat((torch.zeros((curr_batch_size, 2), device=self.device), y_ang), dim=1) rot_mat = tgm.angle_axis_to_rotation_matrix(rot_aa.reshape(-1, 3))[:, :3, :3] if self.opt_pose: body_model_output = self.body_model(return_verts=True) pelvis = body_model_output.joints[:, 0, :].reshape(1, 3) vertices_local = body_model_output.vertices.squeeze() vertices_local = torch.matmul(self.R_smpl2scene, (vertices_local - pelvis).t()).t() vertices_local.unsqueeze_(0) if down_sample: vertices_local = self.down_sample_fn.forward(vertices_local.permute(0, 2, 1)) vertices_local = self.down_sample_fn2.forward(vertices_local).permute(0, 2, 1) vertices_local = torch.bmm(rot_mat, vertices_local.permute(0, 2, 1)).permute(0, 2, 1) vertices_local += t_free else: # very important to make a local copy, so that you don't change the original variable if vertices is None: vertices_local = torch.bmm(rot_mat, self.vertices.permute(0, 2, 1)).permute(0, 2, 1) else: vertices_local = torch.bmm(rot_mat, vertices.permute(0, 2, 1)).permute(0, 2, 1) vertices_local += t_free return vertices_local, rot_mat def compute_loss(self, t_free, y_ang): pose_loss = torch.tensor(0.0) if self.opt_pose: pose_loss = self.pose_w * F.mse_loss(self.body_model.body_pose, self.init_body_pose) vertices_local, rot_mat = self.compute_vertices(t_free, y_ang) contact_loss, pen_loss, semantic_loss = compute_afford_loss(vertices=vertices_local, scene_data=self.scene_data, gen_batch=self.gen_batch, pen_w=self.pen_w, no_obj_classes=self.no_obj_classes, use_semantics=self.use_semantics, semantics_w=self.semantics_w) return contact_loss, pen_loss, pose_loss, semantic_loss def create_fitting_closure(self, t_free, y_ang): def fitting_func(): self.optimizer.zero_grad() recon_loss, pen_loss, pose_loss, semantic_loss = self.compute_loss(t_free, y_ang) loss_total = recon_loss + pen_loss + pose_loss + semantic_loss loss_total.backward(retain_graph=True) return loss_total return fitting_func
47
120
0.625332
3,983
0.441232
0
0
0
0
0
0
1,020
0.112994
4b90f733d945576384389e3af5e8eb7b26b24785
137
py
Python
gan_provider.py
jiameng1010/pointNet
17d230f46f64136baba2c3d6cb7f05ab4bbb9f31
[ "MIT" ]
null
null
null
gan_provider.py
jiameng1010/pointNet
17d230f46f64136baba2c3d6cb7f05ab4bbb9f31
[ "MIT" ]
null
null
null
gan_provider.py
jiameng1010/pointNet
17d230f46f64136baba2c3d6cb7f05ab4bbb9f31
[ "MIT" ]
1
2019-02-03T12:19:36.000Z
2019-02-03T12:19:36.000Z
from tensorflow.contrib.slim.python.slim.data import data_provider from tensorflow.contrib.slim.python.slim.data import parallel_reader
34.25
68
0.861314
0
0
0
0
0
0
0
0
0
0
4b91ba97fda9b2ee93796afb30a9ecc697c21159
1,205
py
Python
script.module.placenta/lib/resources/lib/modules/thexem.py
parser4life/tantrumrepo
3b37145f4772409e538cbddb0b7aa23be525772a
[ "Beerware" ]
1
2021-05-09T19:55:51.000Z
2021-05-09T19:55:51.000Z
script.module.placenta/lib/resources/lib/modules/thexem.py
parser4life/tantrumrepo
3b37145f4772409e538cbddb0b7aa23be525772a
[ "Beerware" ]
null
null
null
script.module.placenta/lib/resources/lib/modules/thexem.py
parser4life/tantrumrepo
3b37145f4772409e538cbddb0b7aa23be525772a
[ "Beerware" ]
2
2020-04-01T22:11:12.000Z
2020-05-07T23:54:52.000Z
# -*- coding: UTF-8 -*- ####################################################################### # ---------------------------------------------------------------------------- # "THE BEER-WARE LICENSE" (Revision 42): # @tantrumdev wrote this file. As long as you retain this notice you # can do whatever you want with this stuff. If we meet some day, and you think # this stuff is worth it, you can buy me a beer in return. - Muad'Dib # ---------------------------------------------------------------------------- ####################################################################### # Addon Name: Placenta # Addon id: plugin.video.placenta # Addon Provider: MuadDib import json from resources.lib.modules import client URL_PATTERN = 'http://thexem.de/map/single?id=%s&origin=tvdb&season=%s&episode=%s&destination=scene' def get_scene_episode_number(tvdbid, season, episode): try: url = URL_PATTERN % (tvdbid, season, episode) r = client.request(url) r = json.loads(r) if r['result'] == 'success': data = r['data']['scene'] return data['season'], data['episode'] except: pass return season, episode
36.515152
100
0.480498
0
0
0
0
0
0
0
0
790
0.655602
4b92d579f1edf869213b966f3c57e11cb659219d
1,048
py
Python
Day18/main.py
MHKomeili/100DaysofCode
a5799011a43f777ddc5ac9e649aa27291313b62b
[ "MIT" ]
null
null
null
Day18/main.py
MHKomeili/100DaysofCode
a5799011a43f777ddc5ac9e649aa27291313b62b
[ "MIT" ]
null
null
null
Day18/main.py
MHKomeili/100DaysofCode
a5799011a43f777ddc5ac9e649aa27291313b62b
[ "MIT" ]
null
null
null
# import colorgram # # colors = colorgram.extract('image.jpg', 30) # rgb_colors = [] # for color in colors: # rgb_colors.append((color.rgb.r, color.rgb.g, color.rgb.b)) # # print(rgb_colors) from turtle import Turtle, Screen import random color_list = [(238, 251, 245), (250, 228, 15), (213, 12, 8), (199, 11, 36), (10, 98, 61), (5, 39, 32), (232, 228, 5), (64, 221, 157), (198, 68, 19), (32, 91, 189), (43, 212, 71), (235, 148, 38), (32, 30, 153), (242, 247, 251), (15, 22, 54), (67, 9, 49), (245, 38, 148), (14, 206, 222), (65, 203, 230), (62, 20, 10), (229, 164, 7), (226, 19, 111), (14, 154, 22), (246, 58, 14), (98, 75, 8), (248, 11, 9), (223, 140, 205), (66, 241, 160), ] tim = Turtle() scr = Screen() scr.colormode(255) tim.penup() tim.hideturtle() tim.setposition(-300, -300) for i in range(10): tim.setposition(-300, tim.ycor() + 50) for j in range(10): tim.setx(tim.xcor() + 50) tim.dot(20, random.choice(color_list)) scr.exitonclick()
31.757576
119
0.540076
0
0
0
0
0
0
0
0
187
0.178435
4b9490ebcc233667c0f331f949a3dfce27be8b1f
8,723
py
Python
hirebob/portal/forms.py
shantanub0/hirebob
5a55e97c6e220059964fbb55439b0189abae1307
[ "MIT" ]
null
null
null
hirebob/portal/forms.py
shantanub0/hirebob
5a55e97c6e220059964fbb55439b0189abae1307
[ "MIT" ]
1
2018-06-23T01:20:26.000Z
2018-06-25T21:49:17.000Z
hirebob/portal/forms.py
shantanub0/hirebob
5a55e97c6e220059964fbb55439b0189abae1307
[ "MIT" ]
1
2018-06-14T12:11:59.000Z
2018-06-14T12:11:59.000Z
from django import forms from .models import UserAccount, JobPost, JobPostActivity, UserProfile class FormUserCreation(forms.ModelForm): UserTypes = ((1, 'Applicants'), (2, 'Organisations')) user_type = forms.ChoiceField(choices=UserTypes, widget=forms.Select(attrs={'class': "form-control"})) user_full_name = forms.CharField(max_length=100, widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Full Name'})) email = forms.EmailField(max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Email ID'})) password = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'type': 'password', 'placeholder': 'Enter Password', 'minlength': '6', 'onkeyup': 'check();'})) confirm_password = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'type': 'password', 'placeholder': 'Re-enter Password', 'minlength': '6', 'onkeyup': 'check();'})) class Meta: model = UserAccount fields = ('user_type', 'user_full_name', 'email', 'password') class FormLogin(forms.ModelForm): email = forms.EmailField(max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Email ID'})) password = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'type': 'password', 'placeholder': 'Enter Password', 'minlength': '6'})) class Meta: model = UserAccount fields = ('email', 'password') class FormJobPost(forms.ModelForm): Locations = (('Mumbai', 'Mumbai'), ('Navi Mumbai', 'Navi Mumbai'), ('Pune', 'Pune')) job_types = (('Software Engineer', 'Software Engineer'), ('Database Admin', 'Database Admin'), ('DevOps', 'DevOps')) jobs_skill = (('Java', 'Java'), ('Python', 'Python'), ('C', 'C'), ('C++', 'C++')) job_location = forms.ChoiceField(choices=Locations, widget=forms.Select(attrs={'class': "form-control"})) job_type = forms.ChoiceField(choices=job_types, widget=forms.Select(attrs={'class': "form-control"})) job_skills = forms.ChoiceField(choices=jobs_skill, widget=forms.Select(attrs={'class': "form-control"})) job_title = forms.CharField(max_length=100, required=True, widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter job title'})) posted_by_email = forms.EmailField(max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Email ID', 'readonly': True})) job_description = forms.CharField(widget=forms.Textarea(attrs={'class': "form-control", 'placeholder': 'Enter Job Description'})) class Meta: model = JobPost fields = ('job_type', 'job_skills', 'job_location', 'posted_by_email', 'job_description', 'job_title') class FormApply(forms.ModelForm): email = forms.EmailField(required=True, max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Email ID', 'readonly': True})) to_email = forms.EmailField(required=True, max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Email ID', 'readonly': True})) cover_letter = forms.CharField(required=True, widget=forms.Textarea(attrs={'class': "form-control", 'placeholder': 'Cover Letter'})) post_id = forms.IntegerField(required=True, widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Post ID', 'readonly': True})) job_title = forms.CharField(required=True, widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Job Title'})) class Meta: model = JobPostActivity fields = ('email', 'post_id') class FormUploadImage(forms.Form): user_image = forms.ImageField(widget=forms.FileInput()) class Meta: model = UserAccount fields = ('user_image', ) class FormUploadResume(forms.Form): resume = forms.FileField() class Meta: model = UserAccount fields = ('resume', ) class FormApplicantsInfo(forms.Form): Gender = (('Male', 'Male'), ('Female', 'Female'), ('None', 'None')) gender = forms.ChoiceField(choices=Gender, widget=forms.Select(attrs={'class': "form-control"})) email = forms.EmailField(max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Email ID', 'readonly': True})) gmail = forms.EmailField(max_length=250, help_text="Required. Invalid format", widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter gmail id'})) linkedin = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Linkedin profile'})) skype_id = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter skype id'})) about_me = forms.CharField(widget=forms.Textarea(attrs={'class': "form-control", 'placeholder': 'Enter About you'})) address = forms.CharField(widget=forms.Textarea(attrs={'class': "form-control", 'placeholder': 'Enter your address'})) birthday = forms.DateField(widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter DOB in DD-MM-YYYY'})) job_title = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Job Title'})) location = forms.CharField(widget=forms.TextInput(attrs={'class': "form-control", 'placeholder': 'Enter Your location'})) class Meta: model = UserProfile fields = ('email', 'gmail', 'linkedin', 'skype_id', 'about_me', 'address', 'birthday', 'job_title', 'location', 'gender')
56.642857
129
0.458558
8,606
0.986587
0
0
0
0
0
0
2,168
0.248538
4b95d82a263834a4e169c435b74dfded71be2e85
5,538
py
Python
siemstress/trigger.py
dogoncouch/siemstress
be7f60bb0228a886d48deb4f46309be7fb8aa0af
[ "MIT" ]
28
2017-08-14T12:41:56.000Z
2022-02-18T01:18:11.000Z
siemstress/trigger.py
dogoncouch/siemstress
be7f60bb0228a886d48deb4f46309be7fb8aa0af
[ "MIT" ]
1
2017-08-23T10:47:16.000Z
2017-08-24T18:52:48.000Z
siemstress/trigger.py
dogoncouch/siemstress
be7f60bb0228a886d48deb4f46309be7fb8aa0af
[ "MIT" ]
6
2018-01-07T11:42:18.000Z
2020-06-08T00:04:57.000Z
#!/usr/bin/env python #_MIT License #_ #_Copyright (c) 2017 Dan Persons ([email protected]) #_ #_Permission is hereby granted, free of charge, to any person obtaining a copy #_of this software and associated documentation files (the "Software"), to deal #_in the Software without restriction, including without limitation the rights #_to use, copy, modify, merge, publish, distribute, sublicense, and/or sell #_copies of the Software, and to permit persons to whom the Software is #_furnished to do so, subject to the following conditions: #_ #_The above copyright notice and this permission notice shall be included in all #_copies or substantial portions of the Software. #_ #_THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR #_IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, #_FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE #_AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER #_LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, #_OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE #_SOFTWARE. import time from time import strftime from time import sleep from time import daylight from time import timezone from time import altzone from random import randrange from datetime import datetime import MySQLdb as mdb import json import threading import os from sys import exit import siemstress.manage #import signal class SiemTrigger: def __init__(self, db, rule): """Initialize trigger object""" self.db = db self.rule = rule self.tzone = None def watch_rule(self): """Watch a trigger rule""" # Set time zone: if daylight: self.tzone = \ str(int(float(altzone) / 60 // 60)).rjust(2, '0') + \ str(int(float(altzone) / 60 % 60)).ljust(2, '0') else: self.tzone = \ str(int(float(timezone) / 60 // 60)).rjust(2, '0') + \ str(int(float(timezone) / 60 % 60)).ljust(2, '0') if not '-' in self.tzone: self.tzone = '+' + self.tzone while True: # Check the rule: self.check_rule() # Wait until the next interval sleep(int(self.rule['time_int']) * 60) def check_rule(self): """Check a trigger rule""" # To Do: Add date_stamp_utc/int logic if not self.tzone: # Set time zone: if time.localtime().tm_isdst: self.tzone = \ str(int(float(altzone) / 60 // 60)).rjust(2, '0') + \ str(int(float(altzone) / 60 % 60)).ljust(2, '0') else: self.tzone = \ str(int(float(timezone) / 60 // 60)).rjust(2, '0') + \ str(int(float(timezone) / 60 % 60)).ljust(2, '0') if not '-' in self.tzone: self.tzone = '+' + self.tzone # Query the database: con = mdb.connect(self.db['host'], self.db['user'], self.db['password'], self.db['database']) with con: cur = con.cursor() cur.execute(self.rule['sql_query']) rows = cur.fetchall() cur.close() con.close() # Evaluate the results: if len(rows) > int(self.rule['event_limit']): idtags = json.dumps([int(row[0]) for row in rows]) datestamp = datetime.now().strftime('%Y%m%d%H%M%S') datestamputc = datetime.utcnow().strftime('%Y%m%d%H%M%S') magnitude = (((len(rows) // 2) // \ (self.rule['event_limit'] + 1) // 2) + 5) * \ ( 7 - self.rule['severity']) outstatement = 'INSERT INTO ' + \ self.rule['out_table'] + \ '(date_stamp, date_stamp_utc, t_zone, ' + \ 'source_rule, severity, source_table, event_limit, ' + \ 'event_count, magnitude, time_int, message, source_ids) ' + \ 'VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s)' # Send an event to the database: con = mdb.connect(self.db['host'], self.db['user'], self.db['password'], self.db['database']) with con: cur = con.cursor() cur.execute(outstatement, (datestamp, datestamputc, self.tzone, self.rule['rule_name'], self.rule['severity'], self.rule['source_table'], self.rule['event_limit'], len(rows), magnitude, self.rule['time_int'], self.rule['message'], idtags)) cur.close() con.close() def start_rule(db, rule, oneshot): """Initialize trigger object and start watching""" # Make sure the table exists: siemstress.manage.create_ruleevent_table(rule['out_table']) sentry = SiemTrigger(db, rule) if oneshot: sentry.check_rule() elif int(rule['time_int']) == 0: pass else: # Before starting, sleep randomly up to rule interval to stagger # database use: sleep(randrange(0, int(rule['time_int']) * 60)) sentry.watch_rule()
34.397516
81
0.548754
3,536
0.638498
0
0
0
0
0
0
2,094
0.378115
4b996a561c6739777af3fa1902cca7e146f0eeaf
687
py
Python
TianJiPlanBackend/authentication/migrations/0002_auto_20210912_0929.py
weridolin/tianji-plan
b98a49d92ee2a365095f9e15f4231f5178aca1c0
[ "Apache-2.0" ]
null
null
null
TianJiPlanBackend/authentication/migrations/0002_auto_20210912_0929.py
weridolin/tianji-plan
b98a49d92ee2a365095f9e15f4231f5178aca1c0
[ "Apache-2.0" ]
null
null
null
TianJiPlanBackend/authentication/migrations/0002_auto_20210912_0929.py
weridolin/tianji-plan
b98a49d92ee2a365095f9e15f4231f5178aca1c0
[ "Apache-2.0" ]
1
2021-12-07T11:45:13.000Z
2021-12-07T11:45:13.000Z
# Generated by Django 3.2.7 on 2021-09-12 01:29 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('authentication', '0001_initial'), ] operations = [ migrations.AddField( model_name='userprofile', name='mail', field=models.EmailField(blank=True, db_index=True, max_length=127, null=True, unique=True, verbose_name='电话'), ), migrations.AlterField( model_name='userprofile', name='telephone', field=models.CharField(blank=True, db_index=True, max_length=127, null=True, unique=True, verbose_name='电话'), ), ]
28.625
122
0.615721
602
0.866187
0
0
0
0
0
0
136
0.195683
4b9aca9719a2480581a602385b8fda1e00bcfadc
3,040
py
Python
ooobuild/lo/util/time_with_timezone.py
Amourspirit/ooo_uno_tmpl
64e0c86fd68f24794acc22d63d8d32ae05dd12b8
[ "Apache-2.0" ]
null
null
null
ooobuild/lo/util/time_with_timezone.py
Amourspirit/ooo_uno_tmpl
64e0c86fd68f24794acc22d63d8d32ae05dd12b8
[ "Apache-2.0" ]
null
null
null
ooobuild/lo/util/time_with_timezone.py
Amourspirit/ooo_uno_tmpl
64e0c86fd68f24794acc22d63d8d32ae05dd12b8
[ "Apache-2.0" ]
null
null
null
# coding: utf-8 # # Copyright 2022 :Barry-Thomas-Paul: Moss # # Licensed under the Apache License, Version 2.0 (the "License") # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http: // www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Struct Class # this is a auto generated file generated by Cheetah # Namespace: com.sun.star.util # Libre Office Version: 7.3 from ooo.oenv.env_const import UNO_NONE import typing from .time import Time as Time_604e0855 class TimeWithTimezone(object): """ Struct Class represents a combined time value with time zone. **since** LibreOffice 4.1 See Also: `API TimeWithTimezone <https://api.libreoffice.org/docs/idl/ref/structcom_1_1sun_1_1star_1_1util_1_1TimeWithTimezone.html>`_ """ __ooo_ns__: str = 'com.sun.star.util' __ooo_full_ns__: str = 'com.sun.star.util.TimeWithTimezone' __ooo_type_name__: str = 'struct' typeName: str = 'com.sun.star.util.TimeWithTimezone' """Literal Constant ``com.sun.star.util.TimeWithTimezone``""" def __init__(self, TimeInTZ: typing.Optional[Time_604e0855] = UNO_NONE, Timezone: typing.Optional[int] = 0) -> None: """ Constructor Arguments: TimeInTZ (Time, optional): TimeInTZ value. Timezone (int, optional): Timezone value. """ super().__init__() if isinstance(TimeInTZ, TimeWithTimezone): oth: TimeWithTimezone = TimeInTZ self.TimeInTZ = oth.TimeInTZ self.Timezone = oth.Timezone return kargs = { "TimeInTZ": TimeInTZ, "Timezone": Timezone, } if kargs["TimeInTZ"] is UNO_NONE: kargs["TimeInTZ"] = None self._init(**kargs) def _init(self, **kwargs) -> None: self._time_in_tz = kwargs["TimeInTZ"] self._timezone = kwargs["Timezone"] @property def TimeInTZ(self) -> Time_604e0855: """ the time (in TimeZone) """ return self._time_in_tz @TimeInTZ.setter def TimeInTZ(self, value: Time_604e0855) -> None: self._time_in_tz = value @property def Timezone(self) -> int: """ contains the time zone, as signed offset in minutes from UTC, that is east of UTC, that is the amount of minutes that should be added to UTC time to obtain the time in that timezone. To obtain UTC time from TimeInTZ, you need to subtract TimeZone minutes. """ return self._timezone @Timezone.setter def Timezone(self, value: int) -> None: self._timezone = value __all__ = ['TimeWithTimezone']
30.4
190
0.650329
2,174
0.715132
0
0
706
0.232237
0
0
1,738
0.571711
4b9af91c0efeb81facf6d27474553a4bb9a6505d
2,025
py
Python
tests/unit_tests/tasks/fortran/test_fortran_compiler.py
bblay/fab
bbdac7bae20c5b8695a2d56945c9593b4fda9c74
[ "BSD-3-Clause" ]
null
null
null
tests/unit_tests/tasks/fortran/test_fortran_compiler.py
bblay/fab
bbdac7bae20c5b8695a2d56945c9593b4fda9c74
[ "BSD-3-Clause" ]
null
null
null
tests/unit_tests/tasks/fortran/test_fortran_compiler.py
bblay/fab
bbdac7bae20c5b8695a2d56945c9593b4fda9c74
[ "BSD-3-Clause" ]
null
null
null
from pathlib import Path from unittest import mock import pytest from fab.build_config import AddFlags from fab.dep_tree import AnalysedFile from fab.steps.compile_fortran import CompileFortran # todo: we might have liked to reuse this from test_dep_tree from fab.util import CompiledFile @pytest.fixture def src_tree(): return { Path('src/foo.f90'): AnalysedFile(fpath=Path('src/foo.f90'), file_hash=None), Path('src/root.f90'): AnalysedFile( fpath=Path('src/root.f90'), file_deps={Path('src/a.f90'), Path('src/b.f90')}, file_hash=None), Path('src/a.f90'): AnalysedFile(fpath=Path('src/a.f90'), file_deps={Path('src/c.f90')}, file_hash=None), Path('src/b.f90'): AnalysedFile(fpath=Path('src/b.f90'), file_deps={Path('src/c.f90')}, file_hash=None), Path('src/c.f90'): AnalysedFile(fpath=Path('src/c.f90'), file_deps=set(), file_hash=None), } class Test_run(object): # todo: almost identical to the c compiler test def test_vanilla(self, src_tree): # ensure the compile passes match the build tree config = mock.Mock(workspace=Path('foo/src'), multiprocessing=False) c_compiler = CompileFortran( compiler='gcc', common_flags=['-c'], path_flags=[AddFlags(match='foo/src/*', flags=['-Dhello'])]) def foo(items, func): return [CompiledFile(af, output_fpath=None) for af in items] with mock.patch('fab.steps.Step.run_mp', side_effect=foo) as mock_run_mp: c_compiler.run(artefact_store={'build_tree': src_tree}, config=config) # 1st pass mock_run_mp.assert_any_call( items={src_tree[Path('src/foo.f90')], src_tree[Path('src/c.f90')]}, func=mock.ANY) # 2nd pass mock_run_mp.assert_any_call( items={src_tree[Path('src/a.f90')], src_tree[Path('src/b.f90')]}, func=mock.ANY) # last pass mock_run_mp.assert_called_with(items={src_tree[Path('src/root.f90')]}, func=mock.ANY)
38.942308
112
0.651852
1,116
0.551111
0
0
612
0.302222
0
0
483
0.238519
4b9cdb57c833e7e628efc0c75d61d7090e29a276
393
py
Python
exercicios/Lista6/Q5.py
AlexandrePeBrito/CursoUdemyPython
3de58cb30c9f333b32078309847179ff3f9d7e22
[ "MIT" ]
null
null
null
exercicios/Lista6/Q5.py
AlexandrePeBrito/CursoUdemyPython
3de58cb30c9f333b32078309847179ff3f9d7e22
[ "MIT" ]
null
null
null
exercicios/Lista6/Q5.py
AlexandrePeBrito/CursoUdemyPython
3de58cb30c9f333b32078309847179ff3f9d7e22
[ "MIT" ]
null
null
null
""" 5. Faça um programa que receba do usuário um arquivo texto e um caracter. Mostre na tela quantas vezes aquele caractere ocorre dentro do arquivo. """ arquivo=open('CursoUdemyPython/exercicios/Lista6/arq.txt') texto=arquivo.read() carac=input('Informe um caractere: ') ca=0 for c in texto: if(c == carac): ca+=1 arquivo.close() print(f"Foi identificado {ca} deste caractere")
28.071429
92
0.725191
0
0
0
0
0
0
0
0
264
0.668354
4b9e62db340ea51b4cda5971027dcd23a1f17c3d
3,704
py
Python
superset_config.py
mikiec84/incubator-superset
3a1c32ae2378902a26873113d98bd55d290233ca
[ "Apache-2.0" ]
1
2020-08-07T16:30:54.000Z
2020-08-07T16:30:54.000Z
superset_config.py
mikiec84/incubator-superset
3a1c32ae2378902a26873113d98bd55d290233ca
[ "Apache-2.0" ]
null
null
null
superset_config.py
mikiec84/incubator-superset
3a1c32ae2378902a26873113d98bd55d290233ca
[ "Apache-2.0" ]
1
2020-08-07T16:30:58.000Z
2020-08-07T16:30:58.000Z
#--------------------------------------------------------- # Superset specific config #--------------------------------------------------------- ROW_LIMIT = 5000 SUPERSET_WEBSERVER_PORT = 8088 #--------------------------------------------------------- #--------------------------------------------------------- # Flask App Builder configuration #--------------------------------------------------------- # Your App secret key SECRET_KEY = '\2\1ulan123456\1\2\e\y\y\h' # The SQLAlchemy connection string to your database backend # This connection defines the path to the database that stores your # superset metadata (slices, connections, tables, dashboards, ...). # Note that the connection information to connect to the datasources # you want to explore are managed directly in the web UI #SQLALCHEMY_DATABASE_URI = 'postgresql+psycopg2://superset:superset@localhost:5432/superset' # ------------------------------ # GLOBALS FOR APP Builder # ------------------------------ # Uncomment to setup Your App name APP_NAME = 'Insights' # Uncomment to setup an App icon APP_ICON = '/static/assets/images/qmatic_insights-logo.png' # Extract and use X-Forwarded-For/X-Forwarded-Proto headers? ENABLE_PROXY_FIX = True ENABLE_JAVASCRIPT_CONTROLS = True ''' import os from flask_appbuilder.security.manager import AUTH_OID, AUTH_REMOTE_USER, AUTH_DB, AUTH_LDAP, AUTH_OAUTH basedir = os.path.abspath(os.path.dirname(__file__)) SUPERSET_WORKERS = 8 CSRF_ENABLED = True AUTH_TYPE = AUTH_OAUTH AUTH_USER_REGISTRATION = False AUTH_USER_REGISTRATION_ROLE = "Gamma" #"Public" OAUTH_PROVIDERS = [ { 'name': 'google', 'icon': 'fa-google', 'token_key': 'access_token', 'remote_app': { 'base_url': 'https://www.googleapis.com/oauth2/v2/', 'request_token_params': { 'scope': 'email profile' }, 'request_token_url': None, 'access_token_url': 'https://accounts.google.com/o/oauth2/token', 'authorize_url': 'https://accounts.google.com/o/oauth2/auth', 'consumer_key': '996225546131-1qd2alfrrp1scf6gvkeg63mg2ku85lka.apps.googleusercontent.com', 'consumer_secret': '3fxwT-a8YA1akyuUYFfakMCz' } }, { 'name': 'slatest.qmaticcloud.com', 'icon': 'fa-google', 'token_key': 'access_token', 'remote_app': { #'base_url': 'https://slatest.qmaticcloud.com/oauth2server/oauth/', 'base_url': None, 'request_token_params': { 'scope': 'user_info', 'state': '123' }, 'request_token_url': None, 'access_token_url': 'https://slatest.qmaticcloud.com/oauth2server/oauth/token', 'authorize_url': 'https://slatest.qmaticcloud.com/oauth2server/oauth/authorize', 'consumer_key': 'businessintelligence', 'consumer_secret': 'fSmI0K1uSvnORBk3' } }, { 'name': 'msdemo.qmatic.cloud', 'icon': 'fa-google', 'token_key': 'access_token', 'remote_app': { 'base_url': None, 'request_token_params': { 'scope': 'user_info', 'state': '123' }, 'request_token_url': None, 'access_token_url': 'https://msdemo.qmatic.cloud/oauth2server/oauth/token', 'authorize_url': 'https://msdemo.qmatic.cloud/oauth2server/oauth/authorize', 'consumer_key': 'businessintelligence', 'consumer_secret': 'fSmI0K1uSvnORBk3' } } ] '''
35.615385
107
0.551836
0
0
0
0
0
0
0
0
3,529
0.952754
4b9ee2812f3c3d983291b0a7f5a83dcf6f853ee4
5,038
py
Python
python_code/yolo/extract_car_num.py
mukulbhave/tensorflow
848b16fa32cd0f180ab80a98254edd2147ea3948
[ "CNRI-Python" ]
null
null
null
python_code/yolo/extract_car_num.py
mukulbhave/tensorflow
848b16fa32cd0f180ab80a98254edd2147ea3948
[ "CNRI-Python" ]
null
null
null
python_code/yolo/extract_car_num.py
mukulbhave/tensorflow
848b16fa32cd0f180ab80a98254edd2147ea3948
[ "CNRI-Python" ]
null
null
null
import argparse import cv2 import re import numpy as np import string import PIL import os,glob import ntpath import time import matplotlib.pyplot as plt from PIL import Image from yad2k.models.keras_yolo import (preprocess_true_boxes, yolo_body, yolo_eval, yolo_head, yolo_loss) from yad2k.utils.draw_boxes import draw_boxes from retrain_yolo import (create_model,get_classes) import keras.backend as K from crnn.train_crnn import create_crnn_model from crnn.crnn_data_gen import * char_list = string.ascii_letters+string.digits YOLO_ANCHORS = np.array( ((0.57273, 0.677385), (1.87446, 2.06253), (3.33843, 5.47434), (7.88282, 3.52778), (9.77052, 9.16828))) class_names=['plate','no-plate'] class CarNumberDetector: def __init__(self,viden_yolo_weights_path,viden_crnn_weights_path,classes_path,out_path): self.act_model = create_crnn_model(train=False) self.act_model.load_weights(viden_crnn_weights_path)# 'viden_trained_models\\viden_crnn_14May2021.hdf5') class_names = get_classes(classes_path) #print(class_names) self.out_path=out_path #self.anchors = YOLO_ANCHORS self.yolo_model_body, self.yolo_model = create_model(YOLO_ANCHORS, class_names,load_pretrained=False,freeze_body=False) self.yolo_model_body.load_weights(viden_yolo_weights_path)#'viden_trained_models\\viden_yolo_14May2021.h5') self.yolo_outputs = yolo_head(self.yolo_model_body.output, YOLO_ANCHORS, len(class_names)) self.yolo_input_image_shape = K.placeholder(shape=(2, )) self.boxes, self.scores, self.classes = yolo_eval( self.yolo_outputs, self.yolo_input_image_shape,max_boxes=1, score_threshold=.7, iou_threshold=0.5) def extract_number(self,orig_image_url=None,image_array=None,save=False): """ This is the primary method to detect number plate on car and fetch the number. image_array is the numpy array representing original car image method returns numpy array of image with bounding box ad the extracted car_number string """ if( image_array is None and orig_image_url is None): raise ValueError("image array or url is required") if(orig_image_url is not None): image = PIL.Image.open(orig_image_url) else: image = PIL.Image.fromarray(image_array) pred_boxes,pred_box_classes,img_with_boxes = self.get_bounding_boxes(image) pred_txt='' for i, box in list(enumerate(pred_boxes)): box_class = class_names[pred_box_classes[i]] top, left, bottom, right = box pred_obj_img = image.crop((left,top,right,bottom)) pred_txt=self.get_text(pred_obj_img) # Save the image: if save: time_param= int(round(time.time() * 1000)) orig_img_name = self.out_path+pred_txt+"-"+str(time_param)+".jpg" orig_image = PIL.Image.fromarray(img_with_boxes) orig_image.save(orig_img_name) return (orig_image,pred_txt) def get_text(self,pil_image): img = pil_image.resize((128, 32), Image.BICUBIC) img = np.array(img) /255; img = np.sum(img, axis=2,keepdims=True) img = np.expand_dims(img , axis = 0) prediction = self.act_model.predict(img) # use CTC decoder out = K.get_value(K.ctc_decode(prediction, input_length=np.ones(prediction.shape[0])*prediction.shape[1],greedy=False)[0][0]) x = out[0] le= min(10,out.shape[1]) s='' for x in out: for p in range(0, le): if int(x[p]) != -1: s += char_list[int(x[p])] return s def get_bounding_boxes(self,image): image_shape = (416, 416) resized_image =image.resize(tuple(image_shape), PIL.Image.BICUBIC) image_data = np.array(resized_image, dtype='float32') image_data /= 255. image_data = np.expand_dims(image_data, 0) sess = K.get_session() out_boxes, out_scores, out_classes = sess.run( [self.boxes, self.scores, self.classes], feed_dict={ self.yolo_model_body.input: image_data, self.yolo_input_image_shape: [image_data.shape[1], image_data.shape[2]], K.learning_phase(): 0 }) # Convert pred on 416 to actual image size resized_boxes = out_boxes/416 w,h = image.size box_resize_dim = [h,w,h,w] resized_boxes = resized_boxes * box_resize_dim orig_image_data = np.array(image, dtype='float32') orig_image_with_boxes = draw_boxes(orig_image_data, resized_boxes, out_classes, class_names, out_scores,"rand") return resized_boxes,out_classes,orig_image_with_boxes
39.984127
133
0.639936
4,294
0.852322
0
0
0
0
0
0
581
0.115324
4ba04308181ebd07871e89cce3a567b034f969f9
2,881
py
Python
examples/time_frequency/plot_tfr_topography.py
Anevar/mne-python
15b19ed6b9364ae4787f0df2fd7e689b3c0a30bb
[ "BSD-3-Clause" ]
2
2015-09-27T20:33:49.000Z
2020-04-22T19:10:56.000Z
examples/time_frequency/plot_tfr_topography.py
Anevar/mne-python
15b19ed6b9364ae4787f0df2fd7e689b3c0a30bb
[ "BSD-3-Clause" ]
null
null
null
examples/time_frequency/plot_tfr_topography.py
Anevar/mne-python
15b19ed6b9364ae4787f0df2fd7e689b3c0a30bb
[ "BSD-3-Clause" ]
1
2018-09-15T09:45:38.000Z
2018-09-15T09:45:38.000Z
""" =================================================================== Plot time-frequency representations on topographies for MEG sensors =================================================================== Both induced power and phase locking values are displayed. """ print(__doc__) # Authors: Alexandre Gramfort <[email protected]> # Denis Engemann <[email protected]> # # License: BSD (3-clause) import numpy as np import matplotlib.pyplot as plt import mne from mne import fiff from mne.time_frequency import induced_power from mne.viz import plot_topo_power, plot_topo_phase_lock from mne.datasets import sample data_path = sample.data_path() raw_fname = data_path + '/MEG/sample/sample_audvis_raw.fif' event_fname = data_path + '/MEG/sample/sample_audvis_raw-eve.fif' event_id, tmin, tmax = 1, -0.2, 0.5 # Setup for reading the raw data raw = fiff.Raw(raw_fname) events = mne.read_events(event_fname) include = [] raw.info['bads'] += ['MEG 2443', 'EEG 053'] # bads + 2 more # picks MEG gradiometers picks = fiff.pick_types(raw.info, meg='grad', eeg=False, eog=True, stim=False, include=include, exclude='bads') epochs = mne.Epochs(raw, events, event_id, tmin, tmax, picks=picks, baseline=(None, 0), reject=dict(grad=4000e-13, eog=150e-6)) data = epochs.get_data() # as 3D matrix layout = mne.find_layout(epochs.info, 'meg') ############################################################################### # Calculate power and phase locking value frequencies = np.arange(7, 30, 3) # define frequencies of interest n_cycles = frequencies / float(7) # different number of cycle per frequency Fs = raw.info['sfreq'] # sampling in Hz decim = 3 power, phase_lock = induced_power(data, Fs=Fs, frequencies=frequencies, n_cycles=n_cycles, n_jobs=1, use_fft=False, decim=decim, zero_mean=True) ############################################################################### # Prepare topography plots, set baseline correction parameters baseline = (None, 0) # set the baseline for induced power mode = 'ratio' # set mode for baseline rescaling ############################################################################### # Show topography of power. title = 'Induced power - MNE sample data' plot_topo_power(epochs, power, frequencies, layout, baseline=baseline, mode=mode, decim=decim, vmin=0., vmax=14, title=title) plt.show() ############################################################################### # Show topography of phase locking value (PLV) mode = None # no baseline rescaling for PLV title = 'Phase locking value - MNE sample data' plot_topo_phase_lock(epochs, phase_lock, frequencies, layout, baseline=baseline, mode=mode, decim=decim, title=title) plt.show()
35.567901
79
0.591461
0
0
0
0
0
0
0
0
1,377
0.477959
4ba39b6087c75616c2877cb61a2b0736b03e97e4
2,201
py
Python
A1/greenHouseBluetooth.py
rmit-s3559384-andrew-alvaro/IoT
ec444d0b037ddbd2e3aab01c34ea57fd2bd51d5f
[ "MIT" ]
null
null
null
A1/greenHouseBluetooth.py
rmit-s3559384-andrew-alvaro/IoT
ec444d0b037ddbd2e3aab01c34ea57fd2bd51d5f
[ "MIT" ]
1
2021-06-01T23:39:58.000Z
2021-06-01T23:39:58.000Z
A1/greenHouseBluetooth.py
AndrewAlvaro/IoT
ec444d0b037ddbd2e3aab01c34ea57fd2bd51d5f
[ "MIT" ]
null
null
null
import bluetooth import sys, os import subprocess as sp import datetime from pushBulletForBluetooth import pushNotification from makeReminderforBluetooth import Reminder import csv class blueDev: def findmyDevice(self): sendPushBullet = pushNotification() timestamp = datetime.datetime.now().strftime('%d/%m/%Y') nearby_devices = bluetooth.discover_devices(lookup_names = True) if nearby_devices is not None: print("Scanned device:") for addr, name in nearby_devices: devices = (addr.split("(")[-1]) print(devices) else: print("No device available") print() paired = sp.Popen(["bt-device", "--list"], stdin = sp.PIPE, stdout = sp.PIPE, close_fds = True) (stdout, stdin) = (paired.stdout, paired.stdin) list_of_paired_devices = stdout.readlines() list_of_paired_devices.pop(0) print("Matching devices...") for paired_device in list_of_paired_devices: pairedString = paired_device.decode() pairedSplit = pairedString.split("(")[-1] pairedDevice = pairedSplit[0:-2] for devices, name in nearby_devices: if pairedDevice == devices: print(devices, "=", pairedDevice) with open('bluetoothReminder.csv', 'r') as csvfile: readCSV = csv.reader(csvfile) for row in readCSV: if row[0] != timestamp: print("Device matched!") sendPushBullet.send() else: print("Device matched! Notification has already been sent today.") else: print(devices, "!=", pairedDevice) print("Device not matched...") def main(): bluetooth = blueDev() reminder = Reminder() reminder.makeReminder() bluetooth.findmyDevice() if __name__ == "__main__": main()
31
103
0.52567
1,851
0.840981
0
0
0
0
0
0
236
0.107224
4ba3e0dab8146008256a0da74d6aec2d33aa11e9
127
py
Python
appending_to_files.py
jaiminjerry/Python
eb7013c7560b09d37849d653516257d939e143aa
[ "bzip2-1.0.6" ]
null
null
null
appending_to_files.py
jaiminjerry/Python
eb7013c7560b09d37849d653516257d939e143aa
[ "bzip2-1.0.6" ]
null
null
null
appending_to_files.py
jaiminjerry/Python
eb7013c7560b09d37849d653516257d939e143aa
[ "bzip2-1.0.6" ]
1
2021-08-17T03:46:56.000Z
2021-08-17T03:46:56.000Z
appendMe = '\nNew bit of information' appendFile = open('example.txt','a') appendFile.write(appendMe) appendFile.close()
21.166667
38
0.716535
0
0
0
0
0
0
0
0
42
0.330709
4ba4c531fc5b73ca047fb0191f3bbb5ca13cf62d
209
py
Python
udacity/cloud-native-application-architecture/3-message-passing/lesson-3-implementing-message-passing/kafka-python-demo/producer.py
thomasrobertz/mooc
cb87365bfcbe8ccf972f36d70a251c73b3c15a7b
[ "MIT" ]
null
null
null
udacity/cloud-native-application-architecture/3-message-passing/lesson-3-implementing-message-passing/kafka-python-demo/producer.py
thomasrobertz/mooc
cb87365bfcbe8ccf972f36d70a251c73b3c15a7b
[ "MIT" ]
13
2021-12-14T20:59:34.000Z
2022-03-02T11:09:34.000Z
udacity/cloud-native-application-architecture/3-message-passing/lesson-3-implementing-message-passing/kafka-python-demo/producer.py
thomasrobertz/mooc
cb87365bfcbe8ccf972f36d70a251c73b3c15a7b
[ "MIT" ]
1
2020-08-20T12:53:43.000Z
2020-08-20T12:53:43.000Z
from kafka import KafkaProducer TOPIC_NAME = 'items' KAFKA_SERVER = 'localhost:9092' producer = KafkaProducer(bootstrap_servers=KAFKA_SERVER) producer.send(TOPIC_NAME, b'Test Message!!!') producer.flush()
19
56
0.789474
0
0
0
0
0
0
0
0
41
0.196172
4ba509cf1a05cf33bf195b861b6306f41e7b81ea
954
py
Python
myfitnesspal_to_sqlite/cli.py
seeM/myfitnesspal-to-sqlite
ce4c133009cbeacd5fa5410016f81f5eb45e7a64
[ "Apache-2.0" ]
4
2021-07-14T17:31:40.000Z
2021-12-03T21:50:09.000Z
myfitnesspal_to_sqlite/cli.py
seeM/myfitnesspal-to-sqlite
ce4c133009cbeacd5fa5410016f81f5eb45e7a64
[ "Apache-2.0" ]
null
null
null
myfitnesspal_to_sqlite/cli.py
seeM/myfitnesspal-to-sqlite
ce4c133009cbeacd5fa5410016f81f5eb45e7a64
[ "Apache-2.0" ]
null
null
null
from datetime import datetime import myfitnesspal import sqlite_utils import click from . import utils @click.group() @click.version_option() def cli(): "Save data from MyFitnessPal to a SQLite database" @cli.command() @click.argument( "db_path", type=click.Path(file_okay=True, dir_okay=False, allow_dash=False), required=True, ) @click.argument( "user", type=str, required=True, ) @click.argument( "date", type=str, required=True, ) @click.option( "--measurement", multiple=True, required=True, ) def diary(db_path, user, date, measurement): "Save food, exercise, goal, and measurement entries for a given user and date" date = datetime.fromisoformat(date).date() db = sqlite_utils.Database(db_path) client = myfitnesspal.Client(user) diary_entry = utils.fetch_diary_entry(date, client, measurement) utils.save_diary_entry(db, diary_entry) utils.ensure_db_shape(db)
21.2
82
0.705451
0
0
0
0
845
0.885744
0
0
164
0.171908
4ba51c782d7e269031d5abf6080e2a03357844fd
849
py
Python
mcir/t1_hist.py
omritomer/mcir
1554d352172464c6314339195d6ea9a5e00824af
[ "MIT" ]
null
null
null
mcir/t1_hist.py
omritomer/mcir
1554d352172464c6314339195d6ea9a5e00824af
[ "MIT" ]
null
null
null
mcir/t1_hist.py
omritomer/mcir
1554d352172464c6314339195d6ea9a5e00824af
[ "MIT" ]
null
null
null
import numpy as np class T1Hist: def __init__( self, max_t1: float = 3000, min_t1: float = 200 ): self.max_t1 = max_t1 self.min_t1 = min_t1 def get_t1_histogram(self, t1_matrix: np.ndarray, norm_m0_matrix: np.ndarray): t1_histogram = t1_matrix.astype(float).ravel() t1_weights = norm_m0_matrix.astype(float).ravel() self.t1_histogram, self.t1_weights = self.remove_outliers( t1_histogram, t1_weights ) def remove_outliers(self, t1_histogram, t1_weights): t1_weights = t1_weights[ (t1_histogram > (self.min_t1)) & (t1_histogram < (self.max_t1)) ] t1_histogram = t1_histogram[ (t1_histogram > (self.min_t1)) & (t1_histogram < (self.max_t1)) ] return t1_histogram, t1_weights
30.321429
82
0.61013
827
0.974087
0
0
0
0
0
0
0
0
4ba5d882b2fc5de31e1705b7b18a845f264237e7
209
py
Python
main.py
Javert899/pm4py-tool-plugin-skeleton
cfc4aefd02499b323ae60e33f059a6b90e48a95f
[ "MIT" ]
null
null
null
main.py
Javert899/pm4py-tool-plugin-skeleton
cfc4aefd02499b323ae60e33f059a6b90e48a95f
[ "MIT" ]
null
null
null
main.py
Javert899/pm4py-tool-plugin-skeleton
cfc4aefd02499b323ae60e33f059a6b90e48a95f
[ "MIT" ]
null
null
null
import pluginpackageRENAME import preload import os if __name__ == "__main__": preload.preload() app = pluginpackageRENAME.app app.static_folder = os.path.join(os.getcwd(), "html") app.run()
19
57
0.703349
0
0
0
0
0
0
0
0
16
0.076555
4ba61c47eb12a3d8f57c257b4b752059384399df
6,948
py
Python
plot_curve.py
wenhuchen/Hierarchical-DSA
2dbdacde25ee82c9d42fe980694673d285b1f7f3
[ "MIT" ]
45
2019-02-27T02:04:08.000Z
2022-03-21T04:49:22.000Z
plot_curve.py
wenhuchen/Hierarchical-DSA
2dbdacde25ee82c9d42fe980694673d285b1f7f3
[ "MIT" ]
2
2019-08-18T03:05:11.000Z
2020-07-26T13:45:05.000Z
plot_curve.py
wenhuchen/Hierarchical-DSA
2dbdacde25ee82c9d42fe980694673d285b1f7f3
[ "MIT" ]
4
2019-03-12T17:40:12.000Z
2021-06-10T07:59:39.000Z
import matplotlib.pyplot as plt import numpy as np import math import numpy as np import matplotlib.pyplot as plt from scipy.optimize import curve_fit import json from scipy.interpolate import interp1d from data_utils import * def integral(y, x): area = 0 for xi, xj, yi, yj in zip(x[:-1], x[1:], y[:-1], y[1:]): area += (yi + yj) / 2 * abs(xj - xi) return area def preprocess(x): #x = [math.log10(_) for _ in x] x = [_/float(max(x)) for _ in x] return x def func(y, x, y_int): for y1, y2, x1, x2 in zip(y[:-1], y[1:], x[:-1], x[1:]): if y_int == y1: return x1 elif y_int == y2: return x2 elif y_int > y1 and y_int < y2: x_int = (y_int - y1) / (y2 - y1) * (x2 - x1) + x1 return x_int def draw_curve(): x = [1, 100, 200, 500, 1000, 2000, 5000, 10000, 20000, 40000] y_mean = [13.3, 29.6, 33.9, 43.8, 50.81, 67.7, 75.6, 81.5, 91.4, 95.6] plt.plot(x, y_mean, 'black') #plt.fill(x + x[::-1], y_mean + [95.6] * len(y_min), '#f4df42', alpha=.5, ec='None') plt.fill(x + x[::-1], [0] * len(y_mean) + y_mean[::-1], '#0099ff', alpha=.5, ec='None') plt.xlabel("Vocab") plt.xscale('log') plt.xlim(left=0) plt.ylim(ymin=0) plt.ylabel("Accuracy") #plt.show() plt.savefig("metrics.eps", dpi=1000, format='eps') #draw_curve() def draw_uncerntainy_curve(): x = [0, 100, 200, 500, 1000, 2000, 5000, 10000] y_max = [13.3, 51.2, 67.5, 80.4, 85.1, 87.5, 90.5, 91.4] y_mean = [13.3, 29.6, 33.9, 43.9, 50.81, 67.7, 81.5, 91.4] y_min = [13.3, 25.6, 27, 35.1, 42.4, 56, 74.1, 91.4] plt.plot(x, y_mean, 'black', label="Mean Accuracy Curve") plt.plot(x, y_min, 'black', label="Lower Accuracy Curve") plt.plot(x, y_max, 'black', label="Upper Accuracy Curve") #plt.plot(X, y, 'r.', markersize=10, label=u'Observations') #plt.plot(x, y_pred, 'b-', label=u'Prediction') plt.fill(x + x[::-1], y_min + y_max[::-1], '#0099ff', alpha=.5, ec='None', label='Accuracy Range') plt.legend(loc='lower right', prop={'size':14}) plt.xlim(left=0) plt.xlabel("Vocab") plt.ylabel("Accuracy") plt.savefig("accuracy_curve.eps", dpi=1000, format='eps') #plt.show() #plt.fill(np.concatenate([x, x[::-1]]), # np.concatenate([y_pred - 1.9600 * sigma, # (y_pred + 1.9600 * sigma)[::-1]]), # alpha=.5, fc='b', ec='None', label='95% confidence interval') #draw_uncerntainy_curve() def draw_SLU_uncerntainy_curve(): x = [0, 7, 27, 77, 100, 1778, 5134, 10000] x = [str(_) for _ in x] y_max = [13.3, 48.8, 81.3, 92.0, 94.0, 95.3, 95.8, 96.1] y_mean = [13.3, 33.4, 54.3, 77.4, 88.9, 93.5, 94.2, 96.1] y_min = [13.3, 14.2, 33.2, 46.8, 72.8, 88.4, 92.3, 96.1] plt.plot(x, y_mean, color='black', label="Mean Accuracy Curve") plt.plot(x, y_min, color='black', label="Lower Accuracy Curve") plt.plot(x, y_max, color='black', label="Upper Accuracy Curve") #plt.plot(X, y, 'r.', markersize=10, label=u'Observations') #plt.plot(x, y_pred, 'b-', label=u'Prediction') plt.fill(x + x[::-1], y_min + y_max[::-1], color='#0099ff', alpha=.5, ec='None', label='Accuracy Range') plt.xlim(left=0) plt.ylim(bottom=0) plt.legend(loc='lower right', prop={'size':14}) plt.xlabel("Vocab") plt.ylabel("Accuracy") plt.savefig("accuracy_curve.eps", dpi=1000, format='eps') #plt.show() #plt.fill(np.concatenate([x, x[::-1]]), # np.concatenate([y_pred - 1.9600 * sigma, # (y_pred + 1.9600 * sigma)[::-1]]), # alpha=.5, fc='b', ec='None', label='95% confidence interval') #draw_SLU_uncerntainy_curve() def read(string, use_str=False): string = string.strip() result = eval(string) if use_str: result = [str(_) for _ in result] else: result = [float(_) for _ in result] return result def draw_curve(): def savefig(f1, f2, f3, name): x, y = enhanced(read(f1[1]), read(f1[0])) plt.plot(x, y, 'y*', label="Frequency") x, y = enhanced(read(f2[1]), read(f2[0])) plt.plot(x, y, 'b--', label="TF-IDF") x, y = enhanced(read(f3[1]), read(f3[0])) plt.plot(x, y, 'r', label="Variational") #plt.title("{} dataset".format(name)) plt.xlabel('Vocab') plt.ylabel('Accuracy') plt.legend(loc='best') plt.xscale('log') #plt.xlim(left=0.001) #plt.show() plt.savefig("{}.eps".format(name), format="eps", dpi=1000) plt.clf() plt.rcParams.update({'font.size': 14}) file = 'results/ag_news' f1 = open(file + ".txt").readlines() f2 = open(file + "_tf_idf.txt").readlines() f3 = open(file + "_var.txt").readlines() savefig(f1, f2, f3, 'images/ag_news') file = 'results/dbpedia' f1 = open(file + ".txt").readlines() f2 = open(file + "_tf_idf.txt").readlines() f3 = open(file + "_var.txt").readlines() savefig(f1, f2, f3, 'images/dbpedia') file = 'results/yelp_review' f1 = open(file + ".txt").readlines() f2 = open(file + "_tf_idf.txt").readlines() f3 = open(file + "_var.txt").readlines() savefig(f1, f2, f3, 'images/yelp_review') #draw_curve() def compute_score(): from data_utils import * f = 'results/ag_news.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 61673) print CR(y, x) f = 'results/ag_news_tf_idf.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 61673) print CR(y, x) f = 'results/ag_news_var.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 61673) print CR(y, x) print() f = 'results/dbpedia.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 563355) print CR(y, x) f = 'results/dbpedia_tf_idf.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 563355) print CR(y, x) f = 'results/dbpedia_var.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 563355) print CR(y, x) print() f = 'results/yelp_review.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 252712) print CR(y, x) f = 'results/yelp_review_tf_idf.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 252712) print CR(y, x) f = 'results/yelp_review_var.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 252712) print CR(y, x) print() f = 'results/sogou_news.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 254495) print CR(y, x) f = 'results/sogou_news_tf_idf.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 254495) print CR(y, x) f = 'results/snli.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 42391) print CR(y, x) f = 'results/snli_var.txt' y = read(open(f).readlines()[0]) x = read(open(f).readlines()[1]) print ROC(y, x, 42391) print CR(y, x) compute_score()
29.692308
105
0.598877
0
0
0
0
0
0
0
0
1,954
0.281232
4ba90d216dd9521bb1b314598a55d371117b4821
8,392
py
Python
alipay/aop/api/domain/CircleRecommendItemDTO.py
antopen/alipay-sdk-python-all
8e51c54409b9452f8d46c7bb10eea7c8f7e8d30c
[ "Apache-2.0" ]
213
2018-08-27T16:49:32.000Z
2021-12-29T04:34:12.000Z
alipay/aop/api/domain/CircleRecommendItemDTO.py
antopen/alipay-sdk-python-all
8e51c54409b9452f8d46c7bb10eea7c8f7e8d30c
[ "Apache-2.0" ]
29
2018-09-29T06:43:00.000Z
2021-09-02T03:27:32.000Z
alipay/aop/api/domain/CircleRecommendItemDTO.py
antopen/alipay-sdk-python-all
8e51c54409b9452f8d46c7bb10eea7c8f7e8d30c
[ "Apache-2.0" ]
59
2018-08-27T16:59:26.000Z
2022-03-25T10:08:15.000Z
#!/usr/bin/env python # -*- coding: utf-8 -*- import json from alipay.aop.api.constant.ParamConstants import * from alipay.aop.api.domain.AoiInfoDTO import AoiInfoDTO from alipay.aop.api.domain.ItemStoreDTO import ItemStoreDTO class CircleRecommendItemDTO(object): def __init__(self): self._aoi_info = None self._discount = None self._item_cover = None self._item_detail_url = None self._item_id = None self._item_label = None self._item_name = None self._item_store = None self._original_price = None self._sales_info = None self._saved_money = None self._saved_money_info = None self._sell_price = None self._sold_quantity = None self._store_id = None @property def aoi_info(self): return self._aoi_info @aoi_info.setter def aoi_info(self, value): if isinstance(value, AoiInfoDTO): self._aoi_info = value else: self._aoi_info = AoiInfoDTO.from_alipay_dict(value) @property def discount(self): return self._discount @discount.setter def discount(self, value): self._discount = value @property def item_cover(self): return self._item_cover @item_cover.setter def item_cover(self, value): self._item_cover = value @property def item_detail_url(self): return self._item_detail_url @item_detail_url.setter def item_detail_url(self, value): self._item_detail_url = value @property def item_id(self): return self._item_id @item_id.setter def item_id(self, value): self._item_id = value @property def item_label(self): return self._item_label @item_label.setter def item_label(self, value): self._item_label = value @property def item_name(self): return self._item_name @item_name.setter def item_name(self, value): self._item_name = value @property def item_store(self): return self._item_store @item_store.setter def item_store(self, value): if isinstance(value, ItemStoreDTO): self._item_store = value else: self._item_store = ItemStoreDTO.from_alipay_dict(value) @property def original_price(self): return self._original_price @original_price.setter def original_price(self, value): self._original_price = value @property def sales_info(self): return self._sales_info @sales_info.setter def sales_info(self, value): self._sales_info = value @property def saved_money(self): return self._saved_money @saved_money.setter def saved_money(self, value): self._saved_money = value @property def saved_money_info(self): return self._saved_money_info @saved_money_info.setter def saved_money_info(self, value): self._saved_money_info = value @property def sell_price(self): return self._sell_price @sell_price.setter def sell_price(self, value): self._sell_price = value @property def sold_quantity(self): return self._sold_quantity @sold_quantity.setter def sold_quantity(self, value): self._sold_quantity = value @property def store_id(self): return self._store_id @store_id.setter def store_id(self, value): self._store_id = value def to_alipay_dict(self): params = dict() if self.aoi_info: if hasattr(self.aoi_info, 'to_alipay_dict'): params['aoi_info'] = self.aoi_info.to_alipay_dict() else: params['aoi_info'] = self.aoi_info if self.discount: if hasattr(self.discount, 'to_alipay_dict'): params['discount'] = self.discount.to_alipay_dict() else: params['discount'] = self.discount if self.item_cover: if hasattr(self.item_cover, 'to_alipay_dict'): params['item_cover'] = self.item_cover.to_alipay_dict() else: params['item_cover'] = self.item_cover if self.item_detail_url: if hasattr(self.item_detail_url, 'to_alipay_dict'): params['item_detail_url'] = self.item_detail_url.to_alipay_dict() else: params['item_detail_url'] = self.item_detail_url if self.item_id: if hasattr(self.item_id, 'to_alipay_dict'): params['item_id'] = self.item_id.to_alipay_dict() else: params['item_id'] = self.item_id if self.item_label: if hasattr(self.item_label, 'to_alipay_dict'): params['item_label'] = self.item_label.to_alipay_dict() else: params['item_label'] = self.item_label if self.item_name: if hasattr(self.item_name, 'to_alipay_dict'): params['item_name'] = self.item_name.to_alipay_dict() else: params['item_name'] = self.item_name if self.item_store: if hasattr(self.item_store, 'to_alipay_dict'): params['item_store'] = self.item_store.to_alipay_dict() else: params['item_store'] = self.item_store if self.original_price: if hasattr(self.original_price, 'to_alipay_dict'): params['original_price'] = self.original_price.to_alipay_dict() else: params['original_price'] = self.original_price if self.sales_info: if hasattr(self.sales_info, 'to_alipay_dict'): params['sales_info'] = self.sales_info.to_alipay_dict() else: params['sales_info'] = self.sales_info if self.saved_money: if hasattr(self.saved_money, 'to_alipay_dict'): params['saved_money'] = self.saved_money.to_alipay_dict() else: params['saved_money'] = self.saved_money if self.saved_money_info: if hasattr(self.saved_money_info, 'to_alipay_dict'): params['saved_money_info'] = self.saved_money_info.to_alipay_dict() else: params['saved_money_info'] = self.saved_money_info if self.sell_price: if hasattr(self.sell_price, 'to_alipay_dict'): params['sell_price'] = self.sell_price.to_alipay_dict() else: params['sell_price'] = self.sell_price if self.sold_quantity: if hasattr(self.sold_quantity, 'to_alipay_dict'): params['sold_quantity'] = self.sold_quantity.to_alipay_dict() else: params['sold_quantity'] = self.sold_quantity if self.store_id: if hasattr(self.store_id, 'to_alipay_dict'): params['store_id'] = self.store_id.to_alipay_dict() else: params['store_id'] = self.store_id return params @staticmethod def from_alipay_dict(d): if not d: return None o = CircleRecommendItemDTO() if 'aoi_info' in d: o.aoi_info = d['aoi_info'] if 'discount' in d: o.discount = d['discount'] if 'item_cover' in d: o.item_cover = d['item_cover'] if 'item_detail_url' in d: o.item_detail_url = d['item_detail_url'] if 'item_id' in d: o.item_id = d['item_id'] if 'item_label' in d: o.item_label = d['item_label'] if 'item_name' in d: o.item_name = d['item_name'] if 'item_store' in d: o.item_store = d['item_store'] if 'original_price' in d: o.original_price = d['original_price'] if 'sales_info' in d: o.sales_info = d['sales_info'] if 'saved_money' in d: o.saved_money = d['saved_money'] if 'saved_money_info' in d: o.saved_money_info = d['saved_money_info'] if 'sell_price' in d: o.sell_price = d['sell_price'] if 'sold_quantity' in d: o.sold_quantity = d['sold_quantity'] if 'store_id' in d: o.store_id = d['store_id'] return o
32.401544
83
0.596282
8,159
0.972235
0
0
3,824
0.455672
0
0
1,040
0.123928
4bab64e23dd52a8b7e2e5474ebad268f962e7d94
3,599
py
Python
vmcasterpub/uploader_dcap.py
hepix-virtualisation/vmcaster
f4ef1c65bbb81b82aa72a0cd1afc1aa6cf13eb51
[ "Apache-2.0" ]
null
null
null
vmcasterpub/uploader_dcap.py
hepix-virtualisation/vmcaster
f4ef1c65bbb81b82aa72a0cd1afc1aa6cf13eb51
[ "Apache-2.0" ]
null
null
null
vmcasterpub/uploader_dcap.py
hepix-virtualisation/vmcaster
f4ef1c65bbb81b82aa72a0cd1afc1aa6cf13eb51
[ "Apache-2.0" ]
null
null
null
import subprocess import time import logging import os import signal log = logging.getLogger(__name__) def runpreloadcommand(cmd,timeout,preload): newenv = dict(os.environ) newenv["LD_PRELOAD"] = preload process = subprocess.Popen([cmd], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE,env=newenv) processRc = None handleprocess = True counter = 0 stdout = '' stderr = '' while handleprocess: counter += 1 time.sleep(1) cout,cerr = process.communicate() stdout += cout stderr += cerr process.poll() processRc = process.returncode if processRc != None: break if counter == timeout: os.kill(process.pid, signal.SIGQUIT) if counter > timeout: os.kill(process.pid, signal.SIGKILL) processRc = -9 break return (processRc,stdout,stderr) def gsiDcapCopy(src,dest,timeout = 60): cmd = "dccp -C 3000 -d 2 -A %s %s" % (src,dest) process = subprocess.Popen([cmd], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) processRc = None handleprocess = True counter = 0 stdout = '' stderr = '' while handleprocess: counter += 1 time.sleep(1) cout,cerr = process.communicate() stdout += cout stderr += cerr process.poll() processRc = process.returncode if processRc != None: break if counter == timeout: os.kill(process.pid, signal.SIGQUIT) if counter > timeout: os.kill(process.pid, signal.SIGKILL) processRc = -9 break if processRc != 0: log = logging.getLogger("gsiDcapCopy") log.error("failed to execute command '%s'" % (cmd)) return (processRc,stdout,stderr) class uploaderDcap: def __init__(self): self.remotePrefix = None self.log = logging.getLogger("uploaderGsiDcap") def _getfilepath(self,remotePath): if self.remotePrefix != None: return self.remotePrefix + remotePath else: return remotePath def exists(self,remotePath): cmd = "stat %s" % (self._getfilepath(remotePath)) timeout = 10 preload = "/usr/lib64/libpdcap.so.1" return runpreloadcommand(cmd,timeout,preload) def delete(self,remotePath): cmd = "unlink %s" % (self._getfilepath(remotePath)) timeout = 10 preload = "/usr/lib64/libpdcap.so.1" return runpreloadcommand(cmd,timeout,preload) def upload(self,localpath,remotePath): path = self._getfilepath(remotePath) return gsiDcapCopy(localpath,path) def replace(self,localpath,remotePath): path = self._getfilepath(remotePath) (rc,stdout,stderr) = self.exists(path) if rc == 0: (rc,stdout,stderr) = self.delete(path) if rc != 0: msg = "stderr={stderr}".format(stderr=stderr) log.error(msg) return (rc,stdout,stderr) rc,stdout,stderr = gsiDcapCopy(localpath,path) if rc != 0: msg = "stderr={stderr}".format(stderr=stderr) log.error(msg) return (rc,stdout,stderr) return (rc,stdout,stderr) def download(self,remotePath,localpath): rc,stdout,stderr = gsiDcapCopy(self._getfilepath(remotePath),localpath) if rc != 0: for errorLine in stderr.split('\n'): self.log.error("stderr:'%s'" % (errorLine)) return rc,stdout,stderr
31.025862
108
0.594054
1,746
0.485135
0
0
0
0
0
0
233
0.06474
4babaa82bca32126bf21a61b9966b1e6ecb0d62c
3,923
py
Python
drones/serializers.py
maprezdev/restfuldrones
9448a63b148cdf7da8f46d65067ddbb8773e2fd2
[ "MIT" ]
null
null
null
drones/serializers.py
maprezdev/restfuldrones
9448a63b148cdf7da8f46d65067ddbb8773e2fd2
[ "MIT" ]
null
null
null
drones/serializers.py
maprezdev/restfuldrones
9448a63b148cdf7da8f46d65067ddbb8773e2fd2
[ "MIT" ]
null
null
null
# drones/serializers.py file from rest_framework import serializers from drones.models import DroneCategory, Drone, Pilot, Competition from django.contrib.auth.models import User import drones.views class UserDroneSerializer(serializers.HyperlinkedModelSerializer): """serialize the drones related to a User""" class Meta: model: Drone fields = ( 'url', 'name') class UserSerializer(serializers.HyperlinkedModelSerializer): """declare an instance of the UserDroneSerializer class""" drones = UserDroneSerializer( many=True, read_only=True) class Meta: model = User fields = ( 'url', 'pk', 'username', 'drone') class DroneCategorySerializer(serializers.HyperlinkedModelSerializer): """defines a one-to-many relationship that is read- only""" drones = serializers.HyperlinkedRelatedField( many=True, read_only=True, view_name='drone-detail', # browsable API feature ) class Meta: """model related to the serializer, and field names that we want to include in the serialization""" model = DroneCategory fields = ( 'url', 'pk', 'name', 'drones') class DroneSerializer(serializers.HyperlinkedModelSerializer): """display the drone category name""" drone_category = serializers.SlugRelatedField( queryset=DroneCategory.objects.all(), slug_field='name') # Display the owner's username (read-only) owner = serializers.ReadOnlyField(source='owner.username') class Meta: """model related to the serializer, and field names that we want to include in the serialization""" model = Drone fields = ( 'url', 'name', 'drone_category', 'owner', 'manufacturing_date', 'has_it_competed', 'inserted_timestamp', ) class CompetitionSerializer(serializers.HyperlinkedModelSerializer): """display all the details for the related Drone""" drone = DroneSerializer() class Meta: """model related to the serializer, and field names that we want to include in the serialization""" model = Competition fields = ( 'url', 'pk', 'distance_in_feet', 'distance_achievement_date', 'drone') class PilotSerializer(serializers.HyperlinkedModelSerializer): """serialize Pilot instances and serialize all the Competition instances related to the Pilot""" competitions = CompetitionSerializer(many=True, read_only=True) gender = serializers.ChoiceField(choices=Pilot.gender_choices) gender_description = serializers.CharField(source='get_gender_display', read_only=True) class Meta: """model related to the serializer, and field names that we want to include in the serialization""" model = Pilot fields = ( 'url', 'name', 'gender', 'gender_description', 'races_count', 'inserted_timestamp', 'competitions') class PilotCompetitionSerializer(serializers.ModelSerializer): """display the related Pilot name and the related Drone name""" pilot = serializers.SlugRelatedField( queryset=Pilot.objects.all(), slug_field='name') drone = serializers.SlugRelatedField( queryset=Drone.objects.all(), slug_field='name') class Meta: """model related to the serializer, and field names that we want to include in the serialization""" model = Competition fields = ( 'url', 'pk', 'distance_in_feet', 'distance_achievement_date', 'pilot', 'drone')
29.946565
100
0.617894
3,704
0.944175
0
0
0
0
0
0
1,470
0.374713
4baef5968ecd4571dc42ca2e3a144059ebfa9562
1,471
py
Python
Calamous.py
Studio-Pasteque/Pokemon
6b9f457eef8a2dc28cb8b9b69527404b47c9825a
[ "MIT" ]
2
2020-05-27T08:27:58.000Z
2020-05-27T09:31:45.000Z
Calamous.py
Studio-Pasteque/Pokemon
6b9f457eef8a2dc28cb8b9b69527404b47c9825a
[ "MIT" ]
null
null
null
Calamous.py
Studio-Pasteque/Pokemon
6b9f457eef8a2dc28cb8b9b69527404b47c9825a
[ "MIT" ]
null
null
null
import learnables.py # création d'un objet Calamous class Calamous: #définition de son type et de si c'est un pokémon pokemon = True type_ = "Water" # définition de ses talents et de son genre (qui seront random) abilities = "Multicule" or "Glissade" or "Protéen" gender = "M" or "F" # création de ses niveaux levels = 1 exp = 0 needexp = 30 lootexp = 10 difupgexp = 3 # création de la possibilité de monter de niveau if adverse == 0: exp = exp + lootexp * difupgexp if levels == 1 and exp == needexp: levels = 2 needexp = 70 exp = 0 if adverse == 0: exp = exp + lootexp * difupgexp if exp == needexp: levels = levels + 1 needexp = needexp * 2 exp = 0 # définition de ses stats hp = 50 atq = 35 df = 40 spa = 60 spd = 30 spe = 60 pre = 1 esc = 1 # définition de l'apparence et des capacités qu'il pourra apprendre appearence = ('assets/personnages/Poké-avant/Calamous') # définition de valeurs plus cachées happiness = 0 # création de la possibilité de monter certaines d'entre elles if adverse == 0: happiness = happiness + 0.1 # description du pokédex desc = "Ce pokémon ne peut pas vivre hors de l'eau : sa peau sèche dès qu'elle est déshydratée trop longtemps"
27.754717
114
0.56968
1,434
0.960482
0
0
0
0
0
0
647
0.433356
4bb0265f943903e9ce05ffd83240a67916be1de6
5,186
py
Python
scripts/utils.py
alterapars/drought_classification
585aaed3f00d5835059be1c80ad998189d9726f7
[ "MIT" ]
1
2022-02-19T11:42:24.000Z
2022-02-19T11:42:24.000Z
scripts/utils.py
alterapars/drought_classification
585aaed3f00d5835059be1c80ad998189d9726f7
[ "MIT" ]
null
null
null
scripts/utils.py
alterapars/drought_classification
585aaed3f00d5835059be1c80ad998189d9726f7
[ "MIT" ]
2
2022-02-02T08:24:37.000Z
2022-02-03T12:27:05.000Z
import random import matplotlib.pyplot as plt import numpy as np from scipy import stats ############################ STATS input data ################################################ def return_nan_percentage(input_data): """ prints percentage of nan values in max. 3D sized array Parameters ---------- input_array : array max 3D array Returns ------- None """ total_size = input_data.size nan_sum = np.isnan(input_data).sum() perc = float(nan_sum / total_size) print("percentage of nan values inside dataset is: %.2f" % float(perc) + " %") # #4D example: # for i in Training_data: # return_nan_percentage(i) def describe_with_stats(input_data): flat_array = input_data.flatten() # 'omit'performs the calculations ignoring nan values nobs, minmax, mean, variance, skewness, kurtosis = stats.describe( flat_array, nan_policy="omit" ) print("Number of observations: " + str(nobs)) print("min: " + str(minmax[0])) print("max: " + str(minmax[1])) print("the mean is: " + str(mean)) print("the variance is: " + str(variance)) print("Skewness is: " + str(skewness)) print("Kurtosis: " + str(kurtosis)) print("---") # for i in Training_data_germany: # describe_with_stats(i) ############################ Derive Labels ############################################### def mask_nan_values(input_array): array_with_masked_nans = input_array.fillna(value=10000.00) return array_with_masked_nans # back to xarray with: # label_xarray = xr.DataArray(output_3D_array, dims=['time', 'latitude', 'longitude'] ) # to turn list output into a 3D array use: def list_to_array(output_list): output_3D_array = np.stack(output_list, axis=0) return output_3D_array # TODO: returns list of 2D arrays now, try to return 3D x array to save as net cdf -SEE BELOW # TODO: write test # #Example: # #create data subset of 10 of a data xarray # data_10 = data[0:10] #first 10 items to test # print(data.shape) # #call function with a threshod of 10 # output_array = binary_image_classification(data_10, T=0.5) # #show one image of the masked output images # plt.imshow(output_array[0], origin = 'lower') # #might need to change 'lower' to 'upper" # TODO: def list_of_2D_xarray_to_netcdf(): x = 0 netcdf = x return netcdf def save_plots_from_3Darray( input_array, OUTPUT_PATH, title="drought mask figure Nr:", show_plots=True ): """ saves pngs and/or prints images from 3Darrays as png files Parameters ---------- input_xarray : array 3-D input array in the format [num_samples, height, width] title: str title of the plots, number will be added according to iteration index show_plots: boolean determines if plots will be displayed as output or not Returns ------- None """ for k in range(len(input_array[0])): fig = input_array[k].plot() plt.title(title + str(k)) plt.axis("equal") plt.title("drought mask for SMI, month " + str(k)) if show_plots: plt.show() fig.figure.savefig(OUTPUT_PATH + title + str(k) + ".png", dpi=100) print(OUTPUT_PATH + "drought_mask_" + str(k) + ".png") ############################ class imbalance ###################################### # option 1, faster, combine these 2 fcts (recommended): def hide_random_values(input_value, T=0.68): if input_value == 0: if np.random.rand(1) > T: return -1 return input_value # print(hide_random_values(0)) def reduce_class_size(input_array): output_array = np.copy(input_array) for t in range(0, 472): for xy in range(0, 7171): output_array[t, xy] = hide_random_values(output_array[t, xy]) return output_array # option 2, combine these 2 fcts: def get_indices(dataset, value=0): """dataset = str(), 2D-array value = int(), value to print the indices for""" result = np.where(dataset == value) print("Tuple of arrays returned : ", result) # zip the 2 arrays (array 1: rows, array 2: columns) to get the exact coordinates listOfCoordinates = list(zip(result[0], result[1])) # iterate over the list of coordinates # for cord in listOfCoordinates: # print(cord) print(len(listOfCoordinates)) return listOfCoordinates def reduce_class_size(input_array, indices_list, T=0.78, value=int(-1)): """set entries in array to value=x, randomly and within set percentage of array list = list, list of indices (2D) T = int() , percentage to be modified returns: """ output_array = np.copy(input_array) # determine the percentage of the array that will be modified len_modifier = int(len(indices_list) * T) # select percentage T randomly from the list random_coords = random.sample(listOfCoordinates, len_modifier) # print(random_coords[:10]) # set selected entries to value print("selected indices will be set to " + str(value)) for i in random_coords: # print(labels_reshaped[i]) output_array[i] == value return output_array
26.459184
94
0.633822
0
0
0
0
0
0
0
0
2,724
0.52526
4bb0f0499ca35cb26e70156806115a77ce9290c6
1,382
py
Python
2021/day8/2.py
tomhel/AoC_2019
c76c34235821864bc763f85d43cbcbfb9ed43469
[ "MIT" ]
1
2021-12-07T13:18:52.000Z
2021-12-07T13:18:52.000Z
2021/day8/2.py
tomhel/AoC
c76c34235821864bc763f85d43cbcbfb9ed43469
[ "MIT" ]
null
null
null
2021/day8/2.py
tomhel/AoC
c76c34235821864bc763f85d43cbcbfb9ed43469
[ "MIT" ]
null
null
null
def load(): with open("input") as f: for x in f: a, b, = x.strip().split("|") yield {frozenset(x) for x in a.strip().split()}, [frozenset(x) for x in b.strip().split()] def decode_signal(signal): num = {} while len(num) < 10: for x in signal.difference(num.values()): if len(x) == 2: num[1] = x elif len(x) == 3: num[7] = x elif len(x) == 4: num[4] = x elif len(x) == 7: num[8] = x elif len(x) == 6 and 4 in num and num[4].issubset(x): num[9] = x elif len(x) == 5 and 1 in num and num[1].issubset(x): num[3] = x elif len(x) == 6 and 7 in num and 9 in num and num[7].issubset(x) and num[9] != x: num[0] = x elif len(x) == 6 and 1 in num and not num[1].issubset(x): num[6] = x elif len(x) == 5 and 6 in num and x.issubset(num[6]): num[5] = x elif len(x) == 5 and 3 in num and 5 in num: num[2] = x return {v: k for k, v in num.items()} def decode_output(): result = 0 for sig, out in load(): mapping = decode_signal(sig) result += int("".join(str(mapping[x]) for x in out)) return result print(decode_output())
28.791667
102
0.447902
0
0
204
0.147612
0
0
0
0
12
0.008683
4bb35fc82ab5a2d2bc09de6a0496b0c17ea21b52
2,469
py
Python
Grove_Base_Hat_for_RPI/grove.py-master/grove/button/button.py
tcmoore/RPI-Environmental-Controller
7f28dcdf08c51db8400ccc0369eb049fdce5e901
[ "Unlicense", "MIT" ]
5
2019-11-18T02:26:18.000Z
2021-02-06T20:31:37.000Z
Grove_Base_Hat_for_RPI/grove.py-master/grove/button/button.py
tcmoore/RPI-Environmental-Controller
7f28dcdf08c51db8400ccc0369eb049fdce5e901
[ "Unlicense", "MIT" ]
null
null
null
Grove_Base_Hat_for_RPI/grove.py-master/grove/button/button.py
tcmoore/RPI-Environmental-Controller
7f28dcdf08c51db8400ccc0369eb049fdce5e901
[ "Unlicense", "MIT" ]
1
2020-08-26T10:22:37.000Z
2020-08-26T10:22:37.000Z
#!/usr/bin/env python # # This is the library for Grove Base Hat. # # Button Base Class # ''' ## License The MIT License (MIT) Grove Base Hat for the Raspberry Pi, used to connect grove sensors. Copyright (C) 2018 Seeed Technology Co.,Ltd. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ''' class Button(object): # event bits EV_RAW_STATUS = 1 << 0 EV_SINGLE_CLICK = 1 << 1 EV_DOUBLE_CLICK = 1 << 2 EV_LONG_PRESS = 1 << 3 EV_LEVEL_CHANGED = 1 << 4 # EV_HAS = 1 << 31 pins = [] def __init__(self, pin): self.__on_obj = None self.__on_event = None self.__event = 0 self.pins.append(pin) # To use with button array self.__index = self.pins.index(pin) def get_on_event(self): return self.__on_obj, self.__on_event def on_event(self, obj, callback): if not obj: return if not callable(callback): return self.__on_obj, self.__on_event = obj, callback def is_pressed(self): return False # call by derivate class def _send_event(self, event, pressed, tm): if not callable(self.__on_event): return evt = { 'index': self.__index, 'code' : event, 'pressed': pressed, 'time' : tm, } self.__on_event(self.__on_obj, evt)
31.253165
78
0.64439
1,159
0.469421
0
0
0
0
0
0
1,418
0.574322
4bb3c9f13001ae9f4765556a61ae26a55cabde2c
1,402
py
Python
Data Structures/Linked Lists/reverse-a-linked-list.py
Owngithub10101/Hackerrank-Problem-Solving
4e35b609c9f5b94c5bda292b9991baa054a944b6
[ "MIT" ]
23
2020-02-28T16:18:48.000Z
2021-12-21T11:51:07.000Z
Data Structures/Linked Lists/reverse-a-linked-list.py
ramanagali/Hackerrank-Problem-Solving
98f654f984013140d52b9a344146e9e38e46fb81
[ "MIT" ]
null
null
null
Data Structures/Linked Lists/reverse-a-linked-list.py
ramanagali/Hackerrank-Problem-Solving
98f654f984013140d52b9a344146e9e38e46fb81
[ "MIT" ]
16
2020-04-08T10:46:39.000Z
2021-11-15T03:46:56.000Z
# Reverse a linked list # Developer: Murillo Grubler # https://www.hackerrank.com/challenges/reverse-a-linked-list/problem # Time complexity of reverse function: O(n) class SinglyLinkedListNode: def __init__(self, node_data): self.data = node_data self.next = None class SinglyLinkedList: def __init__(self): self.head = None self.tail = None def insert_node(self, node_data): node = SinglyLinkedListNode(node_data) if not self.head: self.head = node else: self.tail.next = node self.tail = node # Complete the reverse function below. # # For your reference: # # SinglyLinkedListNode: # int data # SinglyLinkedListNode next # def reverse(head): ln = SinglyLinkedListNode(head.data) temp_node = head.next while temp_node: next_ln = ln ln = SinglyLinkedListNode(temp_node.data) ln.next = next_ln temp_node = temp_node.next return ln if __name__ == '__main__': tests = int(input()) for tests_itr in range(tests): llist_count = int(input()) llist = SinglyLinkedList() for _ in range(llist_count): llist_item = int(input()) llist.insert_node(llist_item) result = reverse(llist.head) while result: print (result.data, end=' ') result = result.next
24.596491
69
0.622682
430
0.306705
0
0
0
0
0
0
306
0.21826
4bb64f1dd8e15adacfcfa40dd94e5cebe3d88bea
1,737
py
Python
web/api/user/utilities.py
cclrobotics/ARTBot
a0bffabebbc09361bf7748741fe3d30c78af8fbd
[ "MIT" ]
5
2020-12-04T19:28:42.000Z
2021-12-07T16:14:28.000Z
web/api/user/utilities.py
cclrobotics/ARTBot
a0bffabebbc09361bf7748741fe3d30c78af8fbd
[ "MIT" ]
50
2019-10-08T19:47:24.000Z
2021-07-26T05:43:37.000Z
web/api/user/utilities.py
cclrobotics/ARTBot
a0bffabebbc09361bf7748741fe3d30c78af8fbd
[ "MIT" ]
4
2019-10-23T04:14:49.000Z
2021-08-01T01:22:37.000Z
import os from PIL import Image import random from functools import wraps from flask import jsonify from flask_jwt_extended import get_current_user from .artpiece import Artpiece from .exceptions import InvalidUsage from web.extensions import cache #decorator to require admin_acccess for a route def access_level_required(level): try: def outer(func): @wraps(func) def inner(*args, **kwargs): if get_current_user().role < level: raise InvalidUsage.forbidden() return func(*args, **kwargs) return inner except TypeError: raise TypeError("Specify an access level to use access_level_required decorator") return outer @cache.memoize(timeout=3600) def get_image_description(image_path): with Image.open(image_path) as image: # Exif ID 270 = ImageDescription return image.getexif().get(270) """ Return a list of images in the 'gallery' folder and their descriptions Output is list of tuples (image_location, image_description) output list is in random order for random display order every time """ def get_gallery_images(): internal_path_prefix = './web' public_gallery_path = '/static/img/gallery/' image_paths = [ public_gallery_path + filename for filename in os.listdir(internal_path_prefix + public_gallery_path) ] image_descriptions = list() for image_path in image_paths: this_image_description = get_image_description(internal_path_prefix + image_path) image_descriptions.append(this_image_description) image_metadata = list(zip(image_paths, image_descriptions)) random.shuffle(image_metadata) return image_metadata
31.017857
89
0.716753
0
0
0
0
390
0.224525
0
0
382
0.219919
4bb717792f0ab03afa44f642bc10364fd9b57993
2,528
py
Python
network/utils.py
Goochaozheng/ChunkFusion
7458a8e08886cc76cfeb87881c51e23b1d0674c3
[ "MIT" ]
3
2022-03-15T08:34:15.000Z
2022-03-15T08:40:06.000Z
network/utils.py
Goochaozheng/ChunkFusion
7458a8e08886cc76cfeb87881c51e23b1d0674c3
[ "MIT" ]
null
null
null
network/utils.py
Goochaozheng/ChunkFusion
7458a8e08886cc76cfeb87881c51e23b1d0674c3
[ "MIT" ]
null
null
null
import spconv import torch from torch import nn def residualBlock(channels, kernel_size=3): return spconv.SparseSequential( spconv.ConcatTable() .add(spconv.Identity()) .add(spconv.SparseSequential( nn.BatchNorm1d(channels), spconv.SubMConv3d(channels, channels, kernel_size=kernel_size), nn.Sigmoid() )), spconv.JoinTable() ) def subMVonvBlock(inChannels, outChannels, kernel_size=3, indiceKey=None): return spconv.SparseSequential( nn.BatchNorm1d(inChannels), spconv.SubMConv3d(inChannels, outChannels, kernel_size=kernel_size, indice_key=indiceKey), nn.Sigmoid() ) def convBlock(inChannels, outChannels, kernel_size=3): return nn.Sequential( nn.BatchNorm3d(inChannels), nn.Conv3d(inChannels, outChannels, kernel_size=kernel_size, stride=1, padding=1), nn.LeakyReLU() ) def kaiming_init(m): classname = m.__class__.__name__ if classname.find('Conv') != -1: torch.nn.init.kaiming_uniform_(m.weight) def toSparseInput(inputTSDF): # Construct Sparse Tensor inputTSDF = inputTSDF.permute(0,2,3,4,1) sparseMask = torch.any(torch.abs(inputTSDF) < 1, dim=4) batchSize = len(inputTSDF) spatialShape = inputTSDF.shape[1:-1] sparseIndice = sparseMask.to_sparse(inputTSDF.ndim-1).indices().permute(1, 0).contiguous().int() sparseValue = inputTSDF[sparseMask] inputData_sparse = spconv.SparseConvTensor(features=sparseValue, indices=sparseIndice, spatial_shape=spatialShape, batch_size=batchSize) return inputData_sparse def sparseFuse(inputSparseTSDF, oldSparseTSDF, inputMask, oldMask): # fuseTSDF = torch.cat((self.toDense(inputSparseTSDF), self.toDense(oldSparseTSDF)), dim=1) oldTSDF = spconv.ToDense(oldSparseTSDF).permute(0,2,3,4,1) inputTSDF = spconv.ToDense(inputSparseTSDF).permute(0,2,3,4,1) # oldTSDF[inputMask] = (oldTSDF[inputMask] * oldWeight[inputMask] + inputTSDF[inputMask] * inputWeight[inputMask]) / (oldWeight[inputMask] + inputWeight[inputMask]) batchSize = inputSparseTSDF.batch_size spatialShape = inputSparseTSDF.spatial_shape fuseMask = torch.logical_or(inputMask, oldMask) sparseIndice = fuseMask.to_sparse(oldTSDF.ndim-1).indices().permute(1, 0).contiguous().int() sparseValue = oldTSDF[fuseMask] return spconv.SparseConvTensor(features=sparseValue, indices=sparseIndice, spatial_shape=spatialShape, batch_size=batchSize)
36.114286
168
0.71163
0
0
0
0
0
0
0
0
286
0.113133
4bba68abed889d99f735d0534602287dd744310e
3,794
py
Python
hemlock/load_scripts/doc_to_mongo.py
Lab41/Hemlock
2c53cfc11bfbe1e4f901b519db578090fe7a17dd
[ "Apache-2.0" ]
4
2015-05-14T18:59:44.000Z
2017-03-09T12:49:36.000Z
hemlock/load_scripts/doc_to_mongo.py
Lab41/Hemlock
2c53cfc11bfbe1e4f901b519db578090fe7a17dd
[ "Apache-2.0" ]
null
null
null
hemlock/load_scripts/doc_to_mongo.py
Lab41/Hemlock
2c53cfc11bfbe1e4f901b519db578090fe7a17dd
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python # # Copyright (c) 2013 In-Q-Tel, Inc/Lab41, All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import fnmatch, os, sys, time, uuid from pymongo import MongoClient def mongo_server(server, port, database, collection): # connect to the redis server try: m_server = MongoClient(server, port) m_database = m_server[database] m_collection = m_database[collection] except: print "Mongo server failure" sys.exit(0) return m_server, m_database, m_collection def process_doc(input, m_server, m_database, m_collection): matches = [] docs = [] for root, dirnames, filenames in os.walk(input): for filename in fnmatch.filter(filenames, '*.txt'): matches.append(os.path.join(root, filename)) j = 0 k = 0 for file in matches: if len(docs) % 100 == 0 and len(docs) > 0: m_collection.insert(docs) print str(j), "total docs." print str(k), "docs failed." docs = [] doc = open(file, 'r').read() try: doc = unicode(doc, "utf-8") doc = {"doc": doc} docs.append(doc) j += 1 except: k += 1 if len(docs) > 0: m_collection.insert(docs) print str(j), "total docs." print str(k), "docs failed." def print_help(): print "-i \t<input path to files> (default is /mnt/)" print "-s \t<mongo server> (default is localhost)" print "-p \t<mongo port> (default is 27017)" print "-d \t<mongo database> (default is local)" print "-c \t<mongo collection> (default is collection)" print "-h \thelp\n" sys.exit(0) def process_args(args): # default initialization input = "/mnt/" server = "localhost" port = 27017 database = "local" collection = "collection" # process args i = 0 while i < len(args): if args[i] == "-s": try: server = args[i+1] i += 1 except: print_help() elif args[i] == "-p": try: port = int(args[i+1]) i += 1 except: print_help() elif args[i] == "-d": try: database = args[i+1] i += 1 except: print_help() elif args[i] == "-c": try: collection = args[i+1] i += 1 except: print_help() elif args[i] == "-i": try: input = args[i+1] i += 1 except: print_help() else: print_help() i += 1 return input, server, port, database, collection def get_args(): args = [] for arg in sys.argv: args.append(arg) return args[1:] if __name__ == "__main__": start_time = time.time() args = get_args() input, server, port, database, collection = process_args(args) m_server, m_database, m_collection = mongo_server(server, port, database, collection) process_doc(input, m_server, m_database, m_collection) print "Took",time.time() - start_time,"seconds to complete."
29.640625
89
0.552978
0
0
0
0
0
0
0
0
1,131
0.298102
4bbaecaa33cf5b0c99d08e0e5f803ac656d6dabe
2,659
py
Python
unn/models/initializer.py
zongdaoming/TinyTransformer
8e64f8816117048c388b4b20e3a56760ce149fe3
[ "Apache-2.0" ]
2
2021-08-08T11:23:14.000Z
2021-09-16T04:05:23.000Z
unn/models/initializer.py
zongdaoming/TinyTransformer
8e64f8816117048c388b4b20e3a56760ce149fe3
[ "Apache-2.0" ]
1
2021-08-08T11:25:47.000Z
2021-08-08T11:26:15.000Z
unn/models/initializer.py
zongdaoming/TinyTransformer
8e64f8816117048c388b4b20e3a56760ce149fe3
[ "Apache-2.0" ]
null
null
null
import copy import logging import math import torch from torch import nn logger = logging.getLogger('global') def init_weights_normal(module, std=0.01): for m in module.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Linear) or isinstance( m, nn.ConvTranspose2d): nn.init.normal_(m.weight.data, std=std) if m.bias is not None: m.bias.data.zero_() def init_weights_xavier(module): for m in module.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Linear) or isinstance( m, nn.ConvTranspose2d): nn.init.xavier_normal_(m.weight.data) if m.bias is not None: m.bias.data.zero_() def init_weights_msra(module): for m in module.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Linear) or isinstance( m, nn.ConvTranspose2d): nn.init.kaiming_normal_(m.weight.data) if m.bias is not None: m.bias.data.zero_() def init_bias_focal(module, cls_loss_type, init_prior, num_classes): if cls_loss_type == 'sigmoid': for m in module.modules(): if isinstance(m, nn.Conv2d): # to keep the torch random state m.bias.data.normal_(-math.log(1.0 / init_prior - 1.0), init_prior) torch.nn.init.constant_(m.bias, -math.log(1.0 / init_prior - 1.0)) elif cls_loss_type == 'softmax': for m in module.modules(): if isinstance(m, nn.Conv2d): m.bias.data.normal_(0, 0.01) for i in range(0, m.bias.data.shape[0], num_classes): fg = m.bias.data[i + 1:i + 1 + num_classes - 1] mu = torch.exp(fg).sum() m.bias.data[i] = math.log(mu * (1.0 - init_prior) / init_prior) else: raise NotImplementedError(f'{cls_loss_type} is not supported') def initialize(model, method, **kwargs): # initialize BN for m in model.modules(): if isinstance(m, nn.BatchNorm2d): m.weight.data.fill_(1) m.bias.data.zero_() # initialize Conv & FC if method == 'normal': init_weights_normal(model, **kwargs) elif method == 'msra': init_weights_msra(model) elif method == 'xavier': init_weights_xavier(model) else: raise NotImplementedError(f'{method} not supported') def initialize_from_cfg(model, cfg): if cfg is None: initialize(model, 'normal', std=0.01) return cfg = copy.deepcopy(cfg) method = cfg.pop('method') initialize(model, method, **cfg)
32.036145
83
0.588943
0
0
0
0
0
0
0
0
193
0.072584
4bbcb3d3943aa14ce46dab08f6f7c37762566694
3,000
py
Python
AudioFile.py
ZZZlax/.Pyrate
42a85213e0557b2988bf62bb8eac540263e0ce30
[ "Unlicense" ]
null
null
null
AudioFile.py
ZZZlax/.Pyrate
42a85213e0557b2988bf62bb8eac540263e0ce30
[ "Unlicense" ]
null
null
null
AudioFile.py
ZZZlax/.Pyrate
42a85213e0557b2988bf62bb8eac540263e0ce30
[ "Unlicense" ]
null
null
null
# This Python file uses the following encoding: utf-8 import os; import sys; import urllib.request; from bs4 import BeautifulSoup; import wikipedia from PyQt5.QtWebEngineWidgets import *; from PyQt5.QtGui import QIcon; from PyQt5.QtWidgets import *; from PyQt5.QtNetwork import QNetworkProxy class AudioFile(QMainWindow): def __init__(self): QMainWindow.__init__(self) proxy = QNetworkProxy() proxy.setType(1); proxy.setHostName("127.0.0.1"); proxy.setPort(9050); proxy.setApplicationProxy(proxy) toolb = QToolBar("URL"); self.lineEdit = QLineEdit(self); self.lineEdit2 = QLineEdit(self); combobox2 = QComboBox(self); self.textEdit = QTextEdit(self) def track_search(): try: soup = BeautifulSoup(urllib.request.urlopen(wikipedia.page(str(self.lineEdit.text()+" "+self.lineEdit2.text())).url).read()) except: soup = BeautifulSoup(urllib.request.urlopen(wikipedia.page(str(self.lineEdit.text()+" "+self.lineEdit2.text()+" album")).url).read()) for link in soup.find_all("td"): w = link.get_text().strip() if w[:1] == '"': self.textEdit.append(w.replace('"', "")) else: pass self.lineEdit2.clear() def save(): bandName = self.lineEdit.text(); script = self.textEdit.toPlainText() with open(os.getcwd()+"/.Pyrate/Set_List/"+bandName+".txt", "w", encoding = 'utf-8') as file: file.write(script) file.close() self.textEdit.clear(); self.lineEdit.clear() def text_changed(): nav = combobox2.currentText() if nav == "Save": save() if nav == "Album Search": track_search() combobox2.setCurrentText("") toolb.setOrientation(0x2); self.addToolBar(toolb); self.lineEdit.setObjectName(u"Artist Name"); self.lineEdit.setPlaceholderText("Artist Name"); toolb.addWidget(self.lineEdit); self.lineEdit2.setObjectName(u"Album Name"); self.lineEdit2.setPlaceholderText("Album Name"); toolb.addWidget(self.lineEdit2); combobox2.addItems(["", "Album Search", "Save"]); combobox2.currentTextChanged.connect(lambda: text_changed()); toolb.addWidget(combobox2); self.textEdit.setObjectName(u"Track List"); self.textEdit.setPlaceholderText("Track List"); self.textEdit.setAcceptRichText(False); self.setCentralWidget(self.textEdit) self.setWindowIcon(QIcon(os.getcwd()+'/.Pyrate/.images/pp.png')); self.setWindowTitle("Shanties"); self.setStyleSheet("color: #fe2023;" "background-color: #000000;" "selection-color: #ffffff;" "selection-background-color: #e01b24;"); self.lineEdit.setStyleSheet("background-color: #ffffff;" "selection-color: #000000;");self.lineEdit2.setStyleSheet("background-color: #ffffff;" "selection-color: #000000;"); self.textEdit.setStyleSheet("background-color: #ffffff;" "selection-color: #000000;") if __name__ == "__main__": app = QApplication([]) AudioFile().show() sys.exit(app.exec_())
69.767442
620
0.676667
2,600
0.866667
0
0
0
0
0
0
572
0.190667
4bbd0476d4a8b8dde9a872c84a83e121621a1703
16,649
py
Python
GeneratorTest.py
Autio/swb_datascraping
ef31fd89c68d86849342495b79985572d0f2fc61
[ "MIT" ]
null
null
null
GeneratorTest.py
Autio/swb_datascraping
ef31fd89c68d86849342495b79985572d0f2fc61
[ "MIT" ]
null
null
null
GeneratorTest.py
Autio/swb_datascraping
ef31fd89c68d86849342495b79985572d0f2fc61
[ "MIT" ]
null
null
null
__author__ = 'petriau' import requests # for HTTP requests from bs4 import BeautifulSoup # for HTML parsing url_SBM_FinanceProgress = 'http://sbm.gov.in/sbmreport/Report/Financial/SBM_StateReleaseAllocationincludingUnapproved.aspx' # Function to return HTML parsed with BeautifulSoup from a POST request URL and parameters. def parsePOSTResponse(URL, parameters=''): responseHTMLParsed = '' r = requests.post(URL, data=parameters) if r.status_code == 200: responseHTML = r.content responseHTMLParsed = BeautifulSoup(responseHTML, 'html.parser') return responseHTMLParsed listTest = [['__EVENTARGUMENT',''],['__EVENTTARGET', 'ctl00$ContentPlaceHolder1$rptr_state$ctl03$lnkbtn_stName'],['__EVENTVALIDATION',"/wEWTAL2js/IBwLq6fiEBwK4qJKGBgL7uLfDBQLMho26CAKkvMv0BAKrp/OzCAKzzOWcCQLfxNm+CQLZ25fbDALc9b7CDALYxrzSBgLK5tedAQLJrc6KBwLD2Nb1DwKAz9S2BQLD2JrzAgKAz/jyBAKUlKjOAgKvk9PyBQKUlOzLBQKvk5ewAQKNhuujBAK2ocCIDQKNhq+hBwK2oaT3BwLW5PDOBwLdiLPJCQLW5LTMCgLdiPeGBQLPqct8AqCWu5oDAs+pj/oDAqCW/9cOArTAhacLArv/sNsBArTAyaQOArv/1JcBApnX36oFAtbopuwKApnXo6gIAtbo6qkGAv7tmdUPAvHRnK0JAv7t3dICAvHRgJwEAtuMv40FAoTehsMOAtuMg4sIAoTeyoAKAtTR8eIFAquz5dgPAtTRhc0LAquz+e4CAs3DtJABAoKu9fINAs3DyPoGAoKuibEHAp7/hZEKAuGFkd0CAp7/mfsPAuGFpfMFApfxyL4FAriAoXcCl/HcqAsCuIC1tQoC4M/OkQ8Cv46O1Q0C4M/i+wQCv46iawLZlNH1CQKqnLrXDQLZlJXzDAKqnM5tAr6ri/gNAsWF0MkLUJ4OhBgatkYSQhamBAvcsSVIgC8="], ['__VIEWSTATE',"/wEPDwUKMTQwNTE3ODMyMg9kFgJmD2QWAgIDD2QWBAIfDw8WBB4EVGV4dAVMPHNwYW4gY2xhc3M9ImdseXBoaWNvbiBnbHlwaGljb24tY2lyY2xlLWFycm93LWxlZnQiPjwvc3Bhbj4gQmFjayB0byBQcmV2aW91cx4HVmlzaWJsZWcWAh4Hb25jbGljawUoamF2YXNjcmlwdDpoaXN0b3J5LmJhY2soKTsgcmV0dXJuIGZhbHNlO2QCIQ9kFgICAQ9kFggCAw8PFgIfAAUIKENlbnRlcilkZAIFDw8WAh8ABQsoMjAxNi0yMDE3KWRkAgcPFgIfAWgWBAIBDxYCHwFoFgQCAw8QZGQWAWZkAgcPEA8WBh4NRGF0YVRleHRGaWVsZAUHRmluWWVhch4ORGF0YVZhbHVlRmllbGQFB0ZpblllYXIeC18hRGF0YUJvdW5kZ2QQFQIKLS1TZWxlY3QtLQkyMDE2LTIwMTcVAgItMgkyMDE2LTIwMTcUKwMCZ2cWAQIBZAIDD2QWAgIBDw8WBB8ABRRSZWNvcmQgTm90IEZvdW5kICEhIR8BaGRkAgkPFgIeC18hSXRlbUNvdW50AiAWQgIBD2QWBmYPFQEBMWQCAQ8PFgIfAAUMQSAmIE4gSWxhbmRzZGQCBA8VCgYzNDAuMDAEMC4wMAQwLjAwBjM0MC4wMAQwLjAwBTEzLjU5BTEzLjU5BDQuMDAGMzQwLjAwBjMyNi40MWQCAg9kFgZmDxUBATJkAgEPDxYCHwAFDkFuZGhyYSBQcmFkZXNoZGQCBA8VCgc4NTk2LjY5BzY3NzIuODUEMS4xNAgxNTM3MC42OAc1NjQ5LjkzBzMzNDMuNjEHODk5My41NAU1OC41MQc5NzIwLjc1BzYzNzcuMTRkAgMPZBYGZg8VAQEzZAIBDw8WAh8ABRFBcnVuYWNoYWwgUHJhZGVzaGRkAgQPFQoHMTQ2NS44OAY5NjguNTEEMC4wMAcyNDM0LjM5BDAuMDAGMTA4LjAzBjEwOC4wMwQ0LjQ0BzI0MzQuMzkHMjMyNi4zNWQCBA9kFgZmDxUBATRkAgEPDxYCHwAFBUFzc2FtZGQCBA8VCggxNjExMC43OAQwLjAwBDAuMDAIMTYxMTAuNzgGNjg2LjE5BjkxNi4yNwcxNjAyLjQ2BDkuOTUIMTU0MjQuNjAIMTQ1MDguMzJkAgUPZBYGZg8VAQE1ZAIBDw8WAh8ABQVCaWhhcmRkAgQPFQoHNDIwMC4zNQgxMzE4Ni4zNwQwLjAwCDE3Mzg2LjcyBjY4Ni45OAcxMjI2LjgwBzE5MTMuNzgFMTEuMDEIMTY2OTkuNzQIMTU0NzIuOTRkAgYPZBYGZg8VAQE2ZAIBDw8WAh8ABQxDaGhhdHRpc2dhcmhkZAIEDxUKCC01OTYyLjUxBzk5NDcuNTcEMC4wMAczOTg1LjA2BjU3MS4xNgcxODY5LjE5BzI0NDAuMzQFNjEuMjQHMzQxMy45MQcxNTQ0LjcyZAIHD2QWBmYPFQEBN2QCAQ8PFgIfAAUMRCAmIE4gSGF2ZWxpZGQCBA8VCgQxLjQ4BDAuMDAEMC4wMAQxLjQ4BDAuMDAEMC4wMAQwLjAwBDAuMDAEMS40OAQxLjQ4ZAIID2QWBmYPFQEBOGQCAQ8PFgIfAAUDR29hZGQCBA8VCgctMzMzLjk1BDAuMDAEMC4wMActMzMzLjk1BDAuMDAHMjA5NC40OAcyMDk0LjQ4BDAuMDAHLTMzMy45NQgtMjQyOC40M2QCCQ9kFgZmDxUBATlkAgEPDxYCHwAFB0d1amFyYXRkZAIEDxUKCC00Njg4LjA0CDI4MDQ5LjI2BDAuMjMIMjMzNjEuNDUHMjAwNS4zNgc0MTc5LjAzBzYxODQuMzkFMjYuNDcIMjEzNTYuMDgIMTcxNzcuMDZkAgoPZBYGZg8VAQIxMGQCAQ8PFgIfAAUHSGFyeWFuYWRkAgQPFQoGNzc0LjQ5BzY4NzkuMDcEMi4zNQc3NjU1LjkyBjIwOC40MgU5MS42MQYzMDAuMDMEMy45Mgc3NDQ3LjUwBzczNTUuODlkAgsPZBYGZg8VAQIxMWQCAQ8PFgIfAAUQSGltYWNoYWwgUHJhZGVzaGRkAgQPFQoHNTI4My4yOAQwLjAwBTI0LjAzBzUzMDcuMzEGMzEzLjY0BjY2Ni41NgY5ODAuMjAFMTguNDcHNDk5My42Nwc0MzI3LjExZAIMD2QWBmYPFQECMTJkAgEPDxYCHwAFD0phbW11ICYgS2FzaG1pcmRkAgQPFQoHNTM5OS4zNwYyMjkuOTAEMC4wMAc1NjI5LjI3BjEwMS43MQU1MS44NQYxNTMuNTYEMi43Mwc1NTI3LjU1BzU0NzUuNzBkAg0PZBYGZg8VAQIxM2QCAQ8PFgIfAAUJSmhhcmtoYW5kZGQCBA8VCgktMTIyNzYuNjMIMTAzNTguOTYENC4xMggtMTkxMy41NQcxMzkwLjc2BzIzNTcuMjIHMzc0Ny45OAQwLjAwCC0zMzA0LjMxCC01NjYxLjUyZAIOD2QWBmYPFQECMTRkAgEPDxYCHwAFCUthcm5hdGFrYWRkAgQPFQoILTUwNDAuNjQIMTI2NzEuNTAEMC4wMAc3NjMwLjg2Bjk0OS40MwczMzA1LjYyBzQyNTUuMDUFNTUuNzYHNjY4MS40MwczMzc1LjgxZAIPD2QWBmYPFQECMTVkAgEPDxYCHwAFBktlcmFsYWRkAgQPFQoHMjg5MC45MgQwLjAwBDIuODIHMjg5My43NAYxMDcuNjkENS4xMQYxMTIuODAEMy45MAcyNzg2LjA1BzI3ODAuOTRkAhAPZBYGZg8VAQIxNmQCAQ8PFgIfAAUOTWFkaHlhIFByYWRlc2hkZAIEDxUKCS0xNTYzMy43NAgzNDIyMy41MwUyNS4wMAgxODYxNC43OQc5MzYwLjU0BzM0NzIuOTUIMTI4MzMuNDkFNjguOTQHOTI1NC4yNAc1NzgxLjI5ZAIRD2QWBmYPFQECMTdkAgEPDxYCHwAFC01haGFyYXNodHJhZGQCBA8VCggtNDMzMy4xNwgyNjQ0Ny4wOQQwLjAwCDIyMTEzLjkyBjMyNy42OAczNDg5LjAxBzM4MTYuNjkFMTcuMjYIMjE3ODYuMjMIMTgyOTcuMjNkAhIPZBYGZg8VAQIxOGQCAQ8PFgIfAAUHTWFuaXB1cmRkAgQPFQoHLTQ2Ni4yOQcyNzI3LjUwBDAuMDAHMjI2MS4yMQQwLjAwBjE1NS42MwYxNTUuNjMENi44OAcyMjYxLjIxBzIxMDUuNThkAhMPZBYGZg8VAQIxOWQCAQ8PFgIfAAUJTWVnaGFsYXlhZGQCBA8VCgcxNzI3LjY3BzQxMjIuMjQEMC4wMAc1ODQ5LjkxBjIyOS42MAYxMDguMjUGMzM3Ljg1BDUuNzgHNTYyMC4zMQc1NTEyLjA2ZAIUD2QWBmYPFQECMjBkAgEPDxYCHwAFB01pem9yYW1kZAIEDxUKBjM2NC4zMwQwLjAwBDAuMDAGMzY0LjMzBTk1LjExBTczLjgyBjE2OC45MwU0Ni4zNwYyNjkuMjMGMTk1LjQwZAIVD2QWBmYPFQECMjFkAgEPDxYCHwAFCE5hZ2FsYW5kZGQCBA8VCgYzMDIuMDMEMC4wMAQ5Ljg3BjMxMS45MAYxNzMuNDMEMi4yNwYxNzUuNzAFNTYuMzMGMTM4LjQ3BjEzNi4yMGQCFg9kFgZmDxUBAjIyZAIBDw8WAh8ABQZPZGlzaGFkZAIEDxUKCS01MDYzMS40Nwg0NTg1Ni42MQQwLjAwCC00Nzc0Ljg2Bzk1MTAuMzgHMjI4MC4zNQgxMTc5MC43MwQwLjAwCS0xNDI4NS4yNAktMTY1NjUuNTlkAhcPZBYGZg8VAQIyM2QCAQ8PFgIfAAUKUHVkdWNoZXJyeWRkAgQPFQoGNjYzLjEyBDAuMDAEMC4wMAY2NjMuMTIEMC4wMAQwLjAwBDAuMDAEMC4wMAY2NjMuMTIGNjYzLjEyZAIYD2QWBmYPFQECMjRkAgEPDxYCHwAFBlB1bmphYmRkAgQPFQoILTE2NTUuMjkHMjQ4Mi44NQQwLjAwBjgyNy41NgYxNTQuOTIGNTE4LjkwBjY3My44MwU4MS40MgY2NzIuNjMGMTUzLjczZAIZD2QWBmYPFQECMjVkAgEPDxYCHwAFCVJhamFzdGhhbmRkAgQPFQoJLTMwNTk3LjUwCDYyNzMwLjA1BDAuMDAIMzIxMzIuNTUHNjQxNC45Mwc1ODA4LjUyCDEyMjIzLjQ1BTM4LjA0CDI1NzE3LjYyCDE5OTA5LjEwZAIaD2QWBmYPFQECMjZkAgEPDxYCHwAFBlNpa2tpbWRkAgQPFQoGNTE1LjM5BjQ4MC45NgQwLjAwBjk5Ni4zNQQwLjAwBDAuMDAEMC4wMAQwLjAwBjk5Ni4zNQY5OTYuMzVkAhsPZBYGZg8VAQIyN2QCAQ8PFgIfAAUKVGFtaWwgTmFkdWRkAgQPFQoJLTI0MTEwLjAxCDI2ODUwLjk0BDAuNjIHMjc0MS41NgY0NzguMTEGMTU3Ljg5BjYzNi4wMQUyMy4yMAcyMjYzLjQ0BzIxMDUuNTVkAhwPZBYGZg8VAQIyOGQCAQ8PFgIfAAUJVGVsYW5nYW5hZGQCBA8VCgc1ODE0LjI1BDAuMDAEMC4wMAc1ODE0LjI1BjY1Ni43OAc0NjgwLjI0BzUzMzcuMDIFOTEuNzkHNTE1Ny40NwY0NzcuMjNkAh0PZBYGZg8VAQIyOWQCAQ8PFgIfAAUHVHJpcHVyYWRkAgQPFQoHMzYwNy40OAQwLjAwBDAuMDAHMzYwNy40OAU0MC4yMgYxMjguOTEGMTY5LjEzBDQuNjkHMzU2Ny4yNgczNDM4LjM2ZAIeD2QWBmYPFQECMzBkAgEPDxYCHwAFDVV0dGFyIFByYWRlc2hkZAIEDxUKCDI2OTIyLjIyCDE3ODY3LjQ0BDYuNjYINDQ3OTYuMzIHMzg1Ni44MAczNTE4LjMwBzczNzUuMTAFMTYuNDYINDA5MzkuNTIIMzc0MjEuMjFkAh8PZBYGZg8VAQIzMWQCAQ8PFgIfAAULVXR0YXJha2hhbmRkZAIEDxUKCC0xNjU4LjI3Bzg1MjkuMTMEMC4wMAc2ODcwLjg2BzEyMjAuMDkGNjQwLjc5BzE4NjAuODgFMjcuMDgHNTY1MC43Nwc1MDA5Ljk3ZAIgD2QWBmYPFQECMzJkAgEPDxYCHwAFC1dlc3QgQmVuZ2FsZGQCBA8VCgktMTYyMTkuNjcIMzI4NzUuNjAEMC4wMAgxNjY1NS45MwcxNzI3LjgxBzY2NDYuMTkHODM3NC4wMAU1MC4yOAgxNDkyOC4xMgc4MjgxLjkzZAIhD2QWAgICDxUKCS04ODYyNy40NQkzNTQyNTcuOTMFNzYuODQJMjY1NzA3LjM0CDQ2OTE3LjY3CDUxOTEwLjk5CDk4ODI4LjY3BTM3LjE5CTIxODc4OS42NAkxNjY4NzguNjRkGAEFHl9fQ29udHJvbHNSZXF1aXJlUG9zdEJhY2tLZXlfXxYDBQ9jdGwwMCRpY29uX3dvcmQFEGN0bDAwJGljb25fZXhjZWwFEmN0bDAwJGljb25fcHJpbnRlcqLkin/PLgDvwcsQ6/a18eF5HbFe"] ] paramDictionary = {key: str(value) for key, value in listTest} def merge_two_dicts(x, y): '''Given two dicts, merge them into a new dict as a shallow copy.''' z = x.copy() z.update(y) return z postParams = { # '__EVENTARGUMENT': '', # '__EVENTTARGET': 'ctl00$ContentPlaceHolder1$rptr_state$ctl03$lnkbtn_stName', # '__EVENTVALIDATION': "/wEWTAL2js/IBwLq6fiEBwK4qJKGBgL7uLfDBQLMho26CAKkvMv0BAKrp/OzCAKzzOWcCQLfxNm+CQLZ25fbDALc9b7CDALYxrzSBgLK5tedAQLJrc6KBwLD2Nb1DwKAz9S2BQLD2JrzAgKAz/jyBAKUlKjOAgKvk9PyBQKUlOzLBQKvk5ewAQKNhuujBAK2ocCIDQKNhq+hBwK2oaT3BwLW5PDOBwLdiLPJCQLW5LTMCgLdiPeGBQLPqct8AqCWu5oDAs+pj/oDAqCW/9cOArTAhacLArv/sNsBArTAyaQOArv/1JcBApnX36oFAtbopuwKApnXo6gIAtbo6qkGAv7tmdUPAvHRnK0JAv7t3dICAvHRgJwEAtuMv40FAoTehsMOAtuMg4sIAoTeyoAKAtTR8eIFAquz5dgPAtTRhc0LAquz+e4CAs3DtJABAoKu9fINAs3DyPoGAoKuibEHAp7/hZEKAuGFkd0CAp7/mfsPAuGFpfMFApfxyL4FAriAoXcCl/HcqAsCuIC1tQoC4M/OkQ8Cv46O1Q0C4M/i+wQCv46iawLZlNH1CQKqnLrXDQLZlJXzDAKqnM5tAr6ri/gNAsWF0MkLUJ4OhBgatkYSQhamBAvcsSVIgC8=", # '__VIEWSTATE': "/wEPDwUKMTQwNTE3ODMyMg9kFgJmD2QWAgIDD2QWBAIfDw8WBB4EVGV4dAVMPHNwYW4gY2xhc3M9ImdseXBoaWNvbiBnbHlwaGljb24tY2lyY2xlLWFycm93LWxlZnQiPjwvc3Bhbj4gQmFjayB0byBQcmV2aW91cx4HVmlzaWJsZWcWAh4Hb25jbGljawUoamF2YXNjcmlwdDpoaXN0b3J5LmJhY2soKTsgcmV0dXJuIGZhbHNlO2QCIQ9kFgICAQ9kFggCAw8PFgIfAAUIKENlbnRlcilkZAIFDw8WAh8ABQsoMjAxNi0yMDE3KWRkAgcPFgIfAWgWBAIBDxYCHwFoFgQCAw8QZGQWAWZkAgcPEA8WBh4NRGF0YVRleHRGaWVsZAUHRmluWWVhch4ORGF0YVZhbHVlRmllbGQFB0ZpblllYXIeC18hRGF0YUJvdW5kZ2QQFQIKLS1TZWxlY3QtLQkyMDE2LTIwMTcVAgItMgkyMDE2LTIwMTcUKwMCZ2cWAQIBZAIDD2QWAgIBDw8WBB8ABRRSZWNvcmQgTm90IEZvdW5kICEhIR8BaGRkAgkPFgIeC18hSXRlbUNvdW50AiAWQgIBD2QWBmYPFQEBMWQCAQ8PFgIfAAUMQSAmIE4gSWxhbmRzZGQCBA8VCgYzNDAuMDAEMC4wMAQwLjAwBjM0MC4wMAQwLjAwBTEzLjU5BTEzLjU5BDQuMDAGMzQwLjAwBjMyNi40MWQCAg9kFgZmDxUBATJkAgEPDxYCHwAFDkFuZGhyYSBQcmFkZXNoZGQCBA8VCgc4NTk2LjY5BzY3NzIuODUEMS4xNAgxNTM3MC42OAc1NjQ5LjkzBzMzNDMuNjEHODk5My41NAU1OC41MQc5NzIwLjc1BzYzNzcuMTRkAgMPZBYGZg8VAQEzZAIBDw8WAh8ABRFBcnVuYWNoYWwgUHJhZGVzaGRkAgQPFQoHMTQ2NS44OAY5NjguNTEEMC4wMAcyNDM0LjM5BDAuMDAGMTA4LjAzBjEwOC4wMwQ0LjQ0BzI0MzQuMzkHMjMyNi4zNWQCBA9kFgZmDxUBATRkAgEPDxYCHwAFBUFzc2FtZGQCBA8VCggxNjExMC43OAQwLjAwBDAuMDAIMTYxMTAuNzgGNjg2LjE5BjkxNi4yNwcxNjAyLjQ2BDkuOTUIMTU0MjQuNjAIMTQ1MDguMzJkAgUPZBYGZg8VAQE1ZAIBDw8WAh8ABQVCaWhhcmRkAgQPFQoHNDIwMC4zNQgxMzE4Ni4zNwQwLjAwCDE3Mzg2LjcyBjY4Ni45OAcxMjI2LjgwBzE5MTMuNzgFMTEuMDEIMTY2OTkuNzQIMTU0NzIuOTRkAgYPZBYGZg8VAQE2ZAIBDw8WAh8ABQxDaGhhdHRpc2dhcmhkZAIEDxUKCC01OTYyLjUxBzk5NDcuNTcEMC4wMAczOTg1LjA2BjU3MS4xNgcxODY5LjE5BzI0NDAuMzQFNjEuMjQHMzQxMy45MQcxNTQ0LjcyZAIHD2QWBmYPFQEBN2QCAQ8PFgIfAAUMRCAmIE4gSGF2ZWxpZGQCBA8VCgQxLjQ4BDAuMDAEMC4wMAQxLjQ4BDAuMDAEMC4wMAQwLjAwBDAuMDAEMS40OAQxLjQ4ZAIID2QWBmYPFQEBOGQCAQ8PFgIfAAUDR29hZGQCBA8VCgctMzMzLjk1BDAuMDAEMC4wMActMzMzLjk1BDAuMDAHMjA5NC40OAcyMDk0LjQ4BDAuMDAHLTMzMy45NQgtMjQyOC40M2QCCQ9kFgZmDxUBATlkAgEPDxYCHwAFB0d1amFyYXRkZAIEDxUKCC00Njg4LjA0CDI4MDQ5LjI2BDAuMjMIMjMzNjEuNDUHMjAwNS4zNgc0MTc5LjAzBzYxODQuMzkFMjYuNDcIMjEzNTYuMDgIMTcxNzcuMDZkAgoPZBYGZg8VAQIxMGQCAQ8PFgIfAAUHSGFyeWFuYWRkAgQPFQoGNzc0LjQ5BzY4NzkuMDcEMi4zNQc3NjU1LjkyBjIwOC40MgU5MS42MQYzMDAuMDMEMy45Mgc3NDQ3LjUwBzczNTUuODlkAgsPZBYGZg8VAQIxMWQCAQ8PFgIfAAUQSGltYWNoYWwgUHJhZGVzaGRkAgQPFQoHNTI4My4yOAQwLjAwBTI0LjAzBzUzMDcuMzEGMzEzLjY0BjY2Ni41NgY5ODAuMjAFMTguNDcHNDk5My42Nwc0MzI3LjExZAIMD2QWBmYPFQECMTJkAgEPDxYCHwAFD0phbW11ICYgS2FzaG1pcmRkAgQPFQoHNTM5OS4zNwYyMjkuOTAEMC4wMAc1NjI5LjI3BjEwMS43MQU1MS44NQYxNTMuNTYEMi43Mwc1NTI3LjU1BzU0NzUuNzBkAg0PZBYGZg8VAQIxM2QCAQ8PFgIfAAUJSmhhcmtoYW5kZGQCBA8VCgktMTIyNzYuNjMIMTAzNTguOTYENC4xMggtMTkxMy41NQcxMzkwLjc2BzIzNTcuMjIHMzc0Ny45OAQwLjAwCC0zMzA0LjMxCC01NjYxLjUyZAIOD2QWBmYPFQECMTRkAgEPDxYCHwAFCUthcm5hdGFrYWRkAgQPFQoILTUwNDAuNjQIMTI2NzEuNTAEMC4wMAc3NjMwLjg2Bjk0OS40MwczMzA1LjYyBzQyNTUuMDUFNTUuNzYHNjY4MS40MwczMzc1LjgxZAIPD2QWBmYPFQECMTVkAgEPDxYCHwAFBktlcmFsYWRkAgQPFQoHMjg5MC45MgQwLjAwBDIuODIHMjg5My43NAYxMDcuNjkENS4xMQYxMTIuODAEMy45MAcyNzg2LjA1BzI3ODAuOTRkAhAPZBYGZg8VAQIxNmQCAQ8PFgIfAAUOTWFkaHlhIFByYWRlc2hkZAIEDxUKCS0xNTYzMy43NAgzNDIyMy41MwUyNS4wMAgxODYxNC43OQc5MzYwLjU0BzM0NzIuOTUIMTI4MzMuNDkFNjguOTQHOTI1NC4yNAc1NzgxLjI5ZAIRD2QWBmYPFQECMTdkAgEPDxYCHwAFC01haGFyYXNodHJhZGQCBA8VCggtNDMzMy4xNwgyNjQ0Ny4wOQQwLjAwCDIyMTEzLjkyBjMyNy42OAczNDg5LjAxBzM4MTYuNjkFMTcuMjYIMjE3ODYuMjMIMTgyOTcuMjNkAhIPZBYGZg8VAQIxOGQCAQ8PFgIfAAUHTWFuaXB1cmRkAgQPFQoHLTQ2Ni4yOQcyNzI3LjUwBDAuMDAHMjI2MS4yMQQwLjAwBjE1NS42MwYxNTUuNjMENi44OAcyMjYxLjIxBzIxMDUuNThkAhMPZBYGZg8VAQIxOWQCAQ8PFgIfAAUJTWVnaGFsYXlhZGQCBA8VCgcxNzI3LjY3BzQxMjIuMjQEMC4wMAc1ODQ5LjkxBjIyOS42MAYxMDguMjUGMzM3Ljg1BDUuNzgHNTYyMC4zMQc1NTEyLjA2ZAIUD2QWBmYPFQECMjBkAgEPDxYCHwAFB01pem9yYW1kZAIEDxUKBjM2NC4zMwQwLjAwBDAuMDAGMzY0LjMzBTk1LjExBTczLjgyBjE2OC45MwU0Ni4zNwYyNjkuMjMGMTk1LjQwZAIVD2QWBmYPFQECMjFkAgEPDxYCHwAFCE5hZ2FsYW5kZGQCBA8VCgYzMDIuMDMEMC4wMAQ5Ljg3BjMxMS45MAYxNzMuNDMEMi4yNwYxNzUuNzAFNTYuMzMGMTM4LjQ3BjEzNi4yMGQCFg9kFgZmDxUBAjIyZAIBDw8WAh8ABQZPZGlzaGFkZAIEDxUKCS01MDYzMS40Nwg0NTg1Ni42MQQwLjAwCC00Nzc0Ljg2Bzk1MTAuMzgHMjI4MC4zNQgxMTc5MC43MwQwLjAwCS0xNDI4NS4yNAktMTY1NjUuNTlkAhcPZBYGZg8VAQIyM2QCAQ8PFgIfAAUKUHVkdWNoZXJyeWRkAgQPFQoGNjYzLjEyBDAuMDAEMC4wMAY2NjMuMTIEMC4wMAQwLjAwBDAuMDAEMC4wMAY2NjMuMTIGNjYzLjEyZAIYD2QWBmYPFQECMjRkAgEPDxYCHwAFBlB1bmphYmRkAgQPFQoILTE2NTUuMjkHMjQ4Mi44NQQwLjAwBjgyNy41NgYxNTQuOTIGNTE4LjkwBjY3My44MwU4MS40MgY2NzIuNjMGMTUzLjczZAIZD2QWBmYPFQECMjVkAgEPDxYCHwAFCVJhamFzdGhhbmRkAgQPFQoJLTMwNTk3LjUwCDYyNzMwLjA1BDAuMDAIMzIxMzIuNTUHNjQxNC45Mwc1ODA4LjUyCDEyMjIzLjQ1BTM4LjA0CDI1NzE3LjYyCDE5OTA5LjEwZAIaD2QWBmYPFQECMjZkAgEPDxYCHwAFBlNpa2tpbWRkAgQPFQoGNTE1LjM5BjQ4MC45NgQwLjAwBjk5Ni4zNQQwLjAwBDAuMDAEMC4wMAQwLjAwBjk5Ni4zNQY5OTYuMzVkAhsPZBYGZg8VAQIyN2QCAQ8PFgIfAAUKVGFtaWwgTmFkdWRkAgQPFQoJLTI0MTEwLjAxCDI2ODUwLjk0BDAuNjIHMjc0MS41NgY0NzguMTEGMTU3Ljg5BjYzNi4wMQUyMy4yMAcyMjYzLjQ0BzIxMDUuNTVkAhwPZBYGZg8VAQIyOGQCAQ8PFgIfAAUJVGVsYW5nYW5hZGQCBA8VCgc1ODE0LjI1BDAuMDAEMC4wMAc1ODE0LjI1BjY1Ni43OAc0NjgwLjI0BzUzMzcuMDIFOTEuNzkHNTE1Ny40NwY0NzcuMjNkAh0PZBYGZg8VAQIyOWQCAQ8PFgIfAAUHVHJpcHVyYWRkAgQPFQoHMzYwNy40OAQwLjAwBDAuMDAHMzYwNy40OAU0MC4yMgYxMjguOTEGMTY5LjEzBDQuNjkHMzU2Ny4yNgczNDM4LjM2ZAIeD2QWBmYPFQECMzBkAgEPDxYCHwAFDVV0dGFyIFByYWRlc2hkZAIEDxUKCDI2OTIyLjIyCDE3ODY3LjQ0BDYuNjYINDQ3OTYuMzIHMzg1Ni44MAczNTE4LjMwBzczNzUuMTAFMTYuNDYINDA5MzkuNTIIMzc0MjEuMjFkAh8PZBYGZg8VAQIzMWQCAQ8PFgIfAAULVXR0YXJha2hhbmRkZAIEDxUKCC0xNjU4LjI3Bzg1MjkuMTMEMC4wMAc2ODcwLjg2BzEyMjAuMDkGNjQwLjc5BzE4NjAuODgFMjcuMDgHNTY1MC43Nwc1MDA5Ljk3ZAIgD2QWBmYPFQECMzJkAgEPDxYCHwAFC1dlc3QgQmVuZ2FsZGQCBA8VCgktMTYyMTkuNjcIMzI4NzUuNjAEMC4wMAgxNjY1NS45MwcxNzI3LjgxBzY2NDYuMTkHODM3NC4wMAU1MC4yOAgxNDkyOC4xMgc4MjgxLjkzZAIhD2QWAgICDxUKCS04ODYyNy40NQkzNTQyNTcuOTMFNzYuODQJMjY1NzA3LjM0CDQ2OTE3LjY3CDUxOTEwLjk5CDk4ODI4LjY3BTM3LjE5CTIxODc4OS42NAkxNjY4NzguNjRkGAEFHl9fQ29udHJvbHNSZXF1aXJlUG9zdEJhY2tLZXlfXxYDBQ9jdGwwMCRpY29uX3dvcmQFEGN0bDAwJGljb25fZXhjZWwFEmN0bDAwJGljb25fcHJpbnRlcqLkin/PLgDvwcsQ6/a18eF5HbFe", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl01$hfd_StateId':"26", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl02$hfd_StateId':"1", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl03$hfd_StateId':"2", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl04$hfd_StateId':"3", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl05$hfd_StateId':"4", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl06$hfd_StateId':"34", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl07$hfd_StateId':"28", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl08$hfd_StateId':"5", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl09$hfd_StateId':"6", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl10$hfd_StateId':"7", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl11$hfd_StateId':"8", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl12$hfd_StateId':"9", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl13$hfd_StateId':"35", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl14$hfd_StateId':"10", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl15$hfd_StateId':"11", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl16$hfd_StateId':"12", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl17$hfd_StateId':"13", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl18$hfd_StateId':"14", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl19$hfd_StateId':"15", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl20$hfd_StateId':"16", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl21$hfd_StateId':"17", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl22$hfd_StateId':"18", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl23$hfd_StateId':"32", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl24$hfd_StateId':"19", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl25$hfd_StateId':"20", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl26$hfd_StateId':"21", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl27$hfd_StateId':"22", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl28$hfd_StateId':"36", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl29$hfd_StateId':"23", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl30$hfd_StateId':"24", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl31$hfd_StateId':"33", 'ctl00$ContentPlaceHolder1$rptr_cen$ctl32$hfd_StateId':"25" } p = merge_two_dicts(postParams, paramDictionary) componentPage = parsePOSTResponse(url_SBM_FinanceProgress, p) print(componentPage) x = 'what'
228.068493
5,731
0.910145
0
0
0
0
0
0
0
0
15,104
0.907202
4bbd48777ac0547cad592918b3cc57a1efffc54a
677
py
Python
project euler/q50.py
milkmeat/thomas
fbc72af34267488d931a4885d4e19fce22fea582
[ "MIT" ]
null
null
null
project euler/q50.py
milkmeat/thomas
fbc72af34267488d931a4885d4e19fce22fea582
[ "MIT" ]
null
null
null
project euler/q50.py
milkmeat/thomas
fbc72af34267488d931a4885d4e19fce22fea582
[ "MIT" ]
null
null
null
def listprime(max): prime=[True]*max prime[0]=False prime[1]=False for x in range(max/2): if prime[x]: bei=x+x while bei<max: prime[bei]=False bei+=x return prime def listprimenumber(lpn): listprimelpn=listprime(lpn) l=[] for x in range(len(listprimelpn)): if listprimelpn[x]: l.append(x) return l b=listprimenumber(100) print b count=3 for x in range(len(b)-count): sum=0 for y in range(count): sum+=b[x+y] if sum in b: print sum #if b[x+0]+b[x+1]+b[x+2] in b: # print b[x],b[x+1],b[x+2]
22.566667
39
0.499261
0
0
0
0
0
0
0
0
60
0.088626
4bbd5d337a02e7405c19f8ae7746f2dbce197b3b
4,189
py
Python
s09_files_and_random/solutions/random_walking_simple.py
silverfield/pythonsessions
bf5d82dded7616a5d6998da4eb445708c728794f
[ "MIT" ]
null
null
null
s09_files_and_random/solutions/random_walking_simple.py
silverfield/pythonsessions
bf5d82dded7616a5d6998da4eb445708c728794f
[ "MIT" ]
null
null
null
s09_files_and_random/solutions/random_walking_simple.py
silverfield/pythonsessions
bf5d82dded7616a5d6998da4eb445708c728794f
[ "MIT" ]
null
null
null
__author__ = 'ferrard' # --------------------------------------------------------------- # Imports # --------------------------------------------------------------- import scipy as sp import random import time # --------------------------------------------------------------- # Class - Graph # --------------------------------------------------------------- class WalkableGraph: """Graph on which we can do random walking""" # --------------------------------------------------------------- # Initialisation # --------------------------------------------------------------- def __init__(self, file_path): """ Loads a graph from file The file should have format: CityName i_1 i_2 ... i_k .... where i_j are indices of neighbouring cities (index given by index of the row) """ self._graph = [] self._cities = [] with open(file_path, 'r') as f: for line in f: city = line.split(' ')[0] neighs = [int(s) for s in line.split(' ')[1:]] self._cities.append(city) self._graph.append(neighs) self.n = len(self._cities) self._transition_matrix = self.__get_transition_matrix() # --------------------------------------------------------------- # Interface # --------------------------------------------------------------- def print(self): """ Prints the neighbourhood table of the graph """ for i in range(self.n): print(str(i) + " " + self._cities[i] + " " + str(len(self._graph[i])) + " " + str(self._graph[i])) def probs_after_k_steps(self, k): """ Prints the probability (for each city) we end up in the city after k steps """ probs = (1/self.n)*sp.ones(self.n) for i in range(k): probs = sp.dot(self._transition_matrix, probs) print("Probabilities: ") for i in range(self.n): print("\t" + self._cities[i] + ": " + str(probs[i])) return probs def random_walk(self, start_city, steps=10, time_in_city=0): """ Does a random walk through the graph, starting at given city, making "steps" random steps and waiting in each city for time_in_city seconds """ # find the index of the start-city current_city_index = None for i in range(len(self._cities)): if self._cities[i] == start_city: current_city_index = i if current_city_index is None: raise Exception("Unknown city " + start_city) # do the random walking print("Random walk with " + str(steps) + " steps. Started in " + self._cities[current_city_index]) visits = [0]*self.n for i in range(steps): visits[current_city_index] += 1 current_city_index = random.choice(self._graph[current_city_index]) print("Moved to " + self._cities[current_city_index]) time.sleep(time_in_city) visits[current_city_index] += 1 # print the statistics print("Finished random walk in: " + self._cities[current_city_index]) print("Visits of cities: ") for i in range(self.n): print("\t%s: %s (%s)" % (self._cities[i], visits[i], visits[i]/steps)) # --------------------------------------------------------------- # Implementation # --------------------------------------------------------------- def __get_transition_matrix(self): """ Gets the transition matrix of the graph """ transition_matrix = sp.zeros((self.n, self.n)) for j in range(self.n): for i in self._graph[j]: transition_matrix[i][j] = 1/len(self._graph[j]) return transition_matrix # --------------------------------------------------------------- # Main # --------------------------------------------------------------- def main(): random.seed() g = WalkableGraph('ghana.txt') g.print() print() print("Let's do some walking") k = 1000 g.random_walk("CapeCoast", k, 0) # g.probs_after_k_steps(k) if __name__ == '__main__': main()
34.336066
116
0.467176
3,436
0.820243
0
0
0
0
0
0
1,792
0.427787
4bbef0cf8a34a2518357110d8f321604ae40180d
337
py
Python
51_60/day-60/forms/main.py
srakhe/100-days-py
4d99ab35eb1376d2f8722c42e0bf98acc18fba20
[ "MIT" ]
null
null
null
51_60/day-60/forms/main.py
srakhe/100-days-py
4d99ab35eb1376d2f8722c42e0bf98acc18fba20
[ "MIT" ]
null
null
null
51_60/day-60/forms/main.py
srakhe/100-days-py
4d99ab35eb1376d2f8722c42e0bf98acc18fba20
[ "MIT" ]
null
null
null
from flask import Flask, render_template, request app = Flask(__name__) @app.route('/') def home(): return render_template('index.html') @app.route('/login', methods=['POST']) def login(): return f'<h2>Name: {request.form["name"]}</h2><br><h2>Password: {request.form["password"]}' if __name__ == '__main__': app.run()
18.722222
95
0.652819
0
0
0
0
215
0.637982
0
0
123
0.364985
4bbf1d6eb8af61adb06a84718e97dce8dddb1ac0
6,094
py
Python
spartify/settings.py
cl1ckname/Spartify
3c45236e3f8803af9d01ac638e3d10a834ab7b7d
[ "Apache-2.0" ]
3
2021-07-26T15:43:20.000Z
2022-02-11T17:22:31.000Z
spartify/settings.py
cl1ckname/Spartify
3c45236e3f8803af9d01ac638e3d10a834ab7b7d
[ "Apache-2.0" ]
2
2021-07-08T14:25:22.000Z
2021-08-19T18:17:14.000Z
spartify/settings.py
cl1ckname/Spartify
3c45236e3f8803af9d01ac638e3d10a834ab7b7d
[ "Apache-2.0" ]
1
2021-08-19T18:17:48.000Z
2021-08-19T18:17:48.000Z
""" Django settings for spartify project. Generated by 'django-admin startproject' using Django 2.2.12. For more information on this file, see https://docs.djangoproject.com/en/2.2/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/2.2/ref/settings/ """ import os from dotenv import load_dotenv load_dotenv() # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/2.2/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = os.environ.get("SECRET_KEY") # SECURITY WARNING: don't run with debug turned on in production! DEBUG = int(os.environ.get("DEBUG", default=1)) ALLOWED_HOSTS = os.environ.get("DJANGO_ALLOWED_HOSTS",'').split(" ") + ['*', '192.168.43.72', '192.168.0.53', '0.0.0.0'] AUTHENTICATION_BACKENDS = ( 'django.contrib.auth.backends.ModelBackend', 'social_core.backends.spotify.SpotifyOAuth2', ) AUTH_USER_MODEL = 'backend.User' SOCIAL_AUTH_USER_MODEL = 'backend.User' # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'channels', 'social_django', 'backend', 'lobby' ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', # 'backend.middlewares.ApiMiddleware', ] ROOT_URLCONF = 'spartify.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'spartify.wsgi.application' ASGI_APPLICATION = 'spartify.routing.application' # Database # https://docs.djangoproject.com/en/2.2/ref/settings/#databases DATABASES = { 'default': { "ENGINE": os.environ.get("SQL_ENGINE", "django.db.backends.sqlite3"), "NAME": os.environ.get("SQL_DATABASE", os.path.join(BASE_DIR, "db.sqlite3")), "USER": os.environ.get("SQL_USER", "user"), "PASSWORD": os.environ.get("SQL_PASSWORD", "password"), "HOST": os.environ.get("SQL_HOST", "localhost"), "PORT": os.environ.get("SQL_PORT", "5432"), } } # Password validation # https://docs.djangoproject.com/en/2.2/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/2.2/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/2.2/howto/static-files/ BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) STATIC_URL = 'staticfiles/' STATIC_ROOT = os.path.join(BASE_DIR, STATIC_URL) SOCIAL_AUTH_SPOTIFY_KEY = os.environ['SOCIAL_AUTH_SPOTIFY_KEY'] SOCIAL_AUTH_SPOTIFY_SECRET = os.environ['SOCIAL_AUTH_SPOTIFY_SECRET'] SOCIAL_AUTH_URL_NAMESPACE = 'social' SOCIAL_AUTH_SPOTIFY_SCOPE = ['user-read-email','user-read-private', 'user-read-playback-state', 'user-modify-playback-state'] # SOCIAL_AUTH_LOGIN_REDIRECT_URL = 'http://{}/complete/spotify/' % os.getenv('HOST') LOGIN_REDIRECT_URL = 'dashboard' LOGIN_URL = 'login' DEFAULT_AUTO_FIELD='django.db.models.AutoField' SOCIAL_AUTH_PIPELINE = ( 'social_core.pipeline.social_auth.social_details', 'social_core.pipeline.social_auth.social_uid', 'social_core.pipeline.social_auth.auth_allowed', 'social_core.pipeline.social_auth.social_user', 'social_core.pipeline.user.get_username', 'social_core.pipeline.user.create_user', 'social_core.pipeline.social_auth.associate_user', 'social_core.pipeline.social_auth.load_extra_data', 'social_core.pipeline.user.user_details', 'backend.pipeline.save_access_token', #save token on login, ) QUEUE_SESSION_ID = 'queue' SESSION_EXPIRE_AT_BROWSER_CLOSE = 15 LOGGING = { 'version': 1, 'disable_existing_loggers': False, 'formatters': { 'api_formatter': { 'format': '{username} -- {endpoint} -- {status_code:d}: {message}', 'style': '{', }, 'lobby_formatter': { 'format': '{id}--{username}: {message} -- {asctime}', 'style': '{', }, }, 'handlers': { 'api_errors': { 'class': 'logging.FileHandler', 'filename': 'logs/api_errors.log', 'formatter': 'api_formatter', 'level': 'ERROR', }, }, 'loggers':{ 'backend': { 'handlers': ['api_errors'], }, }, } REDIS_HOST = os.environ.get("REDIS_HOST", '127.0.0.1') REDIS_PORT = 6379 REDIS_DB = 0 CHANNEL_LAYERS = { 'default': { 'BACKEND': 'channels_redis.core.RedisChannelLayer', "CONFIG": { "hosts": [(REDIS_HOST, REDIS_PORT)], }, } }
29.019048
125
0.675583
0
0
0
0
0
0
0
0
3,888
0.638005
4bc09df45d93aefe38be329327bcf363df1f3d3e
7,018
py
Python
bin/tdgwgeo2csv.py
Bauble/bauble.api
183c97fda076ea870e21e70ecf89a2a94a7f5722
[ "BSD-3-Clause" ]
null
null
null
bin/tdgwgeo2csv.py
Bauble/bauble.api
183c97fda076ea870e21e70ecf89a2a94a7f5722
[ "BSD-3-Clause" ]
1
2015-02-05T13:15:00.000Z
2015-02-05T13:15:00.000Z
bin/tdgwgeo2csv.py
Bauble/bauble.api
183c97fda076ea870e21e70ecf89a2a94a7f5722
[ "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python # # tdwggeo2csv.py # # Description: convert TDWG plant distribution files out of the box to a single # CSV file # # TODO: should create new id's for each entry and have a tdwg_code for # each so we can maintain as much data as possbible # TODO: we should probably include the original text files in bauble # and run the conversion script on build # TODO: add a notes column to geography so we carry over the extra # geography data(kew regions, notes, etc.) and so that we can add # notes to them in bauble import codecs import os import re from optparse import OptionParser # l1 - Continent, tblLevel1.txt, UTF-8 # l2 - Region, tblLevel2.txt, UTF-8 # l3 - BotanicalCountry, tblLevel4, ISO-8859-15 # l4 - BaseUnit, tblLevel4.txt, ISO-8859-15 # gazette (places), tblGazette.txt, ISO-8859-15 parser = OptionParser() parser.add_option('-d', '--directory', dest='directory', help='directory of WGS txt files', metavar='DIR') (options, args) = parser.parse_args() if not options.directory: parser.error('directory required') cwd, _dummy = os.path.split(__file__) src_dir = options.directory class Reader(object): def __init__(self, filename, encoding='utf8'): self.file = codecs.open(filename, "r", encoding) self.headers = self.file.next().strip().split('*') s = "" # sanitize the column headers for h in self.headers: h2 = h.replace(' ', '_') s += '(?P<%s>.*?)\*' % h2 s = s[:-2] + '$' self.line_rx = re.compile(s) def group(self, line): m = self.line_rx.match(line.strip()) if m is None: raise ValueError("could not match:\n%s\n%s" % (unicode(line), (unicode(s)))) return m.groupdict() def __iter__(self): return self def next(self): line = self.file.next() # remove the stupid ,00 decimals at the end of the integers #line = self.file.next().replace(',00','') return self.group(line) # converted rows organized by tdwg_code so we can resolve parents converted_rows = {} id_ctr = 1 class Row(dict): def __init__(self, id=None, name=None, tdwg_code=None, iso_code=None, parent_id=None): super(Row, self).__init__(id=id, name=name, tdwg_code=tdwg_code, iso_code=iso_code, parent_id=parent_id) columns = ['id', 'name', 'tdwg_code', 'iso_code', 'parent_id'] def __getattr__(self, item): if item in self: return self[item] else: return getattr(self, item) def __setattr__(self, key, value): self[key] = value def csv(self): s = [] for c in self.columns: if self[c] is None: #s.append('None') s.append('') elif c is 'id' or c is 'parent_id': s.append(self[c]) else: s.append('"%s"' % self[c].encode('utf8')) # s.append(quote(self[c])) return ','.join(s) def convert_level1(): global converted_data, id_ctr reader = Reader(os.path.join(src_dir, 'tblLevel1.txt'), 'utf8') for line in reader: r = Row(id=str(id_ctr), name=line['L1_continent'], tdwg_code=line['L1_code']) converted_rows[line['L1_code']] = r print(r.csv()) id_ctr += 1 def convert_level2(): global converted_data, id_ctr reader = Reader(os.path.join(src_dir, 'tblLevel2.txt'), 'utf8') for line in reader: r = Row(id=str(id_ctr), name=line['L2_region'], tdwg_code=line['L2_code'], iso_code=line['L2_ISOcode']) r.parent_id = converted_rows[line['L1_code']]['id'] converted_rows[line['L2_code']] = r print(r.csv()) id_ctr += 1 def convert_level3(): global converted_data, id_ctr reader = Reader(os.path.join(src_dir, 'tblLevel3.txt'), 'iso-8859-15') for line in reader: r = Row(id=str(id_ctr), name=line['L3_area'], tdwg_code=line['L3_code'], iso_code=line['L3_ISOcode']) #r.parent_id = converted_rows[line['L2_code']]['id'] r['parent_id'] = converted_rows[line['L2_code']]['id'] converted_rows[line['L3_code']] = r print(r.csv()) id_ctr += 1 def convert_level4(): global converted_data, id_ctr reader = Reader(os.path.join(src_dir, 'tblLevel4.txt'), 'iso-8859-15') for line in reader: # skip redundant lines from level 3 if line['L4_code'].endswith('-OO'): continue r = Row(id=str(id_ctr), name=line['L4_country'], tdwg_code=line['L4_code'], iso_code=line['L4_ISOcode']) r.parent_id = converted_rows[line['L3_code']]['id'] converted_rows[line['L4_code']] = r print(r.csv()) id_ctr += 1 def convert_gazetteer(): global converted_data, id_ctr reader = Reader(os.path.join(src_dir, 'tblGazetteer.txt'), 'iso-8859-15') for line in reader: # try to only include those things that are unique to the gazeteer if line['L3_code'] in converted_rows and \ converted_rows[line['L3_code']]['name'] == line['Gazetteer']: continue elif line['L4_code'] in converted_rows and \ converted_rows[line['L4_code']]['name'] == line['Gazetteer']: continue # TODO: create two rows, one for the gazetteer data and one for the # kew data r = Row(id=str(id_ctr), name=line['Gazetteer'], tdwg_code=line['ID']) # throw out anything that doesn't have a name, there seems # to be at least one row that doesn't have a name and is really just # a place holder for a kew region if line['Synonym'] != '': #print '%s == %s' % (line['Gazetteer'].encode('utf8'), line['Synonym'].encode('utf8')) pass if r.name == '' or line['Synonym'] != '': continue try: r.parent_id = converted_rows[line['L4_code']]['id'] except KeyError as e: try: r.parent_id = converted_rows[line['L3_code']]['id'] except KeyError as e: try: r.parent_id = converted_rows[line['L2_code']]['id'] except KeyError as e: try: r.parent_id = converted_rows[line['L1_code']]['id'] except KeyError as e: pass # add the converted rows and print out the csv line converted_rows[line['ID']] = r print(r.csv()) id_ctr += 1 def main(): global id_ctr, converted_rows print(','.join(['"%s"' % c for c in Row.columns])) convert_level1() convert_level2() convert_level3() convert_level4() convert_gazetteer() print(Row(id='%s' % id_ctr, name='Cultivated').csv()) id_ctr += 1 if __name__ == "__main__": main()
31.470852
98
0.578655
1,840
0.262183
0
0
0
0
0
0
2,269
0.323311
4bc28189f37d50450206554fd6ab1753bd171778
7,386
py
Python
getters/haproxy_haproxylogs.py
gunny26/datalogger
7bd29ab88f2e2749284d80a6a834c94c0955a7e0
[ "Apache-2.0" ]
null
null
null
getters/haproxy_haproxylogs.py
gunny26/datalogger
7bd29ab88f2e2749284d80a6a834c94c0955a7e0
[ "Apache-2.0" ]
null
null
null
getters/haproxy_haproxylogs.py
gunny26/datalogger
7bd29ab88f2e2749284d80a6a834c94c0955a7e0
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/python """ Program to read and parse haproxylogs to put them in shape to upload to DataLogger The input date schould be sorted by date, and finished the uploaded data will immediately split into TimeseriesArray, so no further data of this day could be appended """ import os import sys import gzip import logging logging.basicConfig(level=logging.DEBUG) import datetime import zlib import requests import StringIO import argparse # own modules from datalogger import DataLoggerWeb as DataLoggerWeb import tilak_haproxylog def aggregator(index_keynames, value_keynames, ts_keyname, func, interval = 60 * 5): """ aggregates some protocol data to get a consistent timeseries, with interval """ data = {} ts = None #print ts_keyname for parsts, parsdata in func(): #print parsdata #print parsdata["log_timestamp"] if ts is None: ts = parsts key = tuple((parsdata[key] for key in index_keynames)) values = tuple((int(parsdata[key]) for key in value_keynames)) if key not in data: data[key] = values else: data[key] = tuple((data[key][index] + int(values[index]) for index in range(len(values)))) if parsts > (ts + interval): for keys, values in data.items(): yield "%s\t%s\t%s" % (ts, "\t".join((str(index_key) for index_key in keys)), "\t".join((str(value_key) for value_key in values))) ts = None data = {} def parser_generator(index_keynames, value_keynames, file_obj): """ return specific parser for this set of parameters """ def inner(): """ split line into dict of fields, and append some data according to line """ for line in file_obj: logdata = tilak_haproxylog.parse_line(line) if logdata is not None: logdata["hits"] = 1 for value_key in value_keynames: if value_key not in logdata: logdata[value_key] = 0 status_code = int(logdata["status_code"]) if 100 <= status_code <= 199: logdata["rsp_1xx"] = 1 elif 200 <= status_code <= 299: logdata["rsp_2xx"] = 1 elif 300 <= status_code <= 399: logdata["rsp_3xx"] = 1 elif 400 <= status_code <= 499: logdata["rsp_4xx"] = 1 elif 500 <= status_code <= 599: logdata["rsp_5xx"] = 1 else: logdata["rsp_other"] = 1 ret_data = dict(zip(index_keynames, (logdata[index_key] for index_key in index_keynames))) ret_data.update(dict(zip(value_keynames, (logdata[value_key] for value_key in value_keynames)))) yield (logdata["ts"], ret_data) return inner def generate_datalogger_csv(logdir, datestring, keys, values, ts_keyname): """ create CSV like file with StringIO """ if datestring == datetime.date.today().isoformat(): logging.error("todays Logs are actually written and cannot used in datalogger") return headers = [ts_keyname, ] + list(keys) + list(values) linebuffer = [] linebuffer.append("\t".join(headers)) filename = os.path.join(logdir, "haproxylog_%s.gz" % datestring) logging.info("parsing file %s", filename) try: parser = parser_generator(keys, values, gzip.open(filename, "rb")) for line in aggregator(keys, values, ts_keyname, parser): linebuffer.append(line) except IOError as exc: logging.exception(exc) return StringIO.StringIO("\n".join(linebuffer)) def datestring_to_date(datestring): """ convert string in format YYYY-MM-DD into date object """ year, month, day = datestring.split("-") date = datetime.date(year=int(year), month=int(month), day=int(day)) return date def datewalk(datestring1, datestring2): """ count up from datestring1 to datestring2 in single day steps yield in isoformat() """ date1 = datestring_to_date(datestring1) date2 = datestring_to_date(datestring2) assert date2 > date1 oneday = datetime.timedelta(1) while date1 < date2: yield date1.isoformat() date1 += oneday def main(): """ what do you think, what main should do """ yesterday_datestring = (datetime.date.today() - datetime.timedelta(1)).isoformat() parser = argparse.ArgumentParser(description='generate TimeseriesArrays on local backend') parser.add_argument('--url', default="https://datalogger-api.tirol-kliniken.cc/DataLogger", help="url of DataLogger Webapplication") parser.add_argument('--logdir', default="/data1/haproxy_daily/", help="directory where to find day sorted haproxylogs") parser.add_argument("-b", '--back', help="how many days back from now") parser.add_argument("-s", '--startdate', help="start date in isoformat YYY-MM-DD") parser.add_argument("-e", '--enddate', default=yesterday_datestring, help="stop date in isoformat YYY-MM-DD") parser.add_argument("-q", '--quiet', action='store_true', help="set to loglevel ERROR") parser.add_argument("-v", '--verbose', action='store_true', help="set to loglevel DEBUG") args = parser.parse_args() if args.quiet is True: logging.getLogger("").setLevel(logging.ERROR) if args.verbose is True: logging.getLogger("").setLevel(logging.DEBUG) if (args.back is not None) == (args.startdate is not None): logging.error("option -b and -e are mutual exclusive, use only one") sys.exit(1) startdate = None if args.back is not None: startdate = (datetime.date.today() - datetime.timedelta(int(args.back))).isoformat() elif args.startdate is not None: startdate = args.startdate else: logging.error("you have to provide either -b or -s") sys.exit(1) # lets get started datalogger = DataLoggerWeb(args.url) project = "haproxy" tablename = "http_host" baseurl = "%s/upload_raw_file/" % args.url logdir = args.logdir # where to find haproxy logs keys = ("http_host", ) values = ("bytes_read", "rsp_1xx", "rsp_2xx", "rsp_3xx", "rsp_4xx", "rsp_5xx", "rsp_other", "srv_queue", "backend_queue", "actconn", "feconn", "beconn", "srv_conn", "retries", "tq", "tw", "tc", "tr", "tt", "hits") ts_keyname = "ts" for datestring in datewalk(startdate, args.enddate): caches = datalogger.get_caches(project, tablename, datestring) if caches["tsa"]["raw"] is not None: logging.info("Skipping this datestring, raw data is already available") continue try: stringio = generate_datalogger_csv(logdir, datestring, keys, values, ts_keyname) #upload data files = {'myfile': stringio} url = "/".join((baseurl, project, tablename, datestring)) logging.info("calling %s", url) response = requests.post(url, files=files) print response.content except StandardError as exc: logging.error("Exception on file datestring %si, skipping this date", datestring) except zlib.error as exc: logging.error(exc) if __name__ == "__main__": main()
40.80663
217
0.627268
0
0
2,766
0.374492
0
0
0
0
2,084
0.282155