commit
stringlengths
40
40
old_file
stringlengths
4
264
new_file
stringlengths
4
264
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
624
message
stringlengths
15
4.7k
lang
stringclasses
3 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
0c22d14992427aae0cafb8525cbd11b44761dfd7
pontoon/administration/management/commands/update_projects.py
pontoon/administration/management/commands/update_projects.py
import os import datetime from django.conf import settings from django.core.management.base import BaseCommand, CommandError from pontoon.administration.views import _update_from_repository from pontoon.base.models import Project class Command(BaseCommand): help = 'Update all projects from their repositories and store changes \ to the database' def handle(self, *args, **options): for project in Project.objects.all(): try: repository_type = project.repository_type repository_url = project.repository_url repository_path_master = os.path.join( settings.MEDIA_ROOT, repository_type, project.name) _update_from_repository( project, repository_type, repository_url, repository_path_master) now = datetime.datetime.now() self.stdout.write( '[%s]: Successfully updated project "%s"\n' % (now, project)) except Exception as e: now = datetime.datetime.now() raise CommandError( '[%s]: UpdateProjectsFromRepositoryError: %s\n' % (now, unicode(e)))
import os import datetime from django.conf import settings from django.core.management.base import BaseCommand, CommandError from pontoon.administration.views import _update_from_repository from pontoon.base.models import Project class Command(BaseCommand): help = 'Update all projects from their repositories and store changes \ to the database' def handle(self, *args, **options): for project in Project.objects.all(): try: repository_type = project.repository_type repository_url = project.repository_url repository_path_master = os.path.join( settings.MEDIA_ROOT, repository_type, project.slug) _update_from_repository( project, repository_type, repository_url, repository_path_master) now = datetime.datetime.now() self.stdout.write( '[%s]: Successfully updated project "%s"\n' % (now, project)) except Exception as e: now = datetime.datetime.now() raise CommandError( '[%s]: UpdateProjectsFromRepositoryError: %s\n' % (now, unicode(e)))
Use project slug as folder name instead of project name
Use project slug as folder name instead of project name
Python
bsd-3-clause
sudheesh001/pontoon,m8ttyB/pontoon,jotes/pontoon,vivekanand1101/pontoon,Osmose/pontoon,Jobava/mirror-pontoon,mastizada/pontoon,mastizada/pontoon,Jobava/mirror-pontoon,participedia/pontoon,vivekanand1101/pontoon,mathjazz/pontoon,mathjazz/pontoon,Jobava/mirror-pontoon,jotes/pontoon,m8ttyB/pontoon,sudheesh001/pontoon,participedia/pontoon,m8ttyB/pontoon,vivekanand1101/pontoon,yfdyh000/pontoon,yfdyh000/pontoon,yfdyh000/pontoon,mathjazz/pontoon,mastizada/pontoon,Osmose/pontoon,Jobava/mirror-pontoon,sudheesh001/pontoon,mathjazz/pontoon,Osmose/pontoon,participedia/pontoon,mozilla/pontoon,mozilla/pontoon,m8ttyB/pontoon,mozilla/pontoon,jotes/pontoon,yfdyh000/pontoon,mozilla/pontoon,Osmose/pontoon,participedia/pontoon,mastizada/pontoon,mozilla/pontoon,mathjazz/pontoon,vivekanand1101/pontoon,jotes/pontoon,sudheesh001/pontoon
08cbb4ebd44b5dca26d55a0e177c03930a2beb57
stopspam/forms/widgets.py
stopspam/forms/widgets.py
from django import forms from django.utils.translation import ugettext as _, get_language from django.utils.safestring import mark_safe # RECAPTCHA widgets class RecaptchaResponse(forms.Widget): def render(self, *args, **kwargs): from recaptcha.client import captcha as recaptcha recaptcha_options = "<script> var RecaptchaOptions = { theme: '" + self.theme + \ "', lang: '" + get_language()[0:2] + \ ("', custom_theme_widget: 'recaptcha_widget'" if self.theme == 'custom' else "'") + " }; </script>\n" return mark_safe(recaptcha_options + recaptcha.displayhtml(self.public_key)) class RecaptchaChallenge(forms.Widget): is_hidden = True def render(self, *args, **kwargs): return "" # return mark_safe('') # Honeypot widget -- most automated spam posters will check any checkbox # assuming it's an "I accept terms and conditions" box class HoneypotWidget(forms.CheckboxInput): is_hidden = True def render(self, *args, **kwargs): wrapper_html = '<div style="display:none"><label for="id_accept_terms">' + _('Are you a robot?') + '</label>%s</div>' return mark_safe(wrapper_html % super(HoneypotWidget, self).render(*args, **kwargs))
from django import forms from django.utils.translation import ugettext as _, get_language from django.utils.safestring import mark_safe # RECAPTCHA widgets class RecaptchaResponse(forms.Widget): is_hidden = True def render(self, *args, **kwargs): from recaptcha.client import captcha as recaptcha recaptcha_options = u"<script> var RecaptchaOptions = { theme: '" + self.theme + \ "', lang: '" + get_language()[0:2] + \ ("', custom_theme_widget: 'recaptcha_widget'" if self.theme == 'custom' else "'") + " }; </script>\n" return mark_safe(recaptcha_options + recaptcha.displayhtml(self.public_key)) class RecaptchaChallenge(forms.Widget): is_hidden = True def render(self, *args, **kwargs): return "" # return mark_safe('') # Honeypot widget -- most automated spam posters will check any checkbox # assuming it's an "I accept terms and conditions" box class HoneypotWidget(forms.CheckboxInput): is_hidden = True def render(self, *args, **kwargs): wrapper_html = '<div style="display:none"><label for="id_accept_terms">' + _('Are you a robot?') + '</label>%s</div>' return mark_safe(wrapper_html % super(HoneypotWidget, self).render(*args, **kwargs))
Fix skipping of recaptcha field widget HTML by marking it is_hidden
Fix skipping of recaptcha field widget HTML by marking it is_hidden
Python
bsd-3-clause
pombredanne/glamkit-stopspam
3d385898592b07249b478b37854d179d27a27bbb
OmniMarkupLib/Renderers/MarkdownRenderer.py
OmniMarkupLib/Renderers/MarkdownRenderer.py
from base_renderer import * import re import markdown @renderer class MarkdownRenderer(MarkupRenderer): FILENAME_PATTERN_RE = re.compile(r'\.(md|mkdn?|mdwn|mdown|markdown)$') def load_settings(self, renderer_options, global_setting): super(MarkdownRenderer, self).load_settings(renderer_options, global_setting) if 'extensions' in renderer_options: self.extensions = renderer_options['extensions'] else: # Fallback to the default GFM style self.extensions = ['tables', 'strikeout', 'fenced_code', 'codehilite'] if global_setting.mathjax_enabled: if 'mathjax' not in self.extensions: self.extensions.append('mathjax') @classmethod def is_enabled(cls, filename, syntax): if syntax == "text.html.markdown": return True return cls.FILENAME_PATTERN_RE.search(filename) is not None def render(self, text, **kwargs): return markdown.markdown(text, output_format='html5', extensions=self.extensions )
from base_renderer import * import re import markdown @renderer class MarkdownRenderer(MarkupRenderer): FILENAME_PATTERN_RE = re.compile(r'\.(md|mkdn?|mdwn|mdown|markdown|litcoffee)$') def load_settings(self, renderer_options, global_setting): super(MarkdownRenderer, self).load_settings(renderer_options, global_setting) if 'extensions' in renderer_options: self.extensions = renderer_options['extensions'] else: # Fallback to the default GFM style self.extensions = ['tables', 'strikeout', 'fenced_code', 'codehilite'] if global_setting.mathjax_enabled: if 'mathjax' not in self.extensions: self.extensions.append('mathjax') @classmethod def is_enabled(cls, filename, syntax): if syntax == "text.html.markdown": return True return cls.FILENAME_PATTERN_RE.search(filename) is not None def render(self, text, **kwargs): return markdown.markdown(text, output_format='html5', extensions=self.extensions )
Add litcoffee to Markdown extensions
Add litcoffee to Markdown extensions
Python
mit
timonwong/OmniMarkupPreviewer,Lyleo/OmniMarkupPreviewer,timonwong/OmniMarkupPreviewer,timonwong/OmniMarkupPreviewer,Lyleo/OmniMarkupPreviewer,timonwong/OmniMarkupPreviewer,Lyleo/OmniMarkupPreviewer,Lyleo/OmniMarkupPreviewer
2216caf836c1f2864103e8930f60713c226a8464
src/sql/parse.py
src/sql/parse.py
from ConfigParser import ConfigParser from sqlalchemy.engine.url import URL def parse(cell, config): parts = [part.strip() for part in cell.split(None, 1)] if not parts: return {'connection': '', 'sql': ''} if parts[0].startswith('[') and parts[0].endswith(']'): parser = ConfigParser() parser.read(config.dsn_filename) section = parts[0].lstrip('[').rstrip(']') connection = str(URL(drivername=parser.get(section, 'drivername'), username=parser.get(section, 'username'), password=parser.get(section, 'password'), host=parser.get(section, 'host'), database=parser.get(section, 'database'))) sql = parts[1] if len(parts) > 1 else '' elif '@' in parts[0] or '://' in parts[0]: connection = parts[0] if len(parts) > 1: sql = parts[1] else: sql = '' else: connection = '' sql = cell return {'connection': connection.strip(), 'sql': sql.strip() }
from ConfigParser import ConfigParser from sqlalchemy.engine.url import URL def parse(cell, config): parts = [part.strip() for part in cell.split(None, 1)] if not parts: return {'connection': '', 'sql': ''} if parts[0].startswith('[') and parts[0].endswith(']'): section = parts[0].lstrip('[').rstrip(']') parser = ConfigParser() parser.read(config.dsn_filename) cfg_dict = dict(parser.items(section)) connection = str(URL(**cfg_dict)) sql = parts[1] if len(parts) > 1 else '' elif '@' in parts[0] or '://' in parts[0]: connection = parts[0] if len(parts) > 1: sql = parts[1] else: sql = '' else: connection = '' sql = cell return {'connection': connection.strip(), 'sql': sql.strip()}
Allow DNS file to be less specific
Allow DNS file to be less specific
Python
mit
catherinedevlin/ipython-sql,catherinedevlin/ipython-sql
86bd0e7717596affceb1c40031855635b798e67b
benches/benchmark_rust.py
benches/benchmark_rust.py
import numpy as np from pypolyline.util import encode_coordinates # London bounding box N = 51.691874116909894 E = 0.3340155643740321 S = 51.28676016315085 W = -0.5103750689005356 num_coords = 100 coords = zip( np.random.uniform(S, N, [num_coords]), np.random.uniform(W, E, [num_coords]) ) if __name__ == "__main__": for x in xrange(50): encode_coordinates(coords, 5)
import numpy as np from pypolyline.cutil import encode_coordinates # London bounding box N = 51.691874116909894 E = 0.3340155643740321 S = 51.28676016315085 W = -0.5103750689005356 num_coords = 100 coords = zip( np.random.uniform(S, N, [num_coords]), np.random.uniform(W, E, [num_coords]) ) if __name__ == "__main__": for x in xrange(50): encode_coordinates(coords, 5)
Use Cython functions in benchmarks
Use Cython functions in benchmarks
Python
mit
urschrei/pypolyline,urschrei/pypolyline,urschrei/pypolyline
4522de348aab4cc99904b0bc210c223b2477b4b7
tests/config.py
tests/config.py
# our constants. import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) tbx_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm_model.tbx'))
# our constants. import os local_path = os.path.dirname(__file__) xml_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.xml')) csv_doc = os.path.abspath(os.path.join(local_path, 'data', 'fagatelebay_zone.csv')) bathy_raster = os.path.abspath(os.path.join(local_path, 'data', 'bathy5m_clip.tif')) pyt_file = os.path.abspath(os.path.join(local_path, '..', 'Install', 'toolbox', 'btm.pyt'))
Use pyt file instead of stand-alone tbx for testing.
Use pyt file instead of stand-alone tbx for testing.
Python
mpl-2.0
EsriOceans/btm
492827e2fc5244c313af4d25b563ad0f69425249
src/test.py
src/test.py
#!/usr/bin/env python from __future__ import division import numpy as np from matplotlib import pyplot as plt def main(): koeffs = [.3, 1.2, .1, 7] p = np.poly1d(koeffs) x = np.linspace(-5, 5, 100) y = p(x) + 2 * np.random.randn(100) - 1 # fit fit = np.polyfit(x, y, 3) p_fit = np.poly1d(fit) print "Real coefficients:", koeffs print "Fitted coefficients:", fit # plot plt.scatter(x, y) plt.plot(x, p_fit(x), 'orange', lw=2) plt.show() if __name__ == '__main__': main()
#!/usr/bin/env python from __future__ import division import numpy as np from matplotlib import pyplot as plt def main(): koeffs = [0.3, 1.2, 0.1, 7] p = np.poly1d(koeffs) x = np.linspace(-5, 5, 100) y = p(x) + 2 * np.random.randn(100) - 1 # fit fit = np.polyfit(x, y, 3) p_fit = np.poly1d(fit) print "Real coefficients:", koeffs print "Fitted coefficients:", fit # plot plt.scatter(x, y) plt.plot(x, p_fit(x), 'orange', lw=2) plt.show() if __name__ == '__main__': main()
Test change for closing an issue.
Test change for closing an issue. This is a place for a longer comment. Closes #1
Python
mit
bbci/playground
cc7253020251bc96d7d7f22a991b094a60bbc104
startServers.py
startServers.py
import sys import time import subprocess import psutil def startServer(command): if sys.platform.startswith('win'): return psutil.Popen(command, creationflags=subprocess.CREATE_NEW_CONSOLE) else: linuxCommand = 'xterm -hold -e "%s"' % command return psutil.Popen(linuxCommand, shell=True) def main(baseCommand, startingPort, count): servers = {} for i in range(1,count + 1): command = baseCommand + ' ' + str(startingPort + i) servers[i] = { 'command': command, 'process': startServer(command), } time.sleep(3) while True: for i, server in servers.iteritems(): if not server['process'].is_running(): servers[i]['process'] = startServer(servers[i]['command']) if __name__ == '__main__': print sys.argv main(sys.argv[1], int(sys.argv[2]), int(sys.argv[3]))
import sys import time import subprocess def main(baseCommand, startingPort, count): procs = [] for i in range(1,count + 1): command = baseCommand + ' ' + str(startingPort + i) if sys.platform.startswith('win'): process = subprocess.Popen(command, creationflags=subprocess.CREATE_NEW_CONSOLE) else: linuxCommand = 'xterm -hold -e "%s"' % command process = subprocess.Popen(linuxCommand, shell=True) procs.append(process) time.sleep(3) try: input('Enter to exit from Python script...') except: pass if __name__ == '__main__': print sys.argv main(sys.argv[1], int(sys.argv[2]), int(sys.argv[3]))
Revert "keep servers running for fun and profit"
Revert "keep servers running for fun and profit" This reverts commit c574ba41fb609db7a2c75340363fe1a1dcc31399.
Python
mit
IngenuityEngine/coren_proxy,IngenuityEngine/coren_proxy
6ac172843dc78ae6af87f00b260ef70f8965b3b7
start_server.py
start_server.py
#!/usr/bin/env python3 # tsuserver3, an Attorney Online server # # Copyright (C) 2016 argoneus <[email protected]> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from server.tsuserver import TsuServer3 # Idiotproof setup def check_pyyaml(): try: import yaml except ModuleNotFoundError: print("Couldn't import PyYAML. Installing it for you...") import pip pip.main(["install", "--user", "pyyaml"]) def main(): server = TsuServer3() server.start() if __name__ == '__main__': check_pyyaml() main()
#!/usr/bin/env python3 # tsuserver3, an Attorney Online server # # Copyright (C) 2016 argoneus <[email protected]> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # Idiotproof setup def check_pyyaml(): try: import yaml except ModuleNotFoundError: print("Couldn't import PyYAML. Installing it for you...") try: import pip pip.main(["install", "--user", "pyyaml"]) except ModuleNotFoundError: print("Couldn't install it for you, because you don't have pip.") def main(): from server.tsuserver import TsuServer3 server = TsuServer3() server.start() if __name__ == '__main__': check_pyyaml() main()
Handle case where pip is not found
Handle case where pip is not found
Python
agpl-3.0
Attorney-Online-Engineering-Task-Force/tsuserver3,Mariomagistr/tsuserver3
e54fa97cb44557454655efd24380da5223a1c5ae
tests/random_object_id/random_object_id_test.py
tests/random_object_id/random_object_id_test.py
import contextlib import re import sys import mock from six.moves import cStringIO from random_object_id.random_object_id import \ gen_random_object_id, parse_args, main @contextlib.contextmanager def captured_output(): new_out = StringIO() old_out = sys.stdout try: sys.stdout = new_out yield sys.stdout finally: sys.stdout = old_out def test_gen_random_object_id(): assert re.match('[0-9a-f]{24}', gen_random_object_id()) def test_gen_random_object_id_time(): with mock.patch('time.time') as mock_time: mock_time.return_value = 1429506585.786924 object_id = gen_random_object_id() assert re.match('55348a19', object_id) def test_parse_args(): assert parse_args(['-l']).long_form def test_main(): with mock.patch('sys.argv', ['random_object_id']): with captured_output() as output: main() assert re.match('[0-9a-f]{24}\n', output.getvalue()) def test_main_l(): with mock.patch('sys.argv', ['random_object_id', '-l']): with captured_output() as output: main() assert re.match('ObjectId\("[0-9a-f]{24}"\)\n', output.getvalue())
import contextlib import re import sys import mock import six from random_object_id.random_object_id import \ gen_random_object_id, parse_args, main @contextlib.contextmanager def captured_output(): old_out = sys.stdout try: sys.stdout = six.StringIO() yield sys.stdout finally: sys.stdout = old_out def test_gen_random_object_id(): assert re.match('[0-9a-f]{24}', gen_random_object_id()) def test_gen_random_object_id_time(): with mock.patch('time.time') as mock_time: mock_time.return_value = 1429506585.786924 object_id = gen_random_object_id() assert re.match('55348a19', object_id) def test_parse_args(): assert parse_args(['-l']).long_form def test_main(): with mock.patch('sys.argv', ['random_object_id']): with captured_output() as output: main() assert re.match('[0-9a-f]{24}\n', output.getvalue()) def test_main_l(): with mock.patch('sys.argv', ['random_object_id', '-l']): with captured_output() as output: main() assert re.match('ObjectId\("[0-9a-f]{24}"\)\n', output.getvalue())
Change how StringIO is imported
Change how StringIO is imported
Python
mit
mxr/random-object-id
eb34aadfcab01da9783688ffc72e23b0264713ad
spec/unit/hooks/for_caller.py
spec/unit/hooks/for_caller.py
import os import sys import json # # input (stdin) message = sys.stdin.read() message = json.loads(message) message["point"] = "receive" message["payload"]["price"] = "CHF 5.00" # # other inputs if len(sys.argv) > 1: message["argument"] = sys.argv[1] fcv = os.environ.get('ENV_VAR') if fcv: message["env_var"] = fcv # # output print json.dumps(message)
import os import sys import json # # input (stdin) message = sys.stdin.read() message = json.loads(message) message["point"] = "receive" message["payload"]["price"] = "CHF 5.00" # # other inputs if len(sys.argv) > 1: message["argument"] = sys.argv[1] fcv = os.environ.get('ENV_VAR') if fcv: message["env_var"] = fcv # # output print json.dumps(message)
Use if oneliners in Python caller sample
Use if oneliners in Python caller sample
Python
mit
dmicky0419/flor,floraison/flor,floraison/flor,dmicky0419/flor,flon-io/flor,dmicky0419/flor,floraison/flor
bb07ae592fbeb51a55b619a9880f4afc57bedee4
jwt_knox/urls.py
jwt_knox/urls.py
"""jwt_knox urls.py """ from django.conf.urls import url, include from django.contrib import admin from .views import DebugVerifyTokenView, LoginView, LogoutView, LogoutOtherView, LogoutAllView, VerifyView urlpatterns = [ url(r'^get_token$', LoginView.as_view()), url(r'^verify$', VerifyView.as_view()), url(r'^debug$', DebugVerifyTokenView.as_view()), url(r'^logout_other$', LogoutOtherView.as_view()), url(r'^logout_all$', LogoutAllView.as_view()), url(r'^logout$', LogoutView.as_view()), ]
"""jwt_knox urls.py """ from django.conf.urls import url, include from django.contrib import admin from .views import DebugVerifyTokenView, LoginView, LogoutView, LogoutOtherView, LogoutAllView, VerifyView app_name = 'jwt_knox' urlpatterns = [ url(r'^get_token$', LoginView.as_view(), name='get_new_token'), url(r'^verify$', VerifyView.as_view(), name='verify_token'), url(r'^debug$', DebugVerifyTokenView.as_view(), name='debug_token_auth_info'), url(r'^logout_other$', LogoutOtherView.as_view(), name='logout_other_tokens'), url(r'^logout_all$', LogoutAllView.as_view(), name='logout_all_user_tokens'), url(r'^logout$', LogoutView.as_view(), name='logout_current_token'), ]
Add names to URLs in JWT-Knox
Add names to URLs in JWT-Knox
Python
agpl-3.0
gpul-org/xea-core
80da397eb882622bc0bf1641bc4ee4e5813cf655
lopypi/pypi.py
lopypi/pypi.py
import re from urlparse import urlsplit from bs4 import BeautifulSoup import requests from urlparse import urldefrag, urljoin class PyPI(object): def __init__(self, index="http://pypi.python.org/simple"): self._index = index def list_packages(self): resp = requests.get(self._index) soup = BeautifulSoup(package_list) for link in soup.find_all("a"): yield link.text def list_files(self, package): package_uri = "%s/%s/" % (self._index, package) resp = requests.get(package_uri) soup = BeautifulSoup(resp.content) for link in soup.find_all("a"): # filter out links to home and download pages if "rel" in link.attrs: continue filename = link.text # filter out links to other web pages if urlsplit(filename).scheme: continue file_uri, frag = urldefrag(link.attrs['href']) file_uri = urljoin(package_uri, file_uri) mo = re.match(r"^md5=([a-fA-F0-9]{32})$", frag) md5 = mo.group(1) if mo else "" yield dict(filename=filename, remote_uri=file_uri, md5=md5)
import re from urlparse import urlsplit from bs4 import BeautifulSoup import requests from urlparse import urldefrag, urljoin class PyPI(object): def __init__(self, index="http://pypi.python.org/simple"): self._index = index def list_packages(self): resp = requests.get(self._index) soup = BeautifulSoup(resp.content) for link in soup.find_all("a"): yield link.text def list_files(self, package): package_uri = "%s/%s/" % (self._index, package) resp = requests.get(package_uri) soup = BeautifulSoup(resp.content) for link in soup.find_all("a"): # filter out links to home and download pages if "rel" in link.attrs: continue filename = link.text # filter out links to other web pages if urlsplit(filename).scheme: continue file_uri, frag = urldefrag(link.attrs['href']) file_uri = urljoin(package_uri, file_uri) mo = re.match(r"^md5=([a-fA-F0-9]{32})$", frag) md5 = mo.group(1) if mo else "" yield dict(filename=filename, remote_uri=file_uri, md5=md5)
Replace reference to previously factored out variable
Replace reference to previously factored out variable
Python
mit
bwhmather/LoPyPI,bwhmather/LoPyPI
4b6ae0eb113689515ba38e85c33a2ba40e58a163
src/minerva/storage/trend/engine.py
src/minerva/storage/trend/engine.py
from contextlib import closing from operator import contains from functools import partial from minerva.util import k, identity from minerva.directory import EntityType from minerva.storage import Engine from minerva.storage.trend import TableTrendStore class TrendEngine(Engine): @staticmethod def store_cmd(package, filter_package=k(identity)): """ Return a function to bind a data source to the store command. :param package: A DataPackageBase subclass instance :param filter_package: (TableTrendStore) -> (DataPackage) -> DataPackage :return: function that binds a data source to the store command :rtype: (data_source) -> (conn) -> None """ def bind_data_source(data_source): def execute(conn): entity_type_name = package.entity_type_name() with closing(conn.cursor()) as cursor: entity_type = EntityType.get_by_name(entity_type_name)( cursor ) trend_store = TableTrendStore.get( data_source, entity_type, package.granularity )(cursor) trend_store.store( filter_package(trend_store)(package) ).run(conn) return execute return bind_data_source def filter_existing_trends(trend_store): existing_trend_names = {trend.name for trend in trend_store.trends} def f(package): return package.filter_trends(partial(contains, existing_trend_names)) return f
from contextlib import closing from operator import contains from functools import partial from minerva.util import k, identity from minerva.directory import EntityType from minerva.storage import Engine from minerva.storage.trend import TableTrendStore class TrendEngine(Engine): @staticmethod def store_cmd(package, transform_package=k(identity)): """ Return a function to bind a data source to the store command. :param package: A DataPackageBase subclass instance :param transform_package: (TableTrendStore) -> (DataPackage) -> DataPackage :return: function that binds a data source to the store command :rtype: (data_source) -> (conn) -> None """ def bind_data_source(data_source): def execute(conn): entity_type_name = package.entity_type_name() with closing(conn.cursor()) as cursor: entity_type = EntityType.get_by_name(entity_type_name)( cursor ) trend_store = TableTrendStore.get( data_source, entity_type, package.granularity )(cursor) trend_store.store( transform_package(trend_store)(package) ).run(conn) return execute return bind_data_source def filter_existing_trends(trend_store): existing_trend_names = {trend.name for trend in trend_store.trends} def f(package): return package.filter_trends(partial(contains, existing_trend_names)) return f
Rename parameter filter_package to a more appropriate transform_package
Rename parameter filter_package to a more appropriate transform_package
Python
agpl-3.0
hendrikx-itc/minerva,hendrikx-itc/minerva
1db5ed3fa2fbb724c480bbf52c1d40c390dc857f
examples/example1.py
examples/example1.py
import fte.encoder regex = '^(a|b)+$' fixed_slice = 512 input_plaintext = 'test' fteObj = fte.encoder.RegexEncoder(regex, fixed_slice) ciphertext = fteObj.encode(input_plaintext) output_plaintext = fteObj.decode(ciphertext) print 'regex='+regex print 'fixed_slice='+str(fixed_slice) print 'input_plaintext='+input_plaintext print 'ciphertext='+ciphertext[:16]+'...'+ciphertext[-16:] print 'output_plaintext='+output_plaintext[0]
import regex2dfa import fte.encoder regex = '^(a|b)+$' fixed_slice = 512 input_plaintext = 'test' dfa = regex2dfa.regex2dfa(regex) fteObj = fte.encoder.DfaEncoder(dfa, fixed_slice) ciphertext = fteObj.encode(input_plaintext) [output_plaintext, remainder] = fteObj.decode(ciphertext) print 'input_plaintext='+input_plaintext print 'ciphertext='+ciphertext[:16]+'...'+ciphertext[-16:] print 'output_plaintext='+output_plaintext
Update example code to represent current FTE API and usage.
Update example code to represent current FTE API and usage.
Python
apache-2.0
kpdyer/libfte,kpdyer/libfte
f38b117316039042f3c00c73bbb7ceaeb0f2e6e1
src/python/pants/core_tasks/noop.py
src/python/pants/core_tasks/noop.py
# coding=utf-8 # Copyright 2014 Pants project contributors (see CONTRIBUTORS.md). # Licensed under the Apache License, Version 2.0 (see LICENSE). from __future__ import (absolute_import, division, generators, nested_scopes, print_function, unicode_literals, with_statement) from pants.task.noop_exec_task import NoopExecTask class NoopCompile(NoopExecTask): """A no-op that provides a product type that can be used to force scheduling.""" @classmethod def product_types(cls): return ['ran_compile'] class NoopTest(NoopExecTask): """A no-op that provides a product type that can be used to force scheduling.""" @classmethod def product_types(cls): return ['ran_tests']
# coding=utf-8 # Copyright 2014 Pants project contributors (see CONTRIBUTORS.md). # Licensed under the Apache License, Version 2.0 (see LICENSE). from __future__ import (absolute_import, division, generators, nested_scopes, print_function, unicode_literals, with_statement) from pants.task.noop_exec_task import NoopExecTask class NoopCompile(NoopExecTask): """A no-op that provides a product type that can be used to force scheduling. :API: public """ @classmethod def product_types(cls): return ['ran_compile'] class NoopTest(NoopExecTask): """A no-op that provides a product type that can be used to force scheduling. :API: public """ @classmethod def product_types(cls): return ['ran_tests']
Add public api markers for core_tasks
Add public api markers for core_tasks The following modules were reviewed and all api's were left as private. As far as I can tell these modules are not currently used by plugins. * pants.core_tasks.bash_completion.py * pants.core_tasks.changed_target_tasks.py * pants.core_tasks.clean.py * pants.core_tasks.deferred_sources_mapper.py * pants.core_tasks.explain_options_task.py * pants.core_tasks.invalidate.py * pants.core_tasks.list_goals.py * pants.core_tasks.pantsd_kill.py * pants.core_tasks.register.py * pants.core_tasks.reporting_server_kill.py * pants.core_tasks.reporting_server_run.py * pants.core_tasks.roots.py * pants.core_tasks.run_prep_command.py * pants.core_tasks.targets_help.py * pants.core_tasks.what_changed.py Testing Done: CI green: https://travis-ci.org/pantsbuild/pants/builds/111549646 Bugs closed: 2710, 2955 Reviewed at https://rbcommons.com/s/twitter/r/3490/
Python
apache-2.0
manasapte/pants,twitter/pants,fkorotkov/pants,jsirois/pants,pantsbuild/pants,peiyuwang/pants,pombredanne/pants,cevaris/pants,fkorotkov/pants,mateor/pants,baroquebobcat/pants,gmalmquist/pants,peiyuwang/pants,fkorotkov/pants,wisechengyi/pants,fkorotkov/pants,UnrememberMe/pants,wisechengyi/pants,ericzundel/pants,ericzundel/pants,landism/pants,ericzundel/pants,cevaris/pants,tdyas/pants,pombredanne/pants,jsirois/pants,gmalmquist/pants,kwlzn/pants,baroquebobcat/pants,gmalmquist/pants,15Dkatz/pants,manasapte/pants,peiyuwang/pants,peiyuwang/pants,twitter/pants,ericzundel/pants,pombredanne/pants,kwlzn/pants,baroquebobcat/pants,jsirois/pants,baroquebobcat/pants,twitter/pants,pombredanne/pants,foursquare/pants,baroquebobcat/pants,lahosken/pants,mateor/pants,tdyas/pants,lahosken/pants,UnrememberMe/pants,fkorotkov/pants,manasapte/pants,foursquare/pants,twitter/pants,lahosken/pants,benjyw/pants,lahosken/pants,wisechengyi/pants,ity/pants,pombredanne/pants,foursquare/pants,ity/pants,mateor/pants,landism/pants,fkorotkov/pants,gmalmquist/pants,ericzundel/pants,wisechengyi/pants,lahosken/pants,mateor/pants,UnrememberMe/pants,15Dkatz/pants,15Dkatz/pants,kwlzn/pants,foursquare/pants,dbentley/pants,dbentley/pants,wisechengyi/pants,pantsbuild/pants,tdyas/pants,benjyw/pants,lahosken/pants,wisechengyi/pants,landism/pants,cevaris/pants,pantsbuild/pants,gmalmquist/pants,UnrememberMe/pants,cevaris/pants,lahosken/pants,benjyw/pants,tdyas/pants,pombredanne/pants,ericzundel/pants,tdyas/pants,baroquebobcat/pants,dbentley/pants,UnrememberMe/pants,benjyw/pants,mateor/pants,UnrememberMe/pants,benjyw/pants,baroquebobcat/pants,ity/pants,foursquare/pants,pantsbuild/pants,manasapte/pants,landism/pants,kwlzn/pants,landism/pants,pantsbuild/pants,landism/pants,twitter/pants,ity/pants,wisechengyi/pants,15Dkatz/pants,peiyuwang/pants,kwlzn/pants,gmalmquist/pants,twitter/pants,cevaris/pants,foursquare/pants,dbentley/pants,twitter/pants,ity/pants,benjyw/pants,mateor/pants,foursquare/pants,manasapte/pants,pantsbuild/pants,cevaris/pants,dbentley/pants,cevaris/pants,fkorotkov/pants,lahosken/pants,mateor/pants,ity/pants,UnrememberMe/pants,15Dkatz/pants,ericzundel/pants,pantsbuild/pants,twitter/pants,UnrememberMe/pants,peiyuwang/pants,dbentley/pants,gmalmquist/pants,ericzundel/pants,landism/pants,peiyuwang/pants,manasapte/pants,tdyas/pants,foursquare/pants,wisechengyi/pants,tdyas/pants,dbentley/pants,landism/pants,benjyw/pants,manasapte/pants,mateor/pants,15Dkatz/pants,ity/pants,peiyuwang/pants,tdyas/pants,UnrememberMe/pants,wisechengyi/pants,baroquebobcat/pants,twitter/pants,fkorotkov/pants,kwlzn/pants,15Dkatz/pants,15Dkatz/pants,baroquebobcat/pants,foursquare/pants,kwlzn/pants,tdyas/pants,pombredanne/pants
6837986db77c9c9bd85392a74faebc019c1395a1
swen/flowexecutor.py
swen/flowexecutor.py
from . import flow class FlowExecutor: """ This class is responsible for flow execution """ def __init__(self, yaml_data): self.flow = flow.Flow(yaml_data) def execute(self): (exit_code, stdout, stderr) = None, None, None for step in self.flow.next_step(): if step.step is not None: (exit_code, stdout, stderr) = step.execute(exit_code=exit_code, stdout=stdout, stderr=stderr) # Terminate the flow if the step exit code is not success # and we don't have on_failure transition set explicitly on that set if exit_code != 0 and step.on_failure is None: break return (exit_code, stdout, stderr)
from . import flow import logging class FlowExecutor: """ This class is responsible for flow execution """ def __init__(self, yaml_data): self.flow = flow.Flow(yaml_data) def execute(self): (exit_code, stdout, stderr) = None, None, None for step in self.flow.next_step(): if step.step is not None: (exit_code, stdout, stderr) = step.execute(exit_code=exit_code, stdout=stdout, stderr=stderr) logging.debug("Executed step: {}, exit_code={!r}, stdout={!r}, stderr={!r}".format(step.id, exit_code, stdout, stderr)) # Terminate the flow if the step exit code is not success # and we don't have on_failure transition set explicitly on that step if exit_code != 0 and step.on_failure is None: break return (exit_code, stdout, stderr)
Add debug logging to flow executor
Add debug logging to flow executor
Python
mit
unix-beard/swen,unix-beard/swen,unix-beard/swen
a4184edab35890673b8b6a67e68a73e6ab7f0b89
tests/runtests.py
tests/runtests.py
#!/usr/bin/env python import os import sys from unittest import defaultTestLoader, TextTestRunner, TestSuite TESTS = ('form', 'fields', 'validators', 'widgets', 'webob_wrapper', 'translations', 'ext_csrf', 'ext_i18n') def make_suite(prefix='', extra=()): tests = TESTS + extra test_names = list(prefix + x for x in tests) suite = TestSuite() suite.addTest(defaultTestLoader.loadTestsFromNames(test_names)) return suite def additional_tests(): """ This is called automatically by setup.py test """ return make_suite('tests.') def main(): extra_tests = tuple(x for x in sys.argv[1:] if '-' not in x) suite = make_suite('', ) sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))) runner = TextTestRunner(verbosity=(sys.argv.count('-v') - sys.argv.count('-q') + 1)) result = runner.run(suite) sys.exit(not result.wasSuccessful()) if __name__ == '__main__': main()
#!/usr/bin/env python import os import sys from unittest import defaultTestLoader, TextTestRunner, TestSuite TESTS = ('form', 'fields', 'validators', 'widgets', 'webob_wrapper', 'translations', 'ext_csrf', 'ext_i18n') def make_suite(prefix='', extra=()): tests = TESTS + extra test_names = list(prefix + x for x in tests) suite = TestSuite() suite.addTest(defaultTestLoader.loadTestsFromNames(test_names)) return suite def additional_tests(): """ This is called automatically by setup.py test """ return make_suite('tests.') def main(): extra_tests = tuple(x for x in sys.argv[1:] if '-' not in x) suite = make_suite('', extra_tests) sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))) runner = TextTestRunner(verbosity=(sys.argv.count('-v') - sys.argv.count('-q') + 1)) result = runner.run(suite) sys.exit(not result.wasSuccessful()) if __name__ == '__main__': main()
Add back in running of extra tests
Add back in running of extra tests
Python
bsd-3-clause
jmagnusson/wtforms,cklein/wtforms,Xender/wtforms,pawl/wtforms,Aaron1992/wtforms,pawl/wtforms,subyraman/wtforms,wtforms/wtforms,skytreader/wtforms,hsum/wtforms,Aaron1992/wtforms,crast/wtforms
e452bee7b2babeec427a84e82ee3b4046f242bfc
process_urls.py
process_urls.py
#!/usr/bin/env python import os import sys import subprocess import logging logging.basicConfig(level=logging.DEBUG) log = logging.getLogger() with open(sys.argv[1], 'r') as handle: for line in handle: if line.startswith('#'): continue data = line.strip().split('\t') sha = data[0] url = data[1] if os.path.exists(sha): log.info("URL exists %s", url) else: log.info("URL missing, downloading %s to %s", url, sha) subprocess.check_call(['wget', url, '-O', sha]) with open(os.path.join('%s.sha256sum' % sha), 'w') as handle: handle.write("%s %s" % (sha, sha)) # Check sha256sum of download try: subprocess.check_call(['sha256sum', '-c', '%s.sha256sum' % sha]) except subprocess.CalledProcessError: log.error("File has bad hash! Refusing to serve this to end users.") os.unlink(sha)
#!/usr/bin/env python import os import sys import subprocess import logging logging.basicConfig(level=logging.DEBUG) log = logging.getLogger() with open(sys.argv[1], 'r') as handle: print """<!DOCTYPE html><html><head><title>Galaxy Package Cache</title></head><body><h1>About</h1><p>This package cache serves to preserve packages permanently. Please see our <a href="https://github/...">Github Repository</a> for more information.</p><h1>Cached URLs</h1><table><thead><tr><th>sha256sum</th><th>URL</th><th>Comment</th></tr></thead><tbody>""" retcode = 0 for line in handle: if line.startswith('#'): continue data = line.strip().split('\t') (sha, url) = data[0:2] comment = data[2] if len(data) > 2 else "" print """<tr><td>{sha}</td><td><a href="{sha}">Link</a></td><td>{comment}</td></tr>""".format(sha=sha, url=url, comment=comment) if os.path.exists(sha): log.info("URL exists %s", url) else: log.info("URL missing, downloading %s to %s", url, sha) subprocess.check_call(['wget', url, '-O', sha]) with open(os.path.join('%s.sha256sum' % sha), 'w') as handle: handle.write("%s %s" % (sha, sha)) # Check sha256sum of download try: subprocess.check_call(['sha256sum', '-c', '%s.sha256sum' % sha]) except subprocess.CalledProcessError: log.error("File has bad hash! Refusing to serve this to end users.") os.unlink(sha) print "</tbody></table></body></html>" sys.exit(retcode)
Update to have html output
Update to have html output
Python
mit
galaxyproject/cargo-port,galaxyproject/cargo-port,erasche/community-package-cache,erasche/community-package-cache,gregvonkuster/cargo-port,erasche/community-package-cache,gregvonkuster/cargo-port,gregvonkuster/cargo-port
52ddec80be8e2c90807a7b07425a6f260c9e86e0
src/zeit/retresco/tests/test_tag.py
src/zeit/retresco/tests/test_tag.py
# coding: utf8 import unittest class TagTest(unittest.TestCase): """Testing ..tag.Tag.""" def test_from_code_generates_a_tag_object_equal_to_its_source(self): from ..tag import Tag tag = Tag(u'Vipraschül', 'Person') self.assertEqual(tag, Tag.from_code(tag.code))
# coding: utf8 import zeit.cms.interfaces import zeit.retresco.testing class TagTest(zeit.retresco.testing.FunctionalTestCase): """Testing ..tag.Tag.""" def test_from_code_generates_a_tag_object_equal_to_its_source(self): from ..tag import Tag tag = Tag(u'Vipraschül', 'Person') self.assertEqual(tag, Tag.from_code(tag.code)) def test_uniqueId_from_tag_can_be_adapted_to_tag(self): from ..tag import Tag tag = Tag(u'Vipraschül', 'Person') self.assertEqual(tag, zeit.cms.interfaces.ICMSContent(tag.uniqueId))
Test that adapter in `zeit.cms` handles unicode escaped uniqueId correctly.
ZON-3199: Test that adapter in `zeit.cms` handles unicode escaped uniqueId correctly.
Python
bsd-3-clause
ZeitOnline/zeit.retresco
0b6e0e09abd007dad504693ca8cae4c7b0222765
gamernews/apps/threadedcomments/views.py
gamernews/apps/threadedcomments/views.py
from django.shortcuts import render_to_response, get_object_or_404 from django.template import RequestContext from django.contrib.auth.decorators import login_required from django.utils.translation import ugettext as _ from django.views.generic.list import ListView from core.models import Account as User from django_comments.models import Comment from news.models import Blob, BlobInstance from .models import ThreadedComment def single_comment(request, id): comment = get_object_or_404(ThreadedComment, id=id) variables = RequestContext(request, {'comment': comment}) return render_to_response('comments/single.html', variables) def comment_posted(request): if request.GET['c']: comment_id, blob_id = request.GET['c'] comment = Comment.objects.get( pk=comment_id ) blob = Blob.objects.get(pk=blob_id) if blob: return HttpResponseRedirect( blob.get_absolute_url() ) return HttpResponseRedirect( "/" )
from django.shortcuts import render_to_response, get_object_or_404 from django.template import RequestContext from django.contrib.auth.decorators import login_required from django.utils.translation import ugettext as _ from django.views.generic.list import ListView from core.models import Account as User from django_comments.models import Comment from news.models import Blob, BlobInstance from .models import ThreadedComment def single_comment(request, id): comment = get_object_or_404(ThreadedComment, id=id) variables = RequestContext(request, {'comment': comment}) return render_to_response('comments/single.html', variables) def comment_posted(request): if request.GET['c']: blob_id = request.GET['c'] blob = Blob.objects.get(pk=blob_id) if blob: return HttpResponseRedirect( blob.get_absolute_url() ) return HttpResponseRedirect( "/" )
Remove name, url and email from comment form
Remove name, url and email from comment form
Python
mit
underlost/GamerNews,underlost/GamerNews
716c0c4ab08266ce42f65afc0cd4bd8e0ed191e0
table_parser.py
table_parser.py
#!/usr/bin/python import sys import latex_table import table_to_file if __name__ == "__main__": # Parse arguments import argparse parser = argparse.ArgumentParser() parser.add_argument("input", help="the LaTeX input file to be parsed") # Add two mutually exclusive arguments: grouped/ungrouped parser_grouped = parser.add_mutually_exclusive_group() parser_grouped.add_argument("--grouped", help="group systematics", action="store_true") parser_grouped.add_argument("--ungrouped", help="do *not* group systematics", action="store_false") # Add optional arguments for file output parser.add_argument("--json", dest="json_file", help="output a JSON file") parser.add_argument("--tex", dest="tex_file", help="output a LaTeX file") args = parser.parse_args() if args.grouped: print "Grouping systematics is not yet implemented" sys.exit(1) table = latex_table.readFromLatex(args.input) rows = table.getRows() columns = table.getColumns() dict = table.getEntries() if args.json_file: table_to_file.storeJSON(table, args.json_file) if args.tex_file: print "Printing to TEX file not yet implemented" table_to_file.storeTEX(table, args.tex_file) sys.exit(1) for row in rows: for column in columns: print "%s %s %s" % (row, column, dict[row][column])
#!/usr/bin/python import sys import latex_table import table_to_file if __name__ == "__main__": # Parse arguments import argparse parser = argparse.ArgumentParser() parser.add_argument("input", help="the LaTeX input file to be parsed") # Add two mutually exclusive arguments: grouped/ungrouped parser_grouped = parser.add_mutually_exclusive_group() parser_grouped.add_argument("--grouped", help="group systematics", action="store_true") parser_grouped.add_argument("--ungrouped", help="do *not* group systematics", action="store_false") # Add optional arguments for file output parser.add_argument("--json", dest="json_file", help="output a JSON file") parser.add_argument("--tex", dest="tex_file", help="output a LaTeX file") args = parser.parse_args() if args.grouped: print "Grouping systematics is not yet implemented" sys.exit(1) table = latex_table.readFromLatex(args.input) rows = table.getRows() columns = table.getColumns() dict = table.getEntries() if args.json_file: table_to_file.storeJSON(table, args.json_file) if args.tex_file: table_to_file.storeTEX(table, args.tex_file) for row in rows: for column in columns: print "%s %s %s" % (row, column, dict[row][column])
Remove exit statement and error message for tex output
Remove exit statement and error message for tex output
Python
mit
knutzk/parse_latex_table
a0fa76a7aeb3dba3b358abeab95fc03a90a0e8b6
members/views.py
members/views.py
from django.shortcuts import render def homepage(request): return render(request, "index.html", {})
from django.shortcuts import render from django.http import HttpResponse from .models import User def homepage(request): return render(request, "index.html", {}) def search(request, name): members = User.objects.filter(first_name__icontains=name) or \ User.objects.filter(last_name__icontains=name) or \ User.objects.filter(username__icontains=name) json_data = [dict( id=member.id, full_name=' '.join([member.first_name, member.last_name])) for member in members] return HttpResponse(json_data, mimetype='application/json')
Add view for searching users and return json format
Add view for searching users and return json format
Python
mit
Hackfmi/Diaphanum,Hackfmi/Diaphanum
4e31496e1d9e0b2af2ce8aa4bb58baa86f352521
flake8_docstrings.py
flake8_docstrings.py
# -*- coding: utf-8 -*- """pep257 docstrings convention needs error code and class parser for be included as module into flake8 """ import io import pep8 import pep257 __version__ = '0.2.2' class pep257Checker(object): """flake8 needs a class to check python file.""" name = 'pep257' version = __version__ STDIN_NAMES = set(['stdin', '-', '(none)', None]) def __init__(self, tree, filename='(none)', builtins=None): self.tree = tree self.filename = filename self.source = self.load_source() self.checker = pep257.PEP257Checker() def run(self): """Use directly check() api from pep257.""" for error in self.checker.check_source(self.source, self.filename): # Ignore AllError, Environment error. if isinstance(error, pep257.Error): # NOTE(sigmavirus24): Fixes GitLab#3 message = '%s %s' % (error.code, error.short_desc) yield (error.line, 0, message, type(self)) def load_source(self): if self.filename in self.STDIN_NAMES: self.filename = 'stdin' self.source = pep8.stdin_get_value() else: with io.open(self.filename, encoding='utf-8') as fd: self.source = fd.read()
# -*- coding: utf-8 -*- """Implementation of pep257 integration with Flake8. pep257 docstrings convention needs error code and class parser for be included as module into flake8 """ import io import pep8 import pep257 __version__ = '0.2.2' class pep257Checker(object): """Flake8 needs a class to check python file.""" name = 'flake8-docstrings' version = __version__ + ', pep257: {0}'.format(pep257.__version__) STDIN_NAMES = set(['stdin', '-', '(none)', None]) def __init__(self, tree, filename='(none)', builtins=None): """Placeholder.""" self.tree = tree self.filename = filename self.source = self.load_source() self.checker = pep257.PEP257Checker() def run(self): """Use directly check() api from pep257.""" for error in self.checker.check_source(self.source, self.filename): # Ignore AllError, Environment error. if isinstance(error, pep257.Error): # NOTE(sigmavirus24): Fixes GitLab#3 message = '%s %s' % (error.code, error.short_desc) yield (error.line, 0, message, type(self)) def load_source(self): """Load the source for the specified file.""" if self.filename in self.STDIN_NAMES: self.filename = 'stdin' self.source = pep8.stdin_get_value() else: with io.open(self.filename, encoding='utf-8') as fd: self.source = fd.read()
Fix up a couple of minor issues
Fix up a couple of minor issues
Python
mit
PyCQA/flake8-docstrings
e5ef9ca9c089ce1da4ff363d0c5a5090785ae0c5
test_scraper.py
test_scraper.py
from scraper import search_CL from scraper import read_search_results def test_search_CL(): test_body, test_encoding = search_CL(minAsk=100) assert "<span class=\"desktop\">craigslist</span>" in test_body assert test_encoding == 'utf-8' def test_read_search_result(): test_body, test_encoding = read_search_results() assert "<span class=\"desktop\">craigslist</span>" in test_body assert test_encoding == 'utf-8'
from scraper import search_CL from scraper import read_search_results from scraper import parse_source from scraper import extract_listings import bs4 def test_search_CL(): test_body, test_encoding = search_CL(minAsk=100, maxAsk=100) assert "<span class=\"desktop\">craigslist</span>" in test_body assert test_encoding == 'utf-8' def test_read_search_result(): test_body, test_encoding = read_search_results() assert "<span class=\"desktop\">craigslist</span>" in test_body assert test_encoding == 'utf-8' def test_parse_source(): test_body, test_encoding = read_search_results() test_parse = parse_source(test_body, test_encoding) assert isinstance(test_parse, bs4.BeautifulSoup) def test_extract_listings(): test_body, test_encoding = read_search_results() test_parse = parse_source(test_body, test_encoding) for row in extract_listings(test_parse): print type(row) assert isinstance(row, bs4.element.Tag)
Add test for extract listings that asserts each listing is a bs4.element.Tag
Add test for extract listings that asserts each listing is a bs4.element.Tag
Python
mit
jefrailey/basic-scraper
aca158817c21b8baeeb64d7290d61c32a79124f9
tests/test_heat_demand.py
tests/test_heat_demand.py
""" Test the electricity demand SPDX-FileCopyrightText: Uwe Krien <[email protected]> SPDX-FileCopyrightText: Patrik Schönfeldt SPDX-License-Identifier: MIT """ import numpy as np from demandlib.examples import heat_demand_example def test_heat_example(): """Test the results of the heat example.""" ann_demands_per_type = {'efh': 25000, 'mfh': 80000, 'ghd': 140000} demands = heat_demand_example.heat_example( ann_demands_per_type=ann_demands_per_type, testmode=True).sum() for key in ann_demands_per_type: assert np.isclose(demands[key], ann_demands_per_type[key])
""" Test the electricity demand SPDX-FileCopyrightText: Uwe Krien <[email protected]> SPDX-FileCopyrightText: Patrik Schönfeldt SPDX-License-Identifier: MIT """ import numpy as np from demandlib.examples import heat_demand_example def test_heat_example(): """Test the results of the heat example.""" ann_demands_per_type = {'efh': 25000, 'mfh': 80000, 'ghd': 140000} demands = heat_demand_example.heat_example( ann_demands_per_type=ann_demands_per_type, testmode=True).sum() for key in ann_demands_per_type: assert np.isclose(demands[key], ann_demands_per_type[key], rtol=1e-04)
Increase tollerance for heat demand test
Increase tollerance for heat demand test
Python
mit
oemof/demandlib
101b4e5fb29195e500103230b3bbdae2369fde75
tests/test_mal_scraper.py
tests/test_mal_scraper.py
import mal_scraper def test_import_mal_scraper(): """Can we import mal_scraper""" assert mal_scraper assert mal_scraper.__version__.split('.') == ['0', '1', '0'] class TestAutomaticUserDicoveryIntegration(object): """Can we discover users as we download pages?""" pass # TODO
import mal_scraper def test_import_mal_scraper(): """Can we import mal_scraper""" assert mal_scraper assert mal_scraper.__version__.split('.') == ['0', '2', '0'] class TestAutomaticUserDicoveryIntegration(object): """Can we discover users as we download pages?""" pass # TODO
Fix failing tests (version number)
Fix failing tests (version number)
Python
mit
QasimK/mal-scraper
0060a32b58c7769ac97ac894cbaf6a2eaa1b389f
mmiisort/main.py
mmiisort/main.py
from isort import SortImports import mothermayi.colors import mothermayi.errors import mothermayi.files def plugin(): return { 'name' : 'isort', 'pre-commit' : pre_commit, } def do_sort(filename): results = SortImports(filename) return getattr(results, 'in_lines', None) and results.in_lines != results.out_lines def get_status(had_changes): return mothermayi.colors.red('unsorted') if had_changes else mothermayi.colors.green('sorted') def pre_commit(config, staged): python_files = list(mothermayi.files.python_source(staged)) if not python_files: return changes = [do_sort(filename) for filename in python_files] messages = [get_status(had_change) for had_change in changes] lines = [" {0:<30} ... {1:<10}".format(filename, message) for filename, message in zip(python_files, messages)] result = "\n".join(lines) if any(changes): raise mothermayi.errors.FailHook(result) return result
from isort import SortImports import mothermayi.colors import mothermayi.errors import mothermayi.files def plugin(): return { 'name' : 'isort', 'pre-commit' : pre_commit, } def do_sort(filename): results = SortImports(filename, check=True) return results.incorrectly_sorted def get_status(had_changes): return mothermayi.colors.red('unsorted') if had_changes else mothermayi.colors.green('sorted') def pre_commit(config, staged): python_files = list(mothermayi.files.python_source(staged)) if not python_files: return changes = [do_sort(filename) for filename in python_files] messages = [get_status(had_change) for had_change in changes] lines = [" {0:<30} ... {1:<10}".format(filename, message) for filename, message in zip(python_files, messages)] result = "\n".join(lines) if any(changes): raise mothermayi.errors.FailHook(result) return result
Leverage isort's check mode to make our logic simpler
Leverage isort's check mode to make our logic simpler This avoids having to check for in_lines or compare against the out_lines by just asking for a check and using the results
Python
mit
EliRibble/mothermayi-isort
ed76f648f60f96216377e4f12fea7043eaed904b
tests/helpers.py
tests/helpers.py
import virtualbox def list_machines(): vbox = virtualbox.vb_get_manager() for machine in vbox.getArray(vbox, "Machines"): print "Machine '%s' logs in '%s'" % ( machine.name, machine.logFolder )
import unittest import virtualbox class VirtualboxTestCase(unittest.TestCase): def setUp(self): self.vbox = virtualbox.vb_get_manager() def assertMachineExists(self, name, msg=None): try: self.vbox.findMachine(name) except Exception as e: if msg: self.fail(msg) else: self.fail(e.message) def assertMachineDoesNotExist(self, name, msg=None): self.assertRaisesRegexp(Exception, "Could not find a registered machine", self.vbox.findMachine, name) def list_machines(): vbox = virtualbox.vb_get_manager() for machine in vbox.getArray(vbox, "Machines"): print "Machine '%s' logs in '%s'" % ( machine.name, machine.logFolder )
Create a basic VirtualBoxTestCase with helper assertions
Create a basic VirtualBoxTestCase with helper assertions
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
990a3266739e5a4d763dd585f7cb722c0fe2b0f5
astroplpython/function/statistic/Maximum.py
astroplpython/function/statistic/Maximum.py
''' Created on Feb 6, 2015 @author: thomas ''' class Maximum (object): @staticmethod def calculate (measurement_list): import numpy as np ''' Find the maximum measurement value for any list of measured values. ''' x = [] for val in measurement_list: x.append(val.x) return measurement_list[np.argmax(x)] def __init__(self, ndarray): ''' Constructor '''
''' Created on Feb 6, 2015 @author: thomas ''' class Maximum (object): @staticmethod def calculate (measurement_list): import numpy as np ''' Find the maximum measurement value for any list of measured values. ''' x = [] for val in measurement_list: x.append(val.x) return measurement_list[np.argmax(x)]
Remove initializer..this is a 'static' class which
Remove initializer..this is a 'static' class which we are using functional approach with, e.g. no instances if we can help it..
Python
mit
brianthomas/astroplpython,brianthomas/astroplpython
554ef995f8c4ba42d00482480bf291bac2fd96e1
utils/database.py
utils/database.py
import json class Database(dict): """Holds a dict that contains all the information about the users in a channel""" def __init__(self, irc): super(Database, self).__init__(json.load(open("userdb.json"))) self.irc = irc def remove_entry(self, event, nick): try: del self[event.target][nick] except KeyError: for i in self[event.target].values(): if i['host'] == event.source.host: del self[event.target][i['hostmask'].split("!")[0]] break def add_entry(self, channel, nick, hostmask, account): temp = { 'hostmask': hostmask, 'host': hostmask.split("@")[1], 'account': account, 'seen': [__import__("time").time(), ""] } failed = False try: user = self[channel][nick] except KeyError: failed = True self[channel][nick] = temp if not failed: del temp['seen'] user.update(temp) def get_user_host(self, channel, nick): try: host = "*!*@" + self[channel][nick]['host'] except KeyError: self.irc.send("WHO {0} nuhs%nhuac".format(channel)) host = "*!*@" + self[channel][nick]['host'] return host def flush(self): with open('userdb.json', 'w') as f: json.dump(self, f, indent=2, separators=(',', ': ')) f.write("\n")
import json class Database(dict): """Holds a dict that contains all the information about the users in a channel""" def __init__(self, irc): super(Database, self).__init__(json.load(open("userdb.json"))) self.irc = irc def remove_entry(self, event, nick): try: del self[event.target][nick] except KeyError: for i in self[event.target].values(): if i['host'] == event.source.host: del self[event.target][i['hostmask'].split("!")[0]] break def add_entry(self, channel, nick, hostmask, account): temp = { 'hostmask': hostmask, 'host': hostmask.split("@")[1], 'account': account, 'seen': [__import__("time").time(), ""] } if nick in self[channel]: del temp['seen'] self[channel][nick].update(temp) else: self[channel][nick] = temp def get_user_host(self, channel, nick): try: host = "*!*@" + self[channel][nick]['host'] except KeyError: self.irc.send("WHO {0} nuhs%nhuac".format(channel)) host = "*!*@" + self[channel][nick]['host'] return host def flush(self): with open('userdb.json', 'w') as f: json.dump(self, f, indent=2, separators=(',', ': ')) f.write("\n")
Reduce code to a simpler form that checks if a user is already in the DB
Reduce code to a simpler form that checks if a user is already in the DB
Python
mit
wolfy1339/Python-IRC-Bot
12f3cc403f6ba0be957d1fb18253fb7529009764
moss/plotting.py
moss/plotting.py
import matplotlib.pyplot as plt def grid_axes_labels(f, xlabel=None, ylabel=None, **kws): axes = f.axes plt.setp(axes.flat, xlabel="", ylabel="") if xlabel is not None: for ax in axes[-1]: ax.set_xlabel(xlabel, **kws) if ylabel is not None: for ax in axes[0]: ax.set_ylabel(ylabel, **kws)
import matplotlib.pyplot as plt def grid_axes_labels(axes, xlabel=None, ylabel=None, **kws): plt.setp(axes.flat, xlabel="", ylabel="") if xlabel is not None: for ax in axes[-1]: ax.set_xlabel(xlabel, **kws) if ylabel is not None: for ax in axes[0]: ax.set_ylabel(ylabel, **kws)
Use matrix of axes not figure
Use matrix of axes not figure
Python
bsd-3-clause
mwaskom/moss,mwaskom/moss
acdbb1a9ca73b43b2a56b9372ded6859f5945721
bpython/test/test_autocomplete.py
bpython/test/test_autocomplete.py
from bpython import autocomplete import unittest try: from unittest import skip except ImportError: def skip(f): return lambda self: None #TODO: Parts of autocompletion to test: # Test that the right matches come back from find_matches (test that priority is correct) # Test the various complete methods (import, filename) to see if right matches # Test that MatchesIterator.substitute correctly subs given a match and a completer class TestSafeEval(unittest.TestCase): def test_catches_syntax_error(self): with self.assertRaises(autocomplete.EvaluationError): autocomplete.safe_eval('1re',{}) class TestFormatters(unittest.TestCase): def test_filename(self): last_part_of_filename = autocomplete.FilenameCompletion.format self.assertEqual(last_part_of_filename('abc'), 'abc') self.assertEqual(last_part_of_filename('abc/'), 'abc/') self.assertEqual(last_part_of_filename('abc/efg'), 'efg') self.assertEqual(last_part_of_filename('abc/efg/'), 'efg/') self.assertEqual(last_part_of_filename('/abc'), 'abc') self.assertEqual(last_part_of_filename('ab.c/e.f.g/'), 'e.f.g/') def test_attribute(self): self.assertEqual(autocomplete.after_last_dot('abc.edf'), 'edf')
from bpython import autocomplete import unittest try: from unittest import skip except ImportError: def skip(f): return lambda self: None #TODO: Parts of autocompletion to test: # Test that the right matches come back from find_matches (test that priority is correct) # Test the various complete methods (import, filename) to see if right matches # Test that MatchesIterator.substitute correctly subs given a match and a completer class TestSafeEval(unittest.TestCase): def test_catches_syntax_error(self): self.assertRaises(autocomplete.EvaluationError, autocomplete.safe_eval, '1re', {}) class TestFormatters(unittest.TestCase): def test_filename(self): last_part_of_filename = autocomplete.FilenameCompletion.format self.assertEqual(last_part_of_filename('abc'), 'abc') self.assertEqual(last_part_of_filename('abc/'), 'abc/') self.assertEqual(last_part_of_filename('abc/efg'), 'efg') self.assertEqual(last_part_of_filename('abc/efg/'), 'efg/') self.assertEqual(last_part_of_filename('/abc'), 'abc') self.assertEqual(last_part_of_filename('ab.c/e.f.g/'), 'e.f.g/') def test_attribute(self): self.assertEqual(autocomplete.after_last_dot('abc.edf'), 'edf')
Make test work under Python 2.6.
Make test work under Python 2.6.
Python
mit
wevial/bpython,aktorion/bpython,wevial/bpython,kdart/bpython,aktorion/bpython,kdart/bpython
ba2938267ac6198242e101d091339152767df557
calexicon/fn/tests/test_julian.py
calexicon/fn/tests/test_julian.py
import unittest from calexicon.calendars.tests.test_calendar import JulianGregorianConversion from calexicon.fn import julian_to_gregorian, gregorian_to_julian from calexicon.fn import julian_to_julian_day_number, julian_day_number_to_julian class TestJulianConversion(JulianGregorianConversion): def setUp(self): pass def Julian_to_Gregorian_conversion(self, julian_args, gregorian_args): result = julian_to_gregorian(*julian_args) self.assertEqual( result, gregorian_args, ) def Gregorian_to_Julian_conversion(self, julian_args, gregorian_args): result = gregorian_to_julian(*gregorian_args) self.assertEqual( result, julian_args, ) class TestJulianNumberConversion(unittest.TestCase): def test_number_to_julian_date(self): self.assertEqual(julian_to_julian_day_number(-4713, 1, 1), 0) self.assertEqual(julian_to_julian_day_number(-4712, 1, 1), 365) def test_julian_date_to_number(self): self.assertEqual(julian_day_number_to_julian(1), (-4713, 1, 1))
import unittest from calexicon.calendars.tests.test_calendar import JulianGregorianConversion from calexicon.fn import julian_to_gregorian, gregorian_to_julian from calexicon.fn import julian_to_julian_day_number, julian_day_number_to_julian class TestJulianConversion(JulianGregorianConversion): def setUp(self): pass def Julian_to_Gregorian_conversion(self, julian_args, gregorian_args): result = julian_to_gregorian(*julian_args) self.assertEqual( result, gregorian_args, ) def Gregorian_to_Julian_conversion(self, julian_args, gregorian_args): result = gregorian_to_julian(*gregorian_args) self.assertEqual( result, julian_args, ) class TestJulianNumberConversion(unittest.TestCase): def test_number_to_julian_date(self): self.assertEqual(julian_to_julian_day_number(-4713, 1, 1), 0) self.assertEqual(julian_to_julian_day_number(-4712, 1, 1), 365) def test_julian_date_to_number(self): self.assertEqual(julian_day_number_to_julian(0), (-4713, 1, 1))
Correct test to match the inverse test.
Correct test to match the inverse test.
Python
apache-2.0
jwg4/calexicon,jwg4/qual
02e03748e66ebf516a4a9b24f52563362e6bb895
command_line/scale_down_images.py
command_line/scale_down_images.py
from __future__ import division def nproc(): from libtbx.introspection import number_of_processors return number_of_processors(return_value_if_unknown=-1) def joiner(args): from dials.util.scale_down_image import scale_down_image scale_down_image(*args) def scale_down_images(in_template, out_template, start, end, scale_factor): from multiprocessing import Pool jobs = [(in_template % j, out_template % j, scale_factor) for j in range(start, end + 1)] pool = Pool(processes=nproc()) result = pool.map_async(joiner, jobs) result.get() return result if __name__ == '__main__': import sys if len(sys.argv) != 6: raise RuntimeError, '%s in_\%04d.cbf out_\%04d.cbf start end scale' % \ sys.argv[0] in_template = sys.argv[1] out_template = sys.argv[2] start = int(sys.argv[3]) end = int(sys.argv[4]) scale_factor = float(sys.argv[5]) scale_down_images(in_template, out_template, start, end, scale_factor)
from __future__ import division def nproc(): from libtbx.introspection import number_of_processors return number_of_processors(return_value_if_unknown=-1) def joiner(args): from dials.util.scale_down_image import scale_down_image scale_down_image(*args) print args[1] def scale_down_images(in_template, out_template, start, end, scale_factor): from multiprocessing import Pool jobs = [(in_template % j, out_template % j, scale_factor) for j in range(start, end + 1)] pool = Pool(processes=nproc()) result = pool.map_async(joiner, jobs) result.get() return result if __name__ == '__main__': import sys if len(sys.argv) != 6: raise RuntimeError, '%s in_\%04d.cbf out_\%04d.cbf start end scale' % \ sys.argv[0] in_template = sys.argv[1] out_template = sys.argv[2] start = int(sys.argv[3]) end = int(sys.argv[4]) scale_factor = float(sys.argv[5]) scale_down_images(in_template, out_template, start, end, scale_factor)
Print out file name after writing
Print out file name after writing
Python
bsd-3-clause
dials/dials,dials/dials,dials/dials,dials/dials,dials/dials
49ce9aa1bdd3479c31b8aa2e606b1768a444aea2
irrigator_pro/farms/templatetags/today_filters.py
irrigator_pro/farms/templatetags/today_filters.py
from django import template from datetime import date, datetime, timedelta register = template.Library() @register.filter(expects_localtime=True) def is_today(value): if isinstance(value, datetime): value = value.date() return value == date.today() @register.filter(expects_localtime=True) def is_past(value): if isinstance(value, datetime): value = value.date() return value < date.today() @register.filter(expects_localtime=True) def is_future(value): if isinstance(value, datetime): value = value.date() return value > date.today() @register.filter(expects_localtime=True) def compare_today(value): if isinstance(value, datetime): value = value.date() return value - date.today()
from django import template from datetime import date, datetime, timedelta register = template.Library() @register.filter(expects_localtime=True) def is_today(value): if isinstance(value, datetime): value = value.date() return value == date.today() @register.filter(expects_localtime=True) def is_past(value): if isinstance(value, datetime): value = value.date() return value < date.today() @register.filter(expects_localtime=True) def is_future(value): if isinstance(value, datetime): value = value.date() return value > date.today() @register.filter(expects_localtime=True) def compare_today(value): if isinstance(value, datetime): value = value.date() return value - date.today() @register.filter(expects_locattime=True) def today_in_season(season): start_date = season.season_start_date end_date = season.season_end_date return (start_date <= date.today() <= end_date)
Add new filter to determine if today is within the time period for a season.
Add new filter to determine if today is within the time period for a season.
Python
mit
warnes/irrigatorpro,warnes/irrigatorpro,warnes/irrigatorpro,warnes/irrigatorpro
0e0b96d0d800716102204cfdca7317ccb92cee95
pytextql/util.py
pytextql/util.py
# -*- coding: utf-8 -*- import csv import itertools def grouper(iterable, n): """ Slice up `iterable` into iterables of `n` items. :param iterable: Iterable to splice. :param n: Number of items per slice. :returns: iterable of iterables """ it = iter(iterable) while True: chunk = itertools.islice(it, n) try: first = next(chunk) except StopIteration: return yield itertools.chain([first], chunk) class UnicodeCSVReader(object): """ An extremely minimal wrapper around csv.reader to assist in reading Unicode data. """ def __init__(self, *args, **kwargs): self.encoding = kwargs.pop('encoding', 'utf8') self.pad_to = kwargs.pop('pad_to', 0) self.pad_with = kwargs.pop('pad_with', '') self.reader = csv.reader(*args, **kwargs) def next(self): row = self.reader.next() padding = [self.pad_with] * (self.pad_to - len(row)) return [unicode(c, self.encoding) for c in row] + padding def __iter__(self): return self @property def dialect(self): return self.reader.dialect @property def line_num(self): return self.reader.line_num
# -*- coding: utf-8 -*- import csv import itertools def grouper(iterable, n): """ Slice up `iterable` into iterables of `n` items. :param iterable: Iterable to splice. :param n: Number of items per slice. :returns: iterable of iterables """ it = iter(iterable) while True: chunk = itertools.islice(it, n) try: first = next(chunk) except StopIteration: return yield itertools.chain([first], chunk) class UnicodeCSVReader(object): """ An extremely minimal wrapper around csv.reader to assist in reading Unicode data. """ def __init__(self, *args, **kwargs): self.encoding = kwargs.pop('encoding', 'utf8') self.pad_to = kwargs.pop('pad_to', 0) self.pad_with = kwargs.pop('pad_with', '') self.reader = csv.reader(*args, **kwargs) def next(self): row = self.reader.next() padding = [self.pad_with] * (self.pad_to - len(row)) return [unicode(c, self.encoding) for c in row] + padding def __iter__(self): return self @property def dialect(self): return self.reader.dialect @property def line_num(self): return self.reader.line_num class UnicodeCSVWriter(object): def __init__(self, *args, **kwargs): self.encoding = kwargs.pop('encoding', 'utf8') self.writer = csv.writer(*args, **kwargs) def writerow(self, row): self.writer.writerow([ column.encode(self.encoding) for column in row ]) def writerows(self, rows): for row in rows: self.writerow(row)
Add a simple UnicodeCSVWriter, probably flawed.
Add a simple UnicodeCSVWriter, probably flawed.
Python
mit
TkTech/pytextql
c67acb72d5ddea8a1e4fb8a12aa3a6913629e0cb
Lib/setup.py
Lib/setup.py
def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('scipy',parent_package,top_path) #config.add_subpackage('cluster') #config.add_subpackage('fftpack') #config.add_subpackage('integrate') #config.add_subpackage('interpolate') #config.add_subpackage('io') config.add_subpackage('lib') config.add_subpackage('linalg') #config.add_subpackage('linsolve') #config.add_subpackage('maxentropy') config.add_subpackage('misc') #config.add_subpackage('montecarlo') config.add_subpackage('optimize') #config.add_subpackage('sandbox') #config.add_subpackage('signal') #config.add_subpackage('sparse') config.add_subpackage('special') config.add_subpackage('stats') #config.add_subpackage('ndimage') #config.add_subpackage('weave') config.make_svn_version_py() # installs __svn_version__.py config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration(top_path='').todict())
def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('scipy',parent_package,top_path) config.add_subpackage('cluster') config.add_subpackage('fftpack') config.add_subpackage('integrate') config.add_subpackage('interpolate') config.add_subpackage('io') config.add_subpackage('lib') config.add_subpackage('linalg') config.add_subpackage('linsolve') config.add_subpackage('maxentropy') config.add_subpackage('misc') #config.add_subpackage('montecarlo') config.add_subpackage('optimize') config.add_subpackage('sandbox') config.add_subpackage('signal') config.add_subpackage('sparse') config.add_subpackage('special') config.add_subpackage('stats') config.add_subpackage('ndimage') config.add_subpackage('weave') config.make_svn_version_py() # installs __svn_version__.py config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration(top_path='').todict())
Fix problem with __all__ variable and update weave docs a bit. Update compiler_cxx too.
Fix problem with __all__ variable and update weave docs a bit. Update compiler_cxx too. git-svn-id: 003f22d385e25de9cff933a5ea4efd77cb5e7b28@2022 d6536bca-fef9-0310-8506-e4c0a848fbcf
Python
bsd-3-clause
lesserwhirls/scipy-cwt,scipy/scipy-svn,lesserwhirls/scipy-cwt,jasonmccampbell/scipy-refactor,scipy/scipy-svn,scipy/scipy-svn,jasonmccampbell/scipy-refactor,lesserwhirls/scipy-cwt,jasonmccampbell/scipy-refactor,scipy/scipy-svn,lesserwhirls/scipy-cwt,jasonmccampbell/scipy-refactor
1441654c46e08b7286999b6887e59c56fa238ff7
python/piling-up.py
python/piling-up.py
from collections import deque def isVerticallyStackable(pile): vertical_stack = [] while pile: largest_cube, cube_sizes = remove_largest_cube_from_pile(pile) if vertical_stack == []: vertical_stack.append(largest_cube) else: top_of_stack = vertical_stack[-1] if(top_of_stack < largest_cube): return False vertical_stack.append(largest_cube) return True def remove_largest_cube_from_pile(cube_sizes): if(cube_sizes == []): return (None, cube_sizes) elif(cube_sizes[0] > cube_sizes[-1]): largest_cube = cube_sizes.popleft() return (largest_cube, cube_sizes) else: largest_cube = cube_sizes.pop() return (largest_cube, cube_sizes) num_test_cases = int(input()) for i in range(num_test_cases): num_cubes = int(input()) pile = deque(map(int, input().strip().split(" "))) if(isVerticallyStackable(pile)): print("Yes") else: print("No")
from collections import deque def isVerticallyStackable(pile): vertical_stack = [] while pile: largest_cube = remove_largest_cube_from_pile(pile) if vertical_stack == []: vertical_stack.append(largest_cube) else: top_of_stack = vertical_stack[-1] if(top_of_stack < largest_cube): return False vertical_stack.append(largest_cube) return True def remove_largest_cube_from_pile(cube_sizes): if(cube_sizes == []): return None elif(cube_sizes[0] > cube_sizes[-1]): return cube_sizes.popleft() else: return cube_sizes.pop() num_test_cases = int(input()) for i in range(num_test_cases): num_cubes = int(input()) pile = deque(map(int, input().strip().split(" "))) if(isVerticallyStackable(pile)): print("Yes") else: print("No")
Remove returned pile b/c mutating directly
Remove returned pile b/c mutating directly
Python
mit
rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank,rootulp/hackerrank
bb229be50e37bb710c32541cec7b159da9508335
tests/functional/subcommands/test_subcommands.py
tests/functional/subcommands/test_subcommands.py
import subprocess def test_subcommand(): """ Test that a command from the example project is registered. """ output = subprocess.check_output(['textx'], stderr=subprocess.STDOUT) assert b'testcommand' in output def test_subcommand_group(): """ Test that a command group is registered. """ output = subprocess.check_output(['textx', 'testgroup'], stderr=subprocess.STDOUT) assert b'groupcommand1' in output assert b'groupcommand2' in output
import sys import pytest import subprocess if (3, 6) <= sys.version_info < (3, 8): pytest.skip("Temporary workaround for Travis problems", allow_module_level=True) def test_subcommand(): """ Test that a command from the example project is registered. """ output = subprocess.check_output(['textx'], stderr=subprocess.STDOUT) assert b'testcommand' in output def test_subcommand_group(): """ Test that a command group is registered. """ output = subprocess.check_output(['textx', 'testgroup'], stderr=subprocess.STDOUT) assert b'groupcommand1' in output assert b'groupcommand2' in output
Add workaround for Travis CI problems
Add workaround for Travis CI problems
Python
mit
igordejanovic/textX,igordejanovic/textX,igordejanovic/textX
82f7a48695bc1dd97f9ab2697548f15b124dc82a
pyoracc/atf/atffile.py
pyoracc/atf/atffile.py
from .atflex import AtfLexer from .atfyacc import AtfParser from mako.template import Template class AtfFile(object): template = Template("${text.serialize()}") def __init__(self, content): self.content = content if content[-1] != '\n': content += "\n" lexer = AtfLexer().lexer parser = AtfParser().parser self.text = parser.parse(content, lexer=lexer) def __str__(self): return AtfFile.template.render_unicode(**vars(self)) def serialize(self): return AtfFile.template.render_unicode(**vars(self)) def _debug_lex_and_yac_file(file): import codecs text = codecs.open(file, encoding='utf-8-sig').read() from pyoracc.atf.atffile import AtfLexer lexer = AtfLexer().lexer lexer.input(text) for tok in lexer: print(tok) print("Lexed file") exer = AtfLexer().lexer parser = AtfParser().parser parser.parse(text, lexer=lexer) print("Parsed file")
from .atflex import AtfLexer from .atfyacc import AtfParser from mako.template import Template class AtfFile(object): template = Template("${text.serialize()}") def __init__(self, content): self.content = content if content[-1] != '\n': content += "\n" lexer = AtfLexer().lexer parser = AtfParser().parser self.text = parser.parse(content, lexer=lexer) def __str__(self): return AtfFile.template.render_unicode(**vars(self)) def serialize(self): return AtfFile.template.render_unicode(**vars(self)) def _debug_lex_and_yac_file(file): import codecs text = codecs.open(file, encoding='utf-8-sig').read() from pyoracc.atf.atffile import AtfLexer lexer = AtfLexer().lexer lexer.input(text) for tok in lexer: print(tok) print("Lexed file") lexer = AtfLexer().lexer parser = AtfParser().parser parser.parse(text, lexer=lexer) print("Parsed file")
Correct typo in debug function
Correct typo in debug function
Python
mit
UCL/pyoracc
b0e5dff69b9e40b916ad8a6655624de7fa85d247
chmvh_website/team/migrations/0002_auto_20161024_2338.py
chmvh_website/team/migrations/0002_auto_20161024_2338.py
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2016-10-24 23:38 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('team', '0001_initial'), ] operations = [ migrations.AlterModelOptions( name='teammember', options={'ordering': ('order',)}, ), migrations.AddField( model_name='teammember', name='order', field=models.PositiveSmallIntegerField(default=0), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2016-10-24 23:38 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('team', '0001_initial'), ] operations = [ migrations.AddField( model_name='teammember', name='order', field=models.PositiveSmallIntegerField(default=0), ), migrations.AlterModelOptions( name='teammember', options={'ordering': ('order',)}, ), ]
Change order of migration operations.
Change order of migration operations.
Python
mit
cdriehuys/chmvh-website,cdriehuys/chmvh-website,cdriehuys/chmvh-website
85fce5f5ab57b6c2144c92ec0d9b185740d7dc91
pyinform/__init__.py
pyinform/__init__.py
# Copyright 2016 ELIFE. All rights reserved. # Use of this source code is governed by a MIT # license that can be found in the LICENSE file. from ctypes import CDLL def get_libpath(): """ Get the library path of the the distributed inform binary. """ import os import re from os.path import dirname, abspath, realpath, join libre = re.compile(r"^inform-(\d+)\.(\d+)\.(\d+)$") root = dirname(abspath(realpath(__file__))) libdir = None major, minor, revision = 0, 0, 0 for _, dirnames, _ in os.walk(root): for dirname in dirnames: match = libre.match(dirname) if match: a, b, c = tuple(int(x) for x in match.group(1,2,3)) if (major, minor, revision) < (a,b,c): major, minor, revision = a, b, c libdir = join(root, match.group()) break break if libdir is None: raise ImportError("cannot find libinform") else: return "{}/lib/libinform.so.{}.{}.{}".format(libdir,major,minor,revision) _inform = CDLL(get_libpath())
# Copyright 2016 ELIFE. All rights reserved. # Use of this source code is governed by a MIT # license that can be found in the LICENSE file. from ctypes import CDLL def get_libpath(): """ Get the library path of the the distributed inform binary. """ import os import re from os.path import dirname, abspath, realpath, join from platform import system libre = re.compile(r"^inform-(\d+)\.(\d+)\.(\d+)$") root = dirname(abspath(realpath(__file__))) libdir = None major, minor, revision = 0, 0, 0 for _, dirnames, _ in os.walk(root): for dirname in dirnames: match = libre.match(dirname) if match: a, b, c = tuple(int(x) for x in match.group(1,2,3)) if (major, minor, revision) < (a,b,c): major, minor, revision = a, b, c libdir = join(root, match.group()) break break if libdir is None: raise ImportError("cannot find libinform") if system() is 'Windows': return "{}/lib/inform.dll".format(libdir) else: return "{}/lib/libinform.so.{}.{}.{}".format(libdir,major,minor,revision) _inform = CDLL(get_libpath())
Resolve the library on windows
Resolve the library on windows
Python
mit
ELIFE-ASU/PyInform
0d1904345d73bf067f8640d62f9d4186757239b6
dsub/_dsub_version.py
dsub/_dsub_version.py
# Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Single source of truth for dsub's version. This must remain small and dependency-free so that any dsub module may import it without creating circular dependencies. Note that this module is parsed as a text file by setup.py and changes to the format of this file could break setup.py. The version should follow formatting requirements specified in PEP-440. - https://www.python.org/dev/peps/pep-0440 A typical release sequence will be versioned as: 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ... """ DSUB_VERSION = '0.3.7'
# Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Single source of truth for dsub's version. This must remain small and dependency-free so that any dsub module may import it without creating circular dependencies. Note that this module is parsed as a text file by setup.py and changes to the format of this file could break setup.py. The version should follow formatting requirements specified in PEP-440. - https://www.python.org/dev/peps/pep-0440 A typical release sequence will be versioned as: 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ... """ DSUB_VERSION = '0.3.8.dev0'
Update dsub version to 0.3.8.dev0
Update dsub version to 0.3.8.dev0 PiperOrigin-RevId: 293000641
Python
apache-2.0
DataBiosphere/dsub,DataBiosphere/dsub
8098b4e73f0b407c47ecbe53318f2a246bd07d37
dsub/_dsub_version.py
dsub/_dsub_version.py
# Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Single source of truth for dsub's version. This must remain small and dependency-free so that any dsub module may import it without creating circular dependencies. Note that this module is parsed as a text file by setup.py and changes to the format of this file could break setup.py. The version should follow formatting requirements specified in PEP-440. - https://www.python.org/dev/peps/pep-0440 A typical release sequence will be versioned as: 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ... """ DSUB_VERSION = '0.1.4.dev0'
# Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Single source of truth for dsub's version. This must remain small and dependency-free so that any dsub module may import it without creating circular dependencies. Note that this module is parsed as a text file by setup.py and changes to the format of this file could break setup.py. The version should follow formatting requirements specified in PEP-440. - https://www.python.org/dev/peps/pep-0440 A typical release sequence will be versioned as: 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ... """ DSUB_VERSION = '0.1.5'
Bump version to 0.1.5 in preparation for a release
Bump version to 0.1.5 in preparation for a release PiperOrigin-RevId: 182068450
Python
apache-2.0
DataBiosphere/dsub,DataBiosphere/dsub
4551732c93b248e669b63d8ea6a9705c52b69dc3
projects/urls.py
projects/urls.py
from django.conf.urls import patterns, url urlpatterns = patterns('projects.views', url(r'^add/$', 'add_project', name='add_project'), url(r'^edit/(?P<project_id>\d+)/$', 'edit_project', name='edit_project'), url(r'^status/(?P<project_id>\d+)/$', 'edit_status', name='edit_status'), url(r'^archive/$', 'projects_archive', name='projects_archive'), url(r'^archive/review/(?P<project_id>\d+)/$', 'show_project', name='show-project'), )
from django.conf.urls import patterns, url urlpatterns = patterns('projects.views', url(r'^add/$', 'add_project', name='add_project'), url(r'^edit/(?P<project_id>\d+)/$', 'edit_project', name='edit_project'), url(r'^edit_status/(?P<project_id>\d+)/$', 'edit_status', name='edit_status'), url(r'^status/(?P<project_id>\d+)/$', 'edit_status', name='edit_status'), url(r'^archive/$', 'projects_archive', name='projects_archive'), url(r'^archive/review/(?P<project_id>\d+)/$', 'show_project', name='show-project'), )
Add url for project_status_edit option
Add url for project_status_edit option
Python
mit
Hackfmi/Diaphanum,Hackfmi/Diaphanum
1cc68fee10975f85ca5a2e2a63b972314a1b62d9
tests/test_redis_storage.py
tests/test_redis_storage.py
import unittest import datetime import hiro import redis from sifr.span import Minute, Day from sifr.storage import MemoryStorage, RedisStorage class RedisStorageTests(unittest.TestCase): def setUp(self): self.redis = redis.Redis() self.redis.flushall() def test_incr_simple_minute(self): span = Minute(datetime.datetime.now(), ["minute_span"]) storage = RedisStorage(self.redis) storage.incr(span) storage.incr(span) self.assertEqual(storage.get(span), 2) def test_incr_unique_minute(self): red = redis.Redis() span = Minute(datetime.datetime.now(), ["minute_span"]) storage = RedisStorage(red) storage.incr_unique(span, "1") storage.incr_unique(span, "1") storage.incr_unique(span, "2") self.assertEqual(storage.get_unique(span), 2) def test_tracker_minute(self): span = Minute(datetime.datetime.now(), ["minute_span"]) storage = RedisStorage(self.redis) storage.track(span, "1", 3) storage.track(span, "1", 3) storage.track(span, "2", 3) storage.track(span, "3", 3) self.assertEqual(storage.enumerate(span), set(["1", "2", "3"]))
import unittest import datetime import hiro import redis from sifr.span import Minute, Day from sifr.storage import MemoryStorage, RedisStorage class RedisStorageTests(unittest.TestCase): def setUp(self): self.redis = redis.Redis() self.redis.flushall() def test_incr_simple_minute(self): span = Minute(datetime.datetime.now(), ["minute_span"]) storage = RedisStorage(self.redis) storage.incr(span) storage.incr(span) self.assertEqual(storage.get(span), 2) def test_incr_unique_minute(self): red = redis.Redis() span = Minute(datetime.datetime.now(), ["minute_span"]) storage = RedisStorage(red) storage.incr_unique(span, "1") storage.incr_unique(span, "1") storage.incr_unique(span, "2") self.assertEqual(storage.get_unique(span), 2) def test_tracker_minute(self): span = Minute(datetime.datetime.now(), ["minute_span"]) storage = RedisStorage(self.redis) storage.track(span, "1") storage.track(span, "1") storage.track(span, "2") storage.track(span, "3") self.assertEqual(storage.enumerate(span), set(["1", "2", "3"]))
Remove old extra argument from track tests
Remove old extra argument from track tests
Python
mit
alisaifee/sifr,alisaifee/sifr
9f1783ac694d91b287dcb5840f54fb3df746a963
bot/action/core/action.py
bot/action/core/action.py
from bot.api.api import Api from bot.multithreading.scheduler import SchedulerApi from bot.storage import Config, State, Cache class Action: def __init__(self): pass def get_name(self): return self.__class__.__name__ def setup(self, api: Api, config: Config, state: State, cache: Cache, scheduler: SchedulerApi): self.api = api self.config = config self.state = state self.cache = cache self.scheduler = scheduler self.post_setup() def post_setup(self): pass def process(self, event): pass class ActionGroup(Action): def __init__(self, *actions): super().__init__() self.actions = list(actions) def add(self, *actions): self.actions.extend(actions) def setup(self, *args): super().setup(*args) self.for_each(lambda action: action.setup(*args)) def process(self, event): self.for_each(lambda action: action.process(event._copy())) def for_each(self, func): for action in self.actions: func(action) class IntermediateAction(ActionGroup): def __init__(self): super().__init__() def then(self, *next_actions): self.add(*next_actions) return self def _continue(self, event): super().process(event)
from bot.api.api import Api from bot.multithreading.scheduler import SchedulerApi from bot.storage import Config, State, Cache class Action: def __init__(self): pass def get_name(self): return self.__class__.__name__ def setup(self, api: Api, config: Config, state: State, cache: Cache, scheduler: SchedulerApi): self.api = api self.config = config self.state = state self.cache = cache self.scheduler = scheduler self.post_setup() def post_setup(self): pass def process(self, event): pass def pre_shutdown(self): pass def shutdown(self): self.pre_shutdown() class ActionGroup(Action): def __init__(self, *actions): super().__init__() self.actions = list(actions) def add(self, *actions): self.actions.extend(actions) def setup(self, *args): super().setup(*args) self.for_each(lambda action: action.setup(*args)) def process(self, event): self.for_each(lambda action: action.process(event._copy())) def shutdown(self): self.for_each(lambda action: action.shutdown()) super().shutdown() def for_each(self, func): for action in self.actions: func(action) class IntermediateAction(ActionGroup): def __init__(self): super().__init__() def then(self, *next_actions): self.add(*next_actions) return self def _continue(self, event): super().process(event)
Add shutdown callback support to Action
Add shutdown callback support to Action
Python
agpl-3.0
alvarogzp/telegram-bot,alvarogzp/telegram-bot
85efa9c105ddb9240a25be433de76ef21b3ed2b3
xutils/const.py
xutils/const.py
# encoding: utf-8 import sys class _const(object): class ConstError(TypeError): pass class ConstCaseError(ConstError): pass def __setattr__(self, name, value): if name in self.__dict__: raise self.ConstError, "Can't change const.{0}".format(name) if not name.isupper(): raise self.ConstCaseError, 'const name "{0}" is not all uppercase'.format(name) self.__dict__[name] = value sys.modules[__name__.title()] = _const sys.modules[__name__] = _const()
# encoding: utf-8 import sys class _const(object): class ConstError(TypeError): pass class ConstCaseError(ConstError): pass def __setattr__(self, name, value): if name in self.__dict__: raise self.ConstError("Can't change const.{0}".format(name)) if not name.isupper(): raise self.ConstCaseError('const name "{0}" is not all uppercase'.format(name)) self.__dict__[name] = value sys.modules[__name__.title()] = _const sys.modules[__name__] = _const()
Fix a bug on Python 3.6
Fix a bug on Python 3.6
Python
mit
xgfone/pycom,xgfone/xutils
2bcc941b015c443c64f08a13012e8caf70028754
ideascube/search/migrations/0001_initial.py
ideascube/search/migrations/0001_initial.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models import ideascube.search.models class Migration(migrations.Migration): dependencies = [ ] operations = [ migrations.CreateModel( name='Search', fields=[ ('rowid', models.IntegerField(serialize=False, primary_key=True)), ('model', models.CharField(max_length=64)), ('model_id', models.IntegerField()), ('public', models.BooleanField(default=True)), ('text', ideascube.search.models.SearchField()), ], options={ 'db_table': 'idx', 'managed': False, }, ), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations from ideascube.search.utils import create_index_table class CreateSearchModel(migrations.CreateModel): def database_forwards(self, *_): # Don't run the parent method, we create the table our own way create_index_table() class Migration(migrations.Migration): dependencies = [ ] operations = [ CreateSearchModel( name='Search', fields=[], options={ 'db_table': 'idx', 'managed': False, }, ), ]
Fix the initial search migration
Fix the initial search migration There is no point in creating the model in this way, that's just not how it's used: instead we want to use the FTS4 extension from SQLite.
Python
agpl-3.0
ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube
fc818ccd0d83ff6b37b38e5e9d03abcae408b503
froide/problem/templatetags/problemreport_tags.py
froide/problem/templatetags/problemreport_tags.py
from collections import defaultdict from django import template from ..models import ProblemReport from ..forms import ProblemReportForm register = template.Library() @register.inclusion_tag('problem/message_toolbar_item.html') def render_problem_button(message): if not hasattr(message, 'problemreports'): # Get all problem reports for all messages request = message.request reports = ProblemReport.objects.filter(message__in=request.messages) message_reports = defaultdict(list) for report in reports: message_reports[report.message_id].append(report) for message in request.messages: message.problemreports = message_reports[message.id] message.problemreports_count = len(message.problemreports) message.problemreports_unresolved_count = len([ r for r in message.problemreports if not r.resolved ]) message.problemreports_form = ProblemReportForm(message=message) return { 'message': message }
from collections import defaultdict from django import template from ..models import ProblemReport from ..forms import ProblemReportForm register = template.Library() @register.inclusion_tag('problem/message_toolbar_item.html') def render_problem_button(message): if not hasattr(message, 'problemreports'): # Get all problem reports for all messages request = message.request reports = ProblemReport.objects.filter(message__in=request.messages) message_reports = defaultdict(list) for report in reports: message_reports[report.message_id].append(report) for mes in request.messages: mes.problemreports = message_reports[mes.id] mes.problemreports_count = len(mes.problemreports) mes.problemreports_unresolved_count = len([ r for r in mes.problemreports if not r.resolved ]) mes.problemreports_form = ProblemReportForm(message=mes) return { 'message': message }
Fix overriding variable in problem report tag
Fix overriding variable in problem report tag
Python
mit
stefanw/froide,stefanw/froide,stefanw/froide,fin/froide,fin/froide,stefanw/froide,stefanw/froide,fin/froide,fin/froide
a3347eff5791c89949a88988a958c45ec50cccdf
runtests.py
runtests.py
#!/usr/bin/env python import os, sys import django from django.conf import settings DIRNAME = os.path.dirname(__file__) settings.configure( DEBUG = True, DATABASES = { 'default': {'ENGINE': 'django.db.backends.sqlite3'}, }, INSTALLED_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.admin', 'hashbrown', ), # Django 1.7 raises a warning if this isn't set. Pollutes test output. MIDDLEWARE_CLASSES = (), ) try: django.setup() except AttributeError: # Running Django<1.7 pass try: from django.test.runner import DiscoverRunner as TestSuiteRunner except ImportError: # Running Django<1.6 from django.test.simple import DjangoTestSuiteRunner as TestSuiteRunner test_runner = TestSuiteRunner(verbosity=1) failures = test_runner.run_tests(['hashbrown', ]) if failures: sys.exit(failures)
#!/usr/bin/env python import os, sys import django from django.conf import settings DIRNAME = os.path.dirname(__file__) settings.configure( DEBUG = True, DATABASES = { 'default': {'ENGINE': 'django.db.backends.sqlite3'}, }, INSTALLED_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.admin', 'hashbrown', ), # Django 1.7 raises a warning if this isn't set. Pollutes test output. MIDDLEWARE_CLASSES = (), TEMPLATES = [{ 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': {}, }], ) try: django.setup() except AttributeError: # Running Django<1.7 pass try: from django.test.runner import DiscoverRunner as TestSuiteRunner except ImportError: # Running Django<1.6 from django.test.simple import DjangoTestSuiteRunner as TestSuiteRunner test_runner = TestSuiteRunner(verbosity=1) failures = test_runner.run_tests(['hashbrown', ]) if failures: sys.exit(failures)
Add `TEMPLATE` config to test runner, to define template backend.
Add `TEMPLATE` config to test runner, to define template backend.
Python
bsd-2-clause
potatolondon/django-hashbrown
af5eae0b477c73c1c8d1bbce646d94858d157142
whip/web.py
whip/web.py
#!/usr/bin/env python import socket from flask import Flask, abort, make_response, request from whip.db import Database app = Flask(__name__) app.config.from_envvar('WHIP_SETTINGS', silent=True) db = None @app.before_first_request def _open_db(): global db db = Database(app.config['DATABASE_DIR']) @app.route('/ip/<ip>') def lookup(ip): try: key = socket.inet_aton(ip) except socket.error: abort(400) dt = request.args.get('datetime') if dt: dt = dt.encode('ascii') else: dt = None # account for empty parameter value info_as_json = db.lookup(key, dt) if info_as_json is None: abort(404) response = make_response(info_as_json) response.headers['Content-type'] = 'application/json' return response
#!/usr/bin/env python import socket from flask import Flask, abort, make_response, request from whip.db import Database app = Flask(__name__) app.config.from_envvar('WHIP_SETTINGS', silent=True) db = None @app.before_first_request def _open_db(): global db db = Database(app.config['DATABASE_DIR']) @app.route('/ip/<ip>') def lookup(ip): try: key = socket.inet_aton(ip) except socket.error: abort(400) dt = request.args.get('datetime') if dt: dt = dt.encode('ascii') else: dt = None # account for empty parameter value info_as_json = db.lookup(key, dt) if info_as_json is None: info_as_json = b'{}' # empty dict, JSON-encoded response = make_response(info_as_json) response.headers['Content-type'] = 'application/json' return response
Return empty responses (not HTTP 404) in REST API for missing data
Return empty responses (not HTTP 404) in REST API for missing data
Python
bsd-3-clause
wbolster/whip
a5b73a7ded0e277662308e0b4d38ac0429c404fb
django_facebook/models.py
django_facebook/models.py
from django.db import models class FacebookProfileModel(models.Model): ''' Abstract class to add to your profile model. NOTE: If you don't use this this abstract class, make sure you copy/paste the fields in. ''' about_me = models.TextField(blank=True, null=True) facebook_id = models.IntegerField(blank=True, null=True) facebook_name = models.CharField(max_length=255, blank=True, null=True) facebook_profile_url = models.TextField(blank=True, null=True) website_url = models.TextField(blank=True, null=True) blog_url = models.TextField(blank=True, null=True) image = models.ImageField(blank=True, null=True, upload_to='profile_images') date_of_birth = models.DateField(blank=True, null=True) class Meta: abstract = True
from django.db import models from django.contrib.auth.models import User class FacebookProfileModel(models.Model): ''' Abstract class to add to your profile model. NOTE: If you don't use this this abstract class, make sure you copy/paste the fields in. ''' user = models.OneToOneField(User) about_me = models.TextField(blank=True, null=True) facebook_id = models.IntegerField(blank=True, null=True) facebook_name = models.CharField(max_length=255, blank=True, null=True) facebook_profile_url = models.TextField(blank=True, null=True) website_url = models.TextField(blank=True, null=True) blog_url = models.TextField(blank=True, null=True) image = models.ImageField(blank=True, null=True, upload_to='profile_images') date_of_birth = models.DateField(blank=True, null=True) def __unicode__(self): return self.user.__unicode__() class Meta: abstract = True
Add reference to user model and __unicode__() method to FacebookProfileModel
Add reference to user model and __unicode__() method to FacebookProfileModel
Python
bsd-3-clause
pjdelport/Django-facebook,QLGu/Django-facebook,Shekharrajak/Django-facebook,VishvajitP/Django-facebook,troygrosfield/Django-facebook,QLGu/Django-facebook,cyrixhero/Django-facebook,rafaelgontijo/Django-facebook-fork,fyndsi/Django-facebook,jcpyun/Django-facebook,abendleiter/Django-facebook,danosaure/Django-facebook,cyrixhero/Django-facebook,andriisoldatenko/Django-facebook,Fiedzia/Django-facebook,sitsbeyou/Django-facebook,tuxos/Django-facebook,takeshineshiro/Django-facebook,troygrosfield/Django-facebook,takeshineshiro/Django-facebook,abhijo89/Django-facebook,selwin/Django-facebook,fivejjs/Django-facebook,ganescoo/Django-facebook,selwin/Django-facebook,PeterWangPo/Django-facebook,fyndsi/Django-facebook,Shekharrajak/Django-facebook,abendleiter/Django-facebook,abhijo89/Django-facebook,selwin/Django-facebook,javipalanca/Django-facebook,rafaelgontijo/Django-facebook-fork,pjdelport/Django-facebook,sitsbeyou/Django-facebook,Fiedzia/Django-facebook,Shekharrajak/Django-facebook,PeterWangPo/Django-facebook,ganescoo/Django-facebook,javipalanca/Django-facebook,andriisoldatenko/Django-facebook,sitsbeyou/Django-facebook,christer155/Django-facebook,troygrosfield/Django-facebook,VishvajitP/Django-facebook,tuxos/Django-facebook,jcpyun/Django-facebook,javipalanca/Django-facebook,VishvajitP/Django-facebook,tuxos/Django-facebook,fyndsi/Django-facebook,danosaure/Django-facebook,christer155/Django-facebook,andriisoldatenko/Django-facebook,cyrixhero/Django-facebook,jcpyun/Django-facebook,pjdelport/Django-facebook,rafaelgontijo/Django-facebook-fork,ganescoo/Django-facebook,christer155/Django-facebook,fivejjs/Django-facebook,QLGu/Django-facebook,danosaure/Django-facebook,takeshineshiro/Django-facebook,abendleiter/Django-facebook,andriisoldatenko/Django-facebook,PeterWangPo/Django-facebook,abhijo89/Django-facebook,Fiedzia/Django-facebook,fivejjs/Django-facebook
3122736e0eccd4d4b1f003faa1db6ec05710883f
addstr.py
addstr.py
#!/usr/bin/python import argparse from dx.dex import Dex from sha1 import update_signature from adler32 import update_checksum def main(): parser = argparse.ArgumentParser(description="Parse and reconstruct dex file") parser.add_argument('target',help='Target DEX file') parser.add_argument('string',help='String to be added') args = parser.parse_args() dex = Dex(args.target) dex.add_string(args.string) dex.save("out2.dex") update_signature("out2.dex") update_checksum("out2.dex") print "Done" if __name__ == '__main__': main()
#!/usr/bin/python import argparse from dx.dex import Dex from dx.hash import update_signature, update_checksum def main(): parser = argparse.ArgumentParser(description="Parse and reconstruct dex file") parser.add_argument('target',help='Target DEX file') parser.add_argument('string',help='String to be added') args = parser.parse_args() dex = Dex(args.target) dex.add_string(args.string) dex.save("out2.dex") update_signature("out2.dex") update_checksum("out2.dex") print "Done" if __name__ == '__main__': main()
Fix attempted import from non-existent module.
Fix attempted import from non-existent module.
Python
bsd-3-clause
strazzere/dexterity,strazzere/dexterity,rchiossi/dexterity,strazzere/dexterity,rchiossi/dexterity,rchiossi/dexterity
edd50431f9c99bcbc765cc85786ead60ba8ba6e4
admin/base/migrations/0002_groups.py
admin/base/migrations/0002_groups.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations from django.contrib.auth.models import Group import logging logger = logging.getLogger(__file__) def add_groups(*args): group, created = Group.objects.get_or_create(name='nodes_and_users') if created: logger.info('nodes_and_users group created') try: group = Group.objects.get(name='prereg_group') group.name = 'prereg' group.save() logger.info('prereg_group renamed to prereg') except Group.DoesNotExist: group, created = Group.objects.get_or_create(name='prereg') if created: logger.info('prereg group created') class Migration(migrations.Migration): dependencies = [ ('base', '0001_groups'), ] operations = [ migrations.RunPython(add_groups), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations from django.contrib.auth.models import Group import logging logger = logging.getLogger(__file__) def add_groups(*args): group, created = Group.objects.get_or_create(name='nodes_and_users') if created: logger.info('nodes_and_users group created') try: group = Group.objects.get(name='prereg_group') group.name = 'prereg' group.save() logger.info('prereg_group renamed to prereg') except Group.DoesNotExist: group, created = Group.objects.get_or_create(name='prereg') if created: logger.info('prereg group created') def remove_groups(*args): Group.objects.filter(name='nodes_and_users').delete() group = Group.objects.get(name='prereg') group.name = 'prereg_group' group.save() class Migration(migrations.Migration): dependencies = [ ('base', '0001_groups'), ] operations = [ migrations.RunPython(add_groups, remove_groups), ]
Add reverse migration for new groups
Add reverse migration for new groups
Python
apache-2.0
brianjgeiger/osf.io,chennan47/osf.io,CenterForOpenScience/osf.io,sloria/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,brianjgeiger/osf.io,monikagrabowska/osf.io,binoculars/osf.io,acshi/osf.io,chrisseto/osf.io,acshi/osf.io,crcresearch/osf.io,aaxelb/osf.io,erinspace/osf.io,brianjgeiger/osf.io,chrisseto/osf.io,erinspace/osf.io,baylee-d/osf.io,mfraezz/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,binoculars/osf.io,saradbowman/osf.io,acshi/osf.io,HalcyonChimera/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,chennan47/osf.io,Johnetordoff/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,leb2dg/osf.io,felliott/osf.io,hmoco/osf.io,cwisecarver/osf.io,pattisdr/osf.io,chrisseto/osf.io,adlius/osf.io,pattisdr/osf.io,monikagrabowska/osf.io,adlius/osf.io,icereval/osf.io,baylee-d/osf.io,cslzchen/osf.io,felliott/osf.io,caneruguz/osf.io,TomBaxter/osf.io,caneruguz/osf.io,baylee-d/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,aaxelb/osf.io,caseyrollins/osf.io,hmoco/osf.io,felliott/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,aaxelb/osf.io,crcresearch/osf.io,mattclark/osf.io,cwisecarver/osf.io,laurenrevere/osf.io,hmoco/osf.io,saradbowman/osf.io,cslzchen/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,monikagrabowska/osf.io,mfraezz/osf.io,icereval/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,TomBaxter/osf.io,laurenrevere/osf.io,sloria/osf.io,caseyrollins/osf.io,Nesiehr/osf.io,leb2dg/osf.io,acshi/osf.io,caseyrollins/osf.io,HalcyonChimera/osf.io,Nesiehr/osf.io,mattclark/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,felliott/osf.io,icereval/osf.io,crcresearch/osf.io,mattclark/osf.io,acshi/osf.io,sloria/osf.io,caneruguz/osf.io,mfraezz/osf.io,binoculars/osf.io,adlius/osf.io,aaxelb/osf.io,caneruguz/osf.io,Nesiehr/osf.io,laurenrevere/osf.io,erinspace/osf.io,hmoco/osf.io,chrisseto/osf.io,Nesiehr/osf.io,chennan47/osf.io,adlius/osf.io
eef28c81f19d7e5eb72635cc2e6bf3b74331c743
quilt/patch.py
quilt/patch.py
# vim: fileencoding=utf-8 et sw=4 ts=4 tw=80: # python-quilt - A Python implementation of the quilt patch system # # Copyright (C) 2012 Björn Ricks <[email protected]> # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA from quilt.utils import Process class Patch(object): def __init__(self, cwd, patch_file, backup=False, prefix=None): cmd = ["patch"] if backup: cmd.append("--backup") if prefix: cmd.append("--prefix") cmd.append(prefix) cmd.append("-i") cmd.append(patch_file) Process(cmd).run(cwd=cwd)
# vim: fileencoding=utf-8 et sw=4 ts=4 tw=80: # python-quilt - A Python implementation of the quilt patch system # # Copyright (C) 2012 Björn Ricks <[email protected]> # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA import os from quilt.utils import Process class Patch(object): def __init__(self, cwd, patch_file, backup=False, prefix=None): cmd = ["patch"] if backup: cmd.append("--backup") if prefix: cmd.append("--prefix") if not prefix[-1] == os.sep: prefix += os.sep cmd.append(prefix) cmd.append("-i") cmd.append(patch_file) Process(cmd).run(cwd=cwd)
Patch parameter --prefix does need a path seperator
Patch parameter --prefix does need a path seperator The --prefix parameter of the patch command needs a path seperator at the end to store the backup in a directory.
Python
mit
vadmium/python-quilt,bjoernricks/python-quilt
306a09153de72f5f9f4043fa45472440065ac473
bindings/pyroot/JupyROOT/kernel/magics/jsrootmagic.py
bindings/pyroot/JupyROOT/kernel/magics/jsrootmagic.py
# -*- coding:utf-8 -*- #----------------------------------------------------------------------------- # Copyright (c) 2016, ROOT Team. # Authors: Danilo Piparo <[email protected]> CERN #----------------------------------------------------------------------------- from JupyROOT.utils import enableJSVis, disableJSVis, enableJSVisDebug, TBufferJSONErrorMessage, TBufferJSONAvailable from metakernel import Magic, option class JSRootMagics(Magic): def __init__(self, kernel): super(JSRootMagics, self).__init__(kernel) @option('arg', default="on", help='Enable or disable JavaScript visualisation. Possible values: on (default), off') def cell_jsroot(self, args): '''Change the visualisation of plots from images to interactive JavaScript objects.''' if args == 'on' or args == '': self.printErrorIfNeeded() enableJSVis() elif args == 'off': disableJSVis() elif args == 'debug': self.printErrorIfNeeded() enableJSVisDebug() def printErrorIfNeeded(self): if not TBufferJSONAvailable(): self.kernel.Error(TBufferJSONErrorMessage) def register_magics(kernel): kernel.register_magics(JSRootMagics)
# -*- coding:utf-8 -*- #----------------------------------------------------------------------------- # Copyright (c) 2016, ROOT Team. # Authors: Danilo Piparo <[email protected]> CERN #----------------------------------------------------------------------------- from JupyROOT.utils import enableJSVis, disableJSVis, enableJSVisDebug, TBufferJSONErrorMessage, TBufferJSONAvailable from metakernel import Magic, option class JSRootMagics(Magic): def __init__(self, kernel): super(JSRootMagics, self).__init__(kernel) @option('arg', default="on", help='Enable or disable JavaScript visualisation. Possible values: on (default), off') def line_jsroot(self, args): '''Change the visualisation of plots from images to interactive JavaScript objects.''' if args == 'on' or args == '': self.printErrorIfNeeded() enableJSVis() elif args == 'off': disableJSVis() elif args == 'debug': self.printErrorIfNeeded() enableJSVisDebug() def printErrorIfNeeded(self): if not TBufferJSONAvailable(): self.kernel.Error(TBufferJSONErrorMessage) def register_magics(kernel): kernel.register_magics(JSRootMagics)
Make the jsroot magic a line magic for the C++ kernel
Make the jsroot magic a line magic for the C++ kernel
Python
lgpl-2.1
veprbl/root,beniz/root,gbitzes/root,simonpf/root,mhuwiler/rootauto,CristinaCristescu/root,georgtroska/root,sawenzel/root,beniz/root,pspe/root,olifre/root,krafczyk/root,abhinavmoudgil95/root,thomaskeck/root,lgiommi/root,thomaskeck/root,root-mirror/root,georgtroska/root,davidlt/root,satyarth934/root,thomaskeck/root,krafczyk/root,georgtroska/root,zzxuanyuan/root-compressor-dummy,Y--/root,CristinaCristescu/root,olifre/root,abhinavmoudgil95/root,simonpf/root,agarciamontoro/root,karies/root,olifre/root,gbitzes/root,mhuwiler/rootauto,buuck/root,BerserkerTroll/root,BerserkerTroll/root,bbockelm/root,veprbl/root,root-mirror/root,sawenzel/root,veprbl/root,pspe/root,bbockelm/root,buuck/root,sawenzel/root,agarciamontoro/root,Y--/root,zzxuanyuan/root,beniz/root,gbitzes/root,karies/root,veprbl/root,krafczyk/root,agarciamontoro/root,gganis/root,veprbl/root,root-mirror/root,zzxuanyuan/root-compressor-dummy,bbockelm/root,bbockelm/root,agarciamontoro/root,krafczyk/root,krafczyk/root,zzxuanyuan/root-compressor-dummy,mhuwiler/rootauto,beniz/root,BerserkerTroll/root,krafczyk/root,satyarth934/root,mhuwiler/rootauto,root-mirror/root,Y--/root,georgtroska/root,abhinavmoudgil95/root,simonpf/root,georgtroska/root,Y--/root,gganis/root,thomaskeck/root,simonpf/root,karies/root,pspe/root,pspe/root,olifre/root,davidlt/root,pspe/root,zzxuanyuan/root,sawenzel/root,CristinaCristescu/root,pspe/root,georgtroska/root,veprbl/root,gganis/root,abhinavmoudgil95/root,gganis/root,zzxuanyuan/root-compressor-dummy,karies/root,CristinaCristescu/root,root-mirror/root,davidlt/root,olifre/root,agarciamontoro/root,abhinavmoudgil95/root,bbockelm/root,olifre/root,root-mirror/root,sawenzel/root,BerserkerTroll/root,buuck/root,beniz/root,karies/root,zzxuanyuan/root-compressor-dummy,pspe/root,beniz/root,pspe/root,agarciamontoro/root,buuck/root,gbitzes/root,BerserkerTroll/root,sawenzel/root,root-mirror/root,krafczyk/root,satyarth934/root,CristinaCristescu/root,karies/root,Y--/root,thomaskeck/root,beniz/root,thomaskeck/root,olifre/root,karies/root,satyarth934/root,gbitzes/root,mhuwiler/rootauto,buuck/root,olifre/root,lgiommi/root,satyarth934/root,olifre/root,zzxuanyuan/root,abhinavmoudgil95/root,simonpf/root,mhuwiler/rootauto,satyarth934/root,veprbl/root,buuck/root,beniz/root,veprbl/root,agarciamontoro/root,veprbl/root,georgtroska/root,lgiommi/root,sawenzel/root,karies/root,zzxuanyuan/root,abhinavmoudgil95/root,zzxuanyuan/root-compressor-dummy,simonpf/root,sawenzel/root,lgiommi/root,simonpf/root,bbockelm/root,buuck/root,lgiommi/root,davidlt/root,zzxuanyuan/root,BerserkerTroll/root,bbockelm/root,agarciamontoro/root,veprbl/root,mhuwiler/rootauto,thomaskeck/root,pspe/root,CristinaCristescu/root,olifre/root,karies/root,gbitzes/root,bbockelm/root,abhinavmoudgil95/root,bbockelm/root,beniz/root,krafczyk/root,veprbl/root,BerserkerTroll/root,bbockelm/root,mhuwiler/rootauto,thomaskeck/root,zzxuanyuan/root-compressor-dummy,davidlt/root,zzxuanyuan/root,gganis/root,root-mirror/root,mhuwiler/rootauto,beniz/root,davidlt/root,CristinaCristescu/root,krafczyk/root,satyarth934/root,agarciamontoro/root,zzxuanyuan/root,agarciamontoro/root,gbitzes/root,BerserkerTroll/root,sawenzel/root,beniz/root,olifre/root,gganis/root,pspe/root,krafczyk/root,buuck/root,BerserkerTroll/root,georgtroska/root,Y--/root,satyarth934/root,gganis/root,lgiommi/root,buuck/root,Y--/root,simonpf/root,Y--/root,zzxuanyuan/root-compressor-dummy,BerserkerTroll/root,buuck/root,satyarth934/root,zzxuanyuan/root,gbitzes/root,zzxuanyuan/root-compressor-dummy,zzxuanyuan/root-compressor-dummy,lgiommi/root,abhinavmoudgil95/root,zzxuanyuan/root,root-mirror/root,Y--/root,simonpf/root,sawenzel/root,mhuwiler/rootauto,buuck/root,karies/root,mhuwiler/rootauto,georgtroska/root,zzxuanyuan/root,thomaskeck/root,zzxuanyuan/root,CristinaCristescu/root,krafczyk/root,davidlt/root,gbitzes/root,thomaskeck/root,simonpf/root,simonpf/root,sawenzel/root,zzxuanyuan/root-compressor-dummy,zzxuanyuan/root,davidlt/root,satyarth934/root,abhinavmoudgil95/root,CristinaCristescu/root,davidlt/root,karies/root,Y--/root,agarciamontoro/root,pspe/root,bbockelm/root,georgtroska/root,davidlt/root,gbitzes/root,davidlt/root,root-mirror/root,gganis/root,gbitzes/root,lgiommi/root,lgiommi/root,abhinavmoudgil95/root,gganis/root,CristinaCristescu/root,gganis/root,CristinaCristescu/root,root-mirror/root,Y--/root,gganis/root,lgiommi/root,georgtroska/root,satyarth934/root,lgiommi/root,BerserkerTroll/root
a37ac8daad8eee1f044d3e19a80a172138460ec3
google_analytics/models.py
google_analytics/models.py
from django.db import models from django.conf import settings from django.contrib.sites.admin import SiteAdmin from django.contrib.sites.models import Site from django.contrib import admin if getattr(settings, 'GOOGLE_ANALYTICS_MODEL', False): class Analytic(models.Model): site = models.ForeignKey(Site, unique=True) analytics_code = models.CharField(blank=True, max_length=100) def __unicode__(self): return u"%s" % (self.analytics_code) class AnalyticAdmin(admin.ModelAdmin): list_display = ('site', 'analytics_code',) admin.site.register(Analytic, AnalyticAdmin)
from django.contrib import admin from django.contrib.sites.models import Site from django.db import models class Analytic(models.Model): site = models.ForeignKey(Site, unique=True) analytics_code = models.CharField(blank=True, max_length=100)
Fix django version problem with new menu options in admin app.
Fix django version problem with new menu options in admin app.
Python
agpl-3.0
OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,OSSystems/lava-server
ec04b842c21cddaef1cf010e419113e83f3be3f1
tests/test_logging_service.py
tests/test_logging_service.py
import os.path import sys import unittest sys.path.append(os.path.join(os.path.dirname(__file__), '../')) sys.path.append(os.path.join(os.path.dirname(__file__), '../vCenterShell')) from vCenterShell.pycommon.logging_service import LoggingService class TestLoggingService(unittest.TestCase): def test_logging_service_01(self): log_file_name = "test_log.log" LoggingService("CRITICAL", "DEBUG", log_file_name) self.assertTrue(os.path.isfile(log_file_name)) os.unlink(log_file_name) def test_logging_service_02(self): log_file_name = "test_log.log" LoggingService("DEBUG", "CRITICAL", log_file_name) self.assertTrue(os.path.isfile(log_file_name)) self.assertEquals(os.path.getsize(log_file_name), 0) os.unlink(log_file_name)
import os.path import sys import unittest sys.path.append(os.path.join(os.path.dirname(__file__), '../')) sys.path.append(os.path.join(os.path.dirname(__file__), '../vCenterShell')) from vCenterShell.pycommon.logging_service import LoggingService class TestLoggingService(unittest.TestCase): def test_logging_service_01(self): log_file_name = "test_log.log" LoggingService("CRITICAL", "DEBUG", None) self.assertFalse(os.path.isfile(log_file_name)) # LoggingService("CRITICAL", "DEBUG", log_file_name) # self.assertTrue(os.path.isfile(log_file_name)) # os.unlink(log_file_name) def test_logging_service_02(self): log_file_name = "test_log.log" LoggingService("DEBUG", "CRITICAL", None) self.assertFalse(os.path.isfile(log_file_name)) # LoggingService("DEBUG", "CRITICAL", log_file_name) # self.assertTrue(os.path.isfile(log_file_name)) # self.assertEquals(os.path.getsize(log_file_name), 0) # os.unlink(log_file_name)
Remove File Creation Command for UnitTesting
Remove File Creation Command for UnitTesting
Python
apache-2.0
QualiSystems/vCenterShell,QualiSystems/vCenterShell
2b99108a817a642c86be06a14ac8d71cdc339555
scripts/speak.py
scripts/speak.py
#!/usr/bin/env python import rospy from sound_play.msg import SoundRequest from sound_play.libsoundplay import SoundClient from std_msgs.msg import String class ChatbotSpeaker: def __init__(self): rospy.init_node('chatbot_speaker') self._client = SoundClient() rospy.Subscriber('chatbot_responses', String, self._response_callback) rospy.spin() def _response_callback(self, data): self._client.say(data.data) def main(): speaker = ChatbotSpeaker() if __name__ == '__main__': main()
#!/usr/bin/env python import os import rospy from sound_play.msg import SoundRequest from sound_play.libsoundplay import SoundClient from std_msgs.msg import String import urllib tts_cmd = ( 'wget -q -U "Mozilla/5.0" -O - "http://translate.google.com/translate_tts?tl=en-uk&q={}" > /tmp/speech.mp3' ) sox_cmd = 'sox /tmp/speech.mp3 /tmp/speech.wav' class ChatbotSpeaker: def __init__(self): rospy.init_node('chatbot_speaker') self._client = SoundClient() rospy.Subscriber('chatbot_responses', String, self._response_callback) rospy.spin() def _response_callback(self, data): query = urllib.quote(data.data) os.system(tts_cmd.format(query)) os.system(sox_cmd) self._client.playWave('/tmp/speech.wav') def main(): speaker = ChatbotSpeaker() if __name__ == '__main__': main()
Use Google Translate API to get a female TTS
Use Google Translate API to get a female TTS
Python
mit
jstnhuang/chatbot
11be4b77e84c721ef8de583b0dcf1035367d4b25
libtmux/__about__.py
libtmux/__about__.py
__title__ = 'libtmux' __package_name__ = 'libtmux' __version__ = '0.8.0' __description__ = 'scripting library / orm for tmux' __email__ = '[email protected]' __author__ = 'Tony Narlock' __github__ = 'https://github.com/tmux-python/libtmux' __license__ = 'MIT' __copyright__ = 'Copyright 2016-2018 Tony Narlock'
__title__ = 'libtmux' __package_name__ = 'libtmux' __version__ = '0.8.0' __description__ = 'scripting library / orm for tmux' __email__ = '[email protected]' __author__ = 'Tony Narlock' __github__ = 'https://github.com/tmux-python/libtmux' __pypi__ = 'https://pypi.python.org/pypi/libtmux' __license__ = 'MIT' __copyright__ = 'Copyright 2016-2018 Tony Narlock'
Add __pypi__ url to metadata
Add __pypi__ url to metadata
Python
bsd-3-clause
tony/libtmux
f3fef8dab576ef5d7a4120a4041ade326868f0ca
flexget/plugins/ui/execute.py
flexget/plugins/ui/execute.py
import logging from flask import render_template, request, Response, redirect, flash from flask import Module, escape from flexget.webui import register_plugin, manager, BufferQueue from Queue import Empty from flask.helpers import jsonify execute = Module(__name__, url_prefix='/execute') log = logging.getLogger('ui.execute') bufferqueue = BufferQueue() @execute.route('/', methods=['POST', 'GET']) def index(): context = {'help': manager.parser.get_help()} if request.method == 'POST': options = manager.parser.parse_args(request.form.get('options', ''))[0] if manager.parser.error_msg: flash(escape(manager.parser.error_msg), 'error') context['options'] = request.form['options'] else: flash('Manual execution started.', 'success') from flexget.webui import executor executor.execute(options=options, output=bufferqueue) return render_template('execute.html', **context) @execute.route('/progress.json') def progress(): ''' Gives takes messages from the queue and exports them to JSON. ''' result = {'items': []} try: while 1: item = bufferqueue.get_nowait() if item != '\n': result['items'].append(item) except Empty: pass return jsonify(result) register_plugin(execute, menu='Execute')
import logging from flask import render_template, request, Response, redirect, flash from flask import Module, escape from flexget.webui import register_plugin, manager, BufferQueue from Queue import Empty from flask.helpers import jsonify execute = Module(__name__, url_prefix='/execute') log = logging.getLogger('ui.execute') bufferqueue = BufferQueue() @execute.route('/', methods=['POST', 'GET']) def index(): context = {'help': manager.parser.get_help()} if request.method == 'POST': options = manager.parser.parse_args(request.form.get('options', ''))[0] if manager.parser.error_msg: flash(escape(manager.parser.error_msg), 'error') context['options'] = request.form['options'] else: flash('Manual execution started.', 'success') from flexget.webui import executor executor.execute(options=options, output=bufferqueue) return render_template('execute.html', **context) @execute.route('/progress.json') def progress(): ''' Gives takes messages from the queue and exports them to JSON. ''' result = {'items': []} try: while 1: item = bufferqueue.get_nowait() if item != '\n': result['items'].append(item) bufferqueue.task_done() except Empty: pass return jsonify(result) register_plugin(execute, menu='Execute')
Fix an issue with repeated messages in json execution output provider.
Fix an issue with repeated messages in json execution output provider. git-svn-id: 555d7295f8287ebc42f8316c6775e40d702c4756@1726 3942dd89-8c5d-46d7-aeed-044bccf3e60c
Python
mit
oxc/Flexget,tsnoam/Flexget,offbyone/Flexget,malkavi/Flexget,ibrahimkarahan/Flexget,ratoaq2/Flexget,asm0dey/Flexget,sean797/Flexget,OmgOhnoes/Flexget,ibrahimkarahan/Flexget,drwyrm/Flexget,jawilson/Flexget,thalamus/Flexget,tarzasai/Flexget,tvcsantos/Flexget,tarzasai/Flexget,Danfocus/Flexget,drwyrm/Flexget,xfouloux/Flexget,xfouloux/Flexget,tsnoam/Flexget,Pretagonist/Flexget,crawln45/Flexget,tobinjt/Flexget,tsnoam/Flexget,tarzasai/Flexget,antivirtel/Flexget,asm0dey/Flexget,patsissons/Flexget,lildadou/Flexget,LynxyssCZ/Flexget,thalamus/Flexget,jacobmetrick/Flexget,jawilson/Flexget,lildadou/Flexget,X-dark/Flexget,Flexget/Flexget,xfouloux/Flexget,vfrc2/Flexget,v17al/Flexget,qvazzler/Flexget,oxc/Flexget,tobinjt/Flexget,Flexget/Flexget,oxc/Flexget,malkavi/Flexget,drwyrm/Flexget,tobinjt/Flexget,spencerjanssen/Flexget,patsissons/Flexget,grrr2/Flexget,antivirtel/Flexget,crawln45/Flexget,vfrc2/Flexget,voriux/Flexget,tvcsantos/Flexget,cvium/Flexget,spencerjanssen/Flexget,cvium/Flexget,Pretagonist/Flexget,sean797/Flexget,ZefQ/Flexget,dsemi/Flexget,thalamus/Flexget,poulpito/Flexget,spencerjanssen/Flexget,sean797/Flexget,camon/Flexget,patsissons/Flexget,jacobmetrick/Flexget,offbyone/Flexget,offbyone/Flexget,voriux/Flexget,LynxyssCZ/Flexget,Pretagonist/Flexget,ibrahimkarahan/Flexget,JorisDeRieck/Flexget,ratoaq2/Flexget,ianstalk/Flexget,antivirtel/Flexget,Danfocus/Flexget,grrr2/Flexget,ZefQ/Flexget,ianstalk/Flexget,v17al/Flexget,Flexget/Flexget,lildadou/Flexget,ianstalk/Flexget,malkavi/Flexget,JorisDeRieck/Flexget,jawilson/Flexget,OmgOhnoes/Flexget,Danfocus/Flexget,OmgOhnoes/Flexget,qk4l/Flexget,Flexget/Flexget,qk4l/Flexget,qk4l/Flexget,JorisDeRieck/Flexget,ZefQ/Flexget,crawln45/Flexget,qvazzler/Flexget,asm0dey/Flexget,ratoaq2/Flexget,dsemi/Flexget,camon/Flexget,crawln45/Flexget,dsemi/Flexget,X-dark/Flexget,gazpachoking/Flexget,tobinjt/Flexget,cvium/Flexget,gazpachoking/Flexget,vfrc2/Flexget,JorisDeRieck/Flexget,malkavi/Flexget,jacobmetrick/Flexget,LynxyssCZ/Flexget,qvazzler/Flexget,Danfocus/Flexget,poulpito/Flexget,jawilson/Flexget,X-dark/Flexget,LynxyssCZ/Flexget,poulpito/Flexget,grrr2/Flexget,v17al/Flexget
9346b34c68fc08dfba0002e907d73829000068cd
labmanager/shell.py
labmanager/shell.py
import cmd class LMShell(cmd.Cmd): def __init__(self, lmapi, completekey='tab', stdin=None, stdout=None): cmd.Cmd.__init__(self, completekey, stdin, stdout) self._lmapi = lmapi def do_list(self, line): configs = self._lmapi.list_library_configurations() print configs def do_EOF(self, line): return True def main(): from labmanager import api from labmanager import config import argparse import getpass parser = argparse.ArgumentParser() parser.add_argument('--hostname') parser.add_argument('--username') parser.add_argument('--organization') parser.add_argument('--workspace', default='Main') parser.add_argument('--timeout', default=None) parser.add_argument('--section', default='default') args = parser.parse_args() api_config = config.load_config(parser, args) if api_config.password is None: api_config.password = getpass.getpass('password: ') client = api.create_soap_client(api_config) labmanager_api = api.LabManager(client) sh = LMShell(labmanager_api) sh.cmdloop()
import cmd class LMShell(cmd.Cmd): def __init__(self, lmapi, completekey='tab', stdin=None, stdout=None): cmd.Cmd.__init__(self, completekey, stdin, stdout) self._lmapi = lmapi def do_list(self, line): configs = self._lmapi.list_library_configurations() print configs def do_EOF(self, line): return True def do_quit(self, line): return True def main(): from labmanager import api from labmanager import config import argparse import getpass parser = argparse.ArgumentParser() parser.add_argument('--hostname') parser.add_argument('--username') parser.add_argument('--organization') parser.add_argument('--workspace', default='Main') parser.add_argument('--timeout', default=None) parser.add_argument('--section', default='default') args = parser.parse_args() api_config = config.load_config(parser, args) if api_config.password is None: api_config.password = getpass.getpass('password: ') client = api.create_soap_client(api_config) labmanager_api = api.LabManager(client) sh = LMShell(labmanager_api) sh.cmdloop()
Add 'quit' command to lmsh
Add 'quit' command to lmsh
Python
bsd-3-clause
jamesls/labmanager-shell
713fcc3f86b4be4d35f0c5ba081a4f786648320a
vim/pythonx/elixir_helpers.py
vim/pythonx/elixir_helpers.py
""" Elixir-related Ultisnips snippet helper functions. NOTE: Changes to this file require restarting Vim! """ import re _DASHES_AND_UNDERSCORES = re.compile("[-_]") _MODULE_FILEPATH = re.compile(r"lib\/([^\/]+)\/([\w+\/]+)*\/([^\/]+).ex") def closing_character(tabstop): """ Return closing character for a tabstop containing an opening character. """ if tabstop.startswith("("): return ")" if tabstop.startswith("{"): return "}" if tabstop.startswith("["): return "]" if tabstop.startswith("\""): return "\"" return "" def module_path_match(path, regex=_MODULE_FILEPATH): """ Return match data for an Elixir module from a file path. """ return re.search(regex, path) def outer_module_name(path): """ Return name for an outer Elixir module from a file path. """ outer_module_path = module_path_match(path).group(1) return to_module_name(outer_module_path) def to_module_name(string): """ Convert string into an Elixir module name """ return ( re.sub(_DASHES_AND_UNDERSCORES, " ", string) .title() .replace(" ", "") .replace(".ex", "") )
""" Elixir-related Ultisnips snippet helper functions. NOTE: Changes to this file require restarting Vim! """ import re _DASHES_AND_UNDERSCORES = re.compile("[-_]") _MODULE_FILEPATH = re.compile(r"lib\/([^\/]+)\/([\w+\/]+)*\/([^\/]+).ex") _CLOSING_CHARACTERS = { "(": ")", "{": "}", "[": "]", "\"": "\"" } def closing_character(tabstop): """ Return closing character for a tabstop containing an opening character. """ if tabstop: return _CLOSING_CHARACTERS.get(tabstop[0], "") return "" def module_path_match(path, regex=_MODULE_FILEPATH): """ Return match data for an Elixir module from a file path. """ return re.search(regex, path) def outer_module_name(path): """ Return name for an outer Elixir module from a file path. """ outer_module_path = module_path_match(path).group(1) return to_module_name(outer_module_path) def to_module_name(string): """ Convert string into an Elixir module name """ return ( re.sub(_DASHES_AND_UNDERSCORES, " ", string) .title() .replace(" ", "") .replace(".ex", "") )
Refactor python if statement into dictionary
Refactor python if statement into dictionary
Python
mit
paulfioravanti/dotfiles,paulfioravanti/dotfiles,paulfioravanti/dotfiles
f1b0b1fc96802bf236cf9bfbc894ebdf47704b33
test_example.py
test_example.py
"""Usage: prog [-vqr] [FILE] prog INPUT OUTPUT prog --help Options: -v print status messages -q report only file names -r show all occurrences of the same error --help """ from docopt import docopt, Options, Arguments, DocoptExit from pytest import raises def test_docopt(): o, a = docopt(__doc__, '-v file.py') assert o == Options(v=True, q=False, r=False, help=False) assert a == Arguments(file='file.py', input=None, output=None) o, a = docopt(__doc__, '-v') assert o == Options(v=True, q=False, r=False, help=False) assert a == Arguments(file=None, input=None, output=None) with raises(DocoptExit): # does not match docopt(__doc__, '-v input.py output.py') with raises(DocoptExit): docopt(__doc__, '--fake') with raises(DocoptExit): docopt(__doc__, '--hel')
"""Usage: prog [-vqr] [FILE] prog INPUT OUTPUT prog --help Options: -v print status messages -q report only file names -r show all occurrences of the same error --help """ from docopt import docopt, Options, Arguments, DocoptExit from pytest import raises def test_docopt(): o, a = docopt(__doc__, '-v file.py') assert o == Options(v=True, q=False, r=False, help=False) assert a == Arguments(file='file.py', input=None, output=None) o, a = docopt(__doc__, '-v') assert o == Options(v=True, q=False, r=False, help=False) assert a == Arguments(file=None, input=None, output=None) with raises(DocoptExit): # does not match docopt(__doc__, '-v input.py output.py') with raises(DocoptExit): docopt(__doc__, '--fake') with raises(SystemExit): docopt(__doc__, '--hel')
Correct test to catch SystemExit on normal exit.
Correct test to catch SystemExit on normal exit.
Python
mit
docopt/docopt,jagguli/docopt,benthomasson/docopt,kenwilcox/docopt,devonjones/docopt,snowsnail/docopt,Zearin/docopt,wkentaro/docopt,crcsmnky/docopt
77744d61918510fcd943d9420ce4c61717a0711b
test/functionalities/connect_remote/TestConnectRemote.py
test/functionalities/connect_remote/TestConnectRemote.py
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = TestBase.compute_mydir(__file__) def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... if sys.platform.startswith('freebsd') or sys.platform.startswith("linux"): self.runCmd("process connect -p gdb-remote connect://localhost:12345") else: self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = TestBase.compute_mydir(__file__) @expectedFailureFreeBSD('llvm.org/pr18313') def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... if sys.platform.startswith('freebsd') or sys.platform.startswith("linux"): self.runCmd("process connect -p gdb-remote connect://localhost:12345") else: self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
Add decorator for GDB connect test failing on FreeBSD
Add decorator for GDB connect test failing on FreeBSD llvm.org/pr18313 git-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@197910 91177308-0d34-0410-b5e6-96231b3b80d8
Python
apache-2.0
llvm-mirror/lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb
4e3f10cc417f28badc34646cc89fcd9d0307b4be
utility/lambdas/s3-static-site-deploy/lambda_function.py
utility/lambdas/s3-static-site-deploy/lambda_function.py
# import boto3 def lambda_handler(event, context): pass
# Invoked by: CloudFormation # Returns: A `Data` object to a pre-signed URL # # Deploys the contents of a versioned zip file object from one bucket in S3 # to a another bucket import sys import boto3 from botocore.client import Config import io import zipfile import os import urllib.request import json import traceback s3 = boto3.client('s3', config=Config(signature_version='s3v4')) STATUS_SUCCESS = 'SUCCESS' STATUS_FAILED = 'FAILED' def send_response(event, context, res_status, res_reason='Done', res_data={}): res_data = json.dumps({ 'Status': res_status, 'Reason': res_reason, 'PhysicalResourceId': context.log_stream_name, 'StackId': event['StackId'], 'RequestId': event['RequestId'], 'LogicalResourceId': event['LogicalResourceId'], 'Data': res_data }).encode() headers = {'content-type': ''} url = event['ResponseURL'] req = urllib.request.Request(url, data=res_data, method='PUT', headers=headers) urllib.request.urlopen(req) def lambda_handler(event, context): try: print(event) if event['RequestType'] == 'Create' or event['RequestType'] == 'Update': # The location of the built static site archive file in S3 bucket = event['ResourceProperties']['StaticSiteArchiveS3Bucket'] key = event['ResourceProperties']['StaticSiteArchiveS3Object'] version = event['ResourceProperties']['StaticSiteArchiveS3ObjectVersion'] # Get the archive object s3_obj = s3.get_object(Bucket=bucket, Key=key, VersionId=version) unzip_dir = f"/tmp/unzip-{event['RequestId']}" # Unzip the archive, to disk with zipfile.ZipFile(io.BytesIO(s3_obj['Body'].read()), 'r') as zip: zip.extractall(unzip_dir) # The bucket to deploy the static to deploy_bucket = event['ResourceProperties']['StaticSiteS3DeployBucket'] # Upload everything from the unzipped archive for root, dirs, files in os.walk(unzip_dir): for filename in files: local_path = os.path.join(root, filename) s3_key = os.path.relpath(local_path, unzip_dir) print(f"Uploading {s3_key} to {deploy_bucket}") s3.upload_file(local_path, deploy_bucket, s3_key) send_response(event, context, STATUS_SUCCESS) else: send_response(event, context, STATUS_SUCCESS) except Exception as e: print('Function failed due to exception.') print(e) traceback.print_exc() send_response(event, context, STATUS_FAILED, res_reason=str(e))
Add S3 static deploy custom resource Lambda function
Add S3 static deploy custom resource Lambda function
Python
mit
PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure
aed82bc0995cf4175c0ab8c521dfc8e89d776a7e
Mac/scripts/zappycfiles.py
Mac/scripts/zappycfiles.py
# Zap .pyc files import os import sys doit = 1 def main(): if os.name == 'mac': import macfs fss, ok = macfs.GetDirectory('Directory to zap pyc files in') if not ok: sys.exit(0) dir = fss.as_pathname() zappyc(dir) else: if not sys.argv[1:]: print 'Usage: zappyc dir ...' sys.exit(1) for dir in sys.argv[1:]: zappyc(dir) def zappyc(dir): os.path.walk(dir, walker, None) def walker(dummy, top, names): for name in names: if name[-4:] == '.pyc': path = os.path.join(top, name) print 'Zapping', path if doit: os.unlink(path) if __name__ == '__main__': main()
#!/usr/local/bin/python """Recursively zap all .pyc files""" import os import sys # set doit true to actually delete files # set doit false to just print what would be deleted doit = 1 def main(): if not sys.argv[1:]: if os.name == 'mac': import macfs fss, ok = macfs.GetDirectory('Directory to zap pyc files in') if not ok: sys.exit(0) dir = fss.as_pathname() zappyc(dir) else: print 'Usage: zappyc dir ...' sys.exit(1) for dir in sys.argv[1:]: zappyc(dir) def zappyc(dir): os.path.walk(dir, walker, None) def walker(dummy, top, names): for name in names: if name[-4:] == '.pyc': path = os.path.join(top, name) print 'Zapping', path if doit: os.unlink(path) if __name__ == '__main__': main()
Patch by Russel Owen: if we have command line arguments zap pyc files in the directories given.
Patch by Russel Owen: if we have command line arguments zap pyc files in the directories given.
Python
mit
sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator
db04d6884c68b1f673a785866155427af86fad65
apps/predict/templatetags/jsonify.py
apps/predict/templatetags/jsonify.py
"""Add a template tag to turn python objects into JSON""" import types import json from django import template from django.utils.safestring import mark_safe register = template.Library() @register.filter def jsonify(obj): if isinstance(obj, types.GeneratorType): obj = list(obj) return mark_safe(json.dumps(obj))
"""Add a template tag to turn python objects into JSON""" import types import json from django import template from django.utils.safestring import mark_safe register = template.Library() @register.filter def jsonify(obj): """Turn object into a json instance""" if isinstance(obj, types.GeneratorType): obj = list(obj) return mark_safe(json.dumps(obj).replace("'", "\\'"))
Remove single quote marks from jsonif
Remove single quote marks from jsonif
Python
agpl-3.0
IQSS/gentb-site,IQSS/gentb-site,IQSS/gentb-site,IQSS/gentb-site,IQSS/gentb-site,IQSS/gentb-site,IQSS/gentb-site,IQSS/gentb-site
73d4aa7fa41117bbbe6447466cf453153f76b5ba
armstrong/core/arm_wells/views.py
armstrong/core/arm_wells/views.py
from django.core.exceptions import ImproperlyConfigured from django.views.generic import TemplateView from django.views.generic.list import MultipleObjectMixin from django.utils.translation import ugettext as _ from .models import Well class SimpleWellView(TemplateView): allow_empty = False well_title = None def __init__(self, *args, **kwargs): super(SimpleWellView, self).__init__(*args, **kwargs) if not self.well_title: raise ImproperlyConfigured( _(u"Expects a `well_title` to be provided")) def get_well(self): try: return Well.objects.get_current(title=self.well_title) except Well.DoesNotExist: if self.allow_empty: return False raise def get_context_data(self, **kwargs): context = super(SimpleWellView, self).get_context_data(**kwargs) context["well"] = self.get_well() return context class QuerySetBackedWellView(SimpleWellView, MultipleObjectMixin): def get_queryset(self): well = self.get_well() return (well.items if well is not False else super(QuerySetBackedWellView, self).get_queryset()) def get_well(self): well = super(QuerySetBackedWellView, self).get_well() if well: well.merge_with(super(QuerySetBackedWellView, self).get_queryset()) return well
from django.core.exceptions import ImproperlyConfigured from django.views.generic import TemplateView from django.views.generic.list import MultipleObjectMixin from django.utils.translation import ugettext as _ from .models import Well class SimpleWellView(TemplateView): allow_empty = False well_title = None def __init__(self, *args, **kwargs): super(SimpleWellView, self).__init__(*args, **kwargs) if not self.well_title: raise ImproperlyConfigured( _(u"Expects a `well_title` to be provided")) def get_well(self): try: return Well.objects.get_current(title=self.well_title) except Well.DoesNotExist: if self.allow_empty: return None raise def get_context_data(self, **kwargs): context = super(SimpleWellView, self).get_context_data(**kwargs) context["well"] = self.get_well() return context class QuerySetBackedWellView(SimpleWellView, MultipleObjectMixin): def get_queryset(self): well = self.get_well() return (well.items if well is not None else super(QuerySetBackedWellView, self).get_queryset()) def get_well(self): well = super(QuerySetBackedWellView, self).get_well() if well: well.merge_with(super(QuerySetBackedWellView, self).get_queryset()) return well
Switch to None based on feedback from @niran
Switch to None based on feedback from @niran
Python
apache-2.0
armstrong/armstrong.core.arm_wells,armstrong/armstrong.core.arm_wells,texastribune/armstrong.core.arm_wells,texastribune/armstrong.core.arm_wells,armstrong/armstrong.core.arm_wells
8ed2aa1a8108ae3a678ff18f4e8fda3539f4b603
avalonstar/components/games/admin.py
avalonstar/components/games/admin.py
# -*- coding: utf-8 -*- from django.contrib import admin from .models import Game, Platform class GameAdmin(admin.ModelAdmin): list_display = ['name', 'platform', 'gbid', 'is_abandoned', 'is_completed'] raw_id_fields = ['platform'] autocomplete_lookup_fields = { 'fk': ['platform'] } admin.site.register(Game, GameAdmin) class PlatformAdmin(admin.ModelAdmin): pass admin.site.register(Platform, PlatformAdmin)
# -*- coding: utf-8 -*- from django.contrib import admin from .models import Game, Platform class GameAdmin(admin.ModelAdmin): list_display = ['name', 'platform', 'gbid', 'is_abandoned', 'is_completed'] list_editable = ['is_abandoned', 'is_completed'] raw_id_fields = ['platform'] autocomplete_lookup_fields = { 'fk': ['platform'] } admin.site.register(Game, GameAdmin) class PlatformAdmin(admin.ModelAdmin): pass admin.site.register(Platform, PlatformAdmin)
Make the game booleans editable.
Make the game booleans editable.
Python
apache-2.0
bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv
12683ea64a875b624230f2dd84609a77eaec1095
cd_wizard.py
cd_wizard.py
#!/usr/bin/env python """Wizard to guide user to: - insert cd - please rip with eac - check for a good rip - upload with metadata (freedb, musicmind) """ from PyQt4 import QtGui def createIntroPage(): page = QtGui.QWizardPage() page.setTitle("Introduction") page.setSubTitle("This wizard will help you archive your CDs in your Personal Music Locker") label = QtGui.QLabel("Please insert a CD") label.setWordWrap(True) layout = QtGui.QVBoxLayout() layout.addWidget(label) page.setLayout(layout) return page def createConclusionPage(): page = QtGui.QWizardPage() page.setTitle("Conclusion") label = QtGui.QLabel("You are now added this CD to your locker!") label.setWordWrap(True) layout = QtGui.QVBoxLayout() layout.addWidget(label) page.setLayout(layout) return page if __name__ == '__main__': import sys app = QtGui.QApplication(sys.argv) wizard = QtGui.QWizard() wizard.addPage(createIntroPage()) wizard.addPage(createConclusionPage()) wizard.setWindowTitle("Music Locker Uploader") wizard.show() sys.exit(wizard.exec_())
#!/usr/bin/env python """Wizard to guide user to: - insert cd - please rip with eac - check for a good rip - upload with metadata (freedb, musicmind) """ from PyQt4 import QtGui def createIntroPage(): page = QtGui.QWizardPage() page.setTitle("Introduction") page.setSubTitle("This wizard will help you archive your CDs in your Personal Music Locker") label = QtGui.QLabel("Please insert a CD") label.setWordWrap(True) layout = QtGui.QVBoxLayout() layout.addWidget(label) page.setLayout(layout) return page def choose_cd(): page = QtGui.QWizardPage() page.setTitle("Choose CD Drive") file_dialog = QtGui.QFileDialog() file_dialog.setFileMode(QtGui.QFileDialog.Directory) file_dialog.setOptions(QtGui.QFileDialog.ShowDirsOnly) file_dialog.setDirectory('/') layout = QtGui.QVBoxLayout() layout.addWidget(file_dialog) page.setLayout(layout) return page def createConclusionPage(): page = QtGui.QWizardPage() page.setTitle("Conclusion") label = QtGui.QLabel("You are now added this CD to your locker!") label.setWordWrap(True) layout = QtGui.QVBoxLayout() layout.addWidget(label) page.setLayout(layout) return page if __name__ == '__main__': import sys app = QtGui.QApplication(sys.argv) wizard = QtGui.QWizard() wizard.addPage(createIntroPage()) wizard.addPage(choose_cd()) wizard.addPage(createConclusionPage()) wizard.setWindowTitle("Music Locker Uploader") wizard.show() sys.exit(wizard.exec_())
Add file browser to choose a CD.
Add file browser to choose a CD.
Python
agpl-3.0
brewsterkahle/archivecd
e7cb5b0be49bc5e811809c56eb4ad3c0dc861cdf
examples/child_watcher.py
examples/child_watcher.py
import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() log.info("There are %d items now: %s", len(children), children) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item)
import logging import random from tornado import gen from zoonado import exc log = logging.getLogger() def arguments(parser): parser.add_argument( "--path", "-p", type=str, default="/examplewatcher", help="ZNode path to use for the example." ) def watcher_callback(children): children.sort() log.info("There are %d items now: %s", len(children), ", ".join(children)) @gen.coroutine def run(client, args): yield client.start() try: yield client.create(args.path) except exc.NodeExists: pass watcher = client.recipes.ChildrenWatcher() watcher.add_callback(args.path, watcher_callback) to_make = ["cat", "dog", "mouse", "human"] random.shuffle(to_make) for item in to_make: yield client.create(args.path + "/" + item, ephemeral=True) yield gen.sleep(1) for item in to_make: yield client.delete(args.path + "/" + item) yield gen.sleep(1)
Fix up to the child watcher example.
Fix up to the child watcher example. Without yielding to the ioloop after each call to client.delete() the child znodes would be deleted but that would never be reported.
Python
apache-2.0
wglass/zoonado
615e57fefa2b3b52ce351ef1d8039216927dc891
Parallel/Testing/Cxx/TestSockets.py
Parallel/Testing/Cxx/TestSockets.py
""" Driver script for testing sockets Unix only """ import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15) sys.exit(os.WEXITSTATUS(retVal))
""" Driver script for testing sockets Unix only """ import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client retVal = os.system('"%s" -D "%s" -V "%s"' % ( sys.argv[2], sys.argv[3], sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) try: os.kill(pid, 15) except: pass sys.exit(os.WEXITSTATUS(retVal))
Fix space problem and put try around os.kill
ERR: Fix space problem and put try around os.kill
Python
bsd-3-clause
SimVascular/VTK,johnkit/vtk-dev,gram526/VTK,daviddoria/PointGraphsPhase1,sankhesh/VTK,hendradarwin/VTK,jeffbaumes/jeffbaumes-vtk,ashray/VTK-EVM,msmolens/VTK,SimVascular/VTK,arnaudgelas/VTK,gram526/VTK,berendkleinhaneveld/VTK,johnkit/vtk-dev,Wuteyan/VTK,aashish24/VTK-old,mspark93/VTK,hendradarwin/VTK,collects/VTK,collects/VTK,sumedhasingla/VTK,ashray/VTK-EVM,SimVascular/VTK,biddisco/VTK,cjh1/VTK,keithroe/vtkoptix,spthaolt/VTK,candy7393/VTK,candy7393/VTK,naucoin/VTKSlicerWidgets,sumedhasingla/VTK,spthaolt/VTK,berendkleinhaneveld/VTK,demarle/VTK,naucoin/VTKSlicerWidgets,sgh/vtk,gram526/VTK,msmolens/VTK,jmerkow/VTK,aashish24/VTK-old,biddisco/VTK,berendkleinhaneveld/VTK,msmolens/VTK,candy7393/VTK,keithroe/vtkoptix,daviddoria/PointGraphsPhase1,mspark93/VTK,demarle/VTK,jmerkow/VTK,spthaolt/VTK,gram526/VTK,johnkit/vtk-dev,jeffbaumes/jeffbaumes-vtk,mspark93/VTK,sankhesh/VTK,johnkit/vtk-dev,demarle/VTK,SimVascular/VTK,jmerkow/VTK,biddisco/VTK,ashray/VTK-EVM,hendradarwin/VTK,jmerkow/VTK,keithroe/vtkoptix,gram526/VTK,candy7393/VTK,berendkleinhaneveld/VTK,arnaudgelas/VTK,sankhesh/VTK,aashish24/VTK-old,aashish24/VTK-old,arnaudgelas/VTK,ashray/VTK-EVM,sumedhasingla/VTK,hendradarwin/VTK,daviddoria/PointGraphsPhase1,sgh/vtk,candy7393/VTK,sankhesh/VTK,johnkit/vtk-dev,demarle/VTK,Wuteyan/VTK,johnkit/vtk-dev,aashish24/VTK-old,demarle/VTK,mspark93/VTK,sankhesh/VTK,msmolens/VTK,hendradarwin/VTK,keithroe/vtkoptix,collects/VTK,candy7393/VTK,SimVascular/VTK,gram526/VTK,mspark93/VTK,sumedhasingla/VTK,mspark93/VTK,sankhesh/VTK,naucoin/VTKSlicerWidgets,mspark93/VTK,cjh1/VTK,msmolens/VTK,jeffbaumes/jeffbaumes-vtk,collects/VTK,jmerkow/VTK,naucoin/VTKSlicerWidgets,daviddoria/PointGraphsPhase1,biddisco/VTK,berendkleinhaneveld/VTK,johnkit/vtk-dev,candy7393/VTK,sumedhasingla/VTK,Wuteyan/VTK,biddisco/VTK,Wuteyan/VTK,demarle/VTK,candy7393/VTK,keithroe/vtkoptix,keithroe/vtkoptix,sgh/vtk,sumedhasingla/VTK,sankhesh/VTK,sumedhasingla/VTK,collects/VTK,arnaudgelas/VTK,Wuteyan/VTK,naucoin/VTKSlicerWidgets,msmolens/VTK,daviddoria/PointGraphsPhase1,demarle/VTK,demarle/VTK,arnaudgelas/VTK,jmerkow/VTK,gram526/VTK,ashray/VTK-EVM,ashray/VTK-EVM,jeffbaumes/jeffbaumes-vtk,jeffbaumes/jeffbaumes-vtk,SimVascular/VTK,msmolens/VTK,Wuteyan/VTK,biddisco/VTK,keithroe/vtkoptix,sumedhasingla/VTK,berendkleinhaneveld/VTK,hendradarwin/VTK,collects/VTK,spthaolt/VTK,sankhesh/VTK,cjh1/VTK,cjh1/VTK,cjh1/VTK,SimVascular/VTK,jeffbaumes/jeffbaumes-vtk,ashray/VTK-EVM,sgh/vtk,ashray/VTK-EVM,Wuteyan/VTK,gram526/VTK,biddisco/VTK,hendradarwin/VTK,berendkleinhaneveld/VTK,msmolens/VTK,sgh/vtk,daviddoria/PointGraphsPhase1,mspark93/VTK,keithroe/vtkoptix,sgh/vtk,jmerkow/VTK,cjh1/VTK,arnaudgelas/VTK,spthaolt/VTK,spthaolt/VTK,jmerkow/VTK,naucoin/VTKSlicerWidgets,SimVascular/VTK,spthaolt/VTK,aashish24/VTK-old
e8d57ef08616b06e5f94da7e01ba96c13b9124d7
perfrunner/celeryremote.py
perfrunner/celeryremote.py
BROKER_URL = 'amqp://couchbase:[email protected]:5672/broker' CELERY_RESULT_BACKEND = 'amqp' CELERY_RESULT_EXCHANGE = 'perf_results' CELERY_RESULT_PERSISTENT = False
BROKER_URL = 'amqp://couchbase:[email protected]:5672/broker' CELERY_RESULT_BACKEND = 'amqp' CELERY_RESULT_EXCHANGE = 'perf_results' CELERY_RESULT_PERSISTENT = False CELERYD_HIJACK_ROOT_LOGGER = False
Disable hijacking of previously configured log handlers
Disable hijacking of previously configured log handlers See also: http://docs.celeryproject.org/en/3.1/configuration.html#celeryd-hijack-root-logger Change-Id: Ibf4618e8bfeb28f877db4a40b4a911ff00442cc9 Reviewed-on: http://review.couchbase.org/82543 Tested-by: Build Bot <[email protected]> Reviewed-by: Pavel Paulau <[email protected]>
Python
apache-2.0
couchbase/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner,couchbase/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner,pavel-paulau/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner
dc883b81a2c5714d9401fb113101639e13e396f5
integration_tests/tests/hello_world_sleep_and_time.py
integration_tests/tests/hello_world_sleep_and_time.py
integration_test = True timeout = 2 SLEEP_INTERVAL = int(100e6) def check_state(state): import re from functools import partial from operator import is_not r = re.compile('^(\d+) \[.*\] Hello World!') lines = map(r.match, state.console.split('\n')) lines = filter(partial(is_not, None), lines) times = map(lambda m: int(m.group(1)), lines) times = list(times) min_times = (timeout - 1) * int(1e9) // SLEEP_INTERVAL assert len(times) >= min_times, "Expected at least {0} hello worlds".format(min_times) prev = 0 for t in times: diff = t - prev assert diff >= SLEEP_INTERVAL, "Sleep interval must be >= {0}".format(SLEEP_INTERVAL) prev = diff
integration_test = True timeout = 2 SLEEP_INTERVAL = int(100e6) MIN_TIME = 1451606400000000000 # 2016-1-1 0:0:0.0 UTC def check_state(state): import re from functools import partial from operator import is_not r = re.compile('^(\d+) \[.*\] Hello World!') lines = map(r.match, state.console.split('\n')) lines = filter(partial(is_not, None), lines) times = map(lambda m: int(m.group(1)), lines) times = list(times) min_times = (timeout - 1) * int(1e9) // SLEEP_INTERVAL assert len(times) >= min_times, "Expected at least {0} hello worlds".format(min_times) prev = 0 for t in times: diff = t - prev assert diff >= SLEEP_INTERVAL, "Sleep interval must be >= {0}".format(SLEEP_INTERVAL) assert t >= MIN_TIME, "Time must be after {0}".format(MIN_TIME) prev = diff
Make sure current date is late enough
Make sure current date is late enough
Python
bsd-2-clause
unigornel/unigornel,unigornel/unigornel
c7e9b65d7951b9757f907da9e4bf35e43dbdbd88
django_env/bin/install.py
django_env/bin/install.py
#!/usr/bin/env python import sys, os def main(): django_env_dir = os.path.abspath('%s/../' % os.path.dirname(__file__)) workon_home = os.environ.get('WORKON_HOME') if not workon_home: print "ERROR: The $WORKON_HOME environment variable is not set. Please check to make sure you've installed and setup virtualenvwrapper correctly." sys.exit() # symlink the django_env directory inside the $WORKON_HOME command = 'ln -sf %s "$WORKON_HOME/django_env"' % django_env_dir os.system(command) # add the ejango_env postmkvirtualenv hook to the virtualenvwrapper postmkvirtualenv hook postmkvirtualenv_cmd = 'source $WORKON_HOME/django_env/bin/postmkvirtualenv' workon_home = os.getenv('WORKON_HOME') postmkvirtualenv_path = os.path.join(workon_home, 'postmkvirtualenv') fh = open(postmkvirtualenv_path, "r") contents = fh.read() fh.close() if contents.find(postmkvirtualenv_cmd) == -1: fh = open(postmkvirtualenv_path, "a") fh.write("\n\n%s\n\n" % postmkvirtualenv_cmd) fh.close() print """ Django-environment is now installed. To create a django-environment run the following. mkvirtualenv [project_name] Example: mkvirtualenv example """ if __name__ == '__main__': main()
#!/usr/bin/env python import sys import os def main(): django_env_dir = os.path.abspath('%s/../' % os.path.dirname(__file__)) workon_home = os.environ.get('WORKON_HOME') if not workon_home: print "ERROR: The $WORKON_HOME environment variable is not set. Please check to make sure you've installed and setup virtualenvwrapper correctly." sys.exit() # symlink the django_env directory inside the $WORKON_HOME command = 'ln -sf %s "$WORKON_HOME/django_env"' % django_env_dir os.system(command) # add the ejango_env postmkvirtualenv hook to the virtualenvwrapper postmkvirtualenv hook postmkvirtualenv_cmd = 'source $WORKON_HOME/django_env/bin/postmkvirtualenv' workon_home = os.getenv('WORKON_HOME') postmkvirtualenv_path = os.path.join(workon_home, 'postmkvirtualenv') fh = open(postmkvirtualenv_path, "r") contents = fh.read() fh.close() if contents.find(postmkvirtualenv_cmd) == -1: fh = open(postmkvirtualenv_path, "a") fh.write("\n\n%s\n\n" % postmkvirtualenv_cmd) fh.close() print """ Django-environment is now installed. To create a django-environment run the following. mkvirtualenv [project_name] Example: mkvirtualenv example """ if __name__ == '__main__': main()
Switch the os import to it's own line for pep8 compliance.
Switch the os import to it's own line for pep8 compliance.
Python
bsd-3-clause
epicserve/django-environment,epicserve/django-environment
f3fb5bd0dbb3e19e58558af015aaee5ec120af71
portal/template_helpers.py
portal/template_helpers.py
""" Module for helper functions used inside jinja2 templates """ # NB, each blueprint must individually load any functions defined below # for them to appear in the namespace when invoked from respective blueprint # See @<blueprint>.context_processor decorator for more info. def split_string(s, delimiter=','): return s.split(delimiter)
""" Module for helper functions used inside jinja2 templates """ # NB, each blueprint must individually load any functions defined below # for them to appear in the namespace when invoked from respective blueprint # See @<blueprint>.context_processor decorator for more info. def split_string(s, delimiter=','): """Given string (or tuple) return the delimited values""" # If given a tuple, split already happened if isinstance(s, (list, tuple)): return s return s.split(delimiter)
Allow for list/tuples in config files when looking for comma delimited strings.
Allow for list/tuples in config files when looking for comma delimited strings.
Python
bsd-3-clause
uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal,uwcirg/true_nth_usa_portal
674f6e0b9fbb76684a9b05d16a5da0d4cc732b1d
scripts/analysis/plot_tracking_vector_estimator_stats.py
scripts/analysis/plot_tracking_vector_estimator_stats.py
#!/usr/bin/env python2 import numpy as np import matplotlib.pyplot as plt import argparse import sys import os parser = argparse.ArgumentParser( prog='plot_tracking_vector_estimator') parser.add_argument('directory', type=str, help='Data directory') args = parser.parse_args() data = np.genfromtxt( os.path.join( args.directory, 'tracking_vector_estimator'), delimiter=',', names=True) state_labels = ['Marker_x', 'Marker_y', 'Marker_z', 'Velocity_x', 'Velocity_y', 'Velocity_z'] noise_labels = ['Noise_x', 'Noise_y', 'Noise_z', 'Noise_vx', 'Noise_vy', 'Noise_vz'] meas_labels = ['Measured_Marker_x', 'Measured_Marker_y', 'Measured_Marker_y', 'Measured_Velocity_x', 'Measured_Velocity_y', 'Measured_Velocity_z'] ts = (data['Time'] - data['Time'][0]) / 1e9 plt.figure(1) for i in range(6): plt.subplot(2, 3, i+1) plt.plot(ts, data[meas_labels[i]]) plt.errorbar(ts, data[state_labels[i]], yerr=data[noise_labels[i]]) plt.ylabel(state_labels[i]) plt.xlabel('Time (seconds)') plt.legend([meas_labels[i], state_labels[i]]) plt.show()
#!/usr/bin/env python2 import numpy as np import matplotlib.pyplot as plt import argparse import sys import os parser = argparse.ArgumentParser( prog='plot_tracking_vector_estimator') parser.add_argument('directory', type=str, help='Data directory') args = parser.parse_args() data = np.genfromtxt( os.path.join( args.directory, 'tracking_vector_estimator'), delimiter=',', names=True) state_labels = ['Marker_x', 'Marker_y', 'Marker_z'] noise_labels = ['Noise_x', 'Noise_y', 'Noise_z'] meas_labels = ['Measured_Marker_x', 'Measured_Marker_y', 'Measured_Marker_y'] meas_noise_labels = ['Meas_noise_x', 'Meas_noise_y', 'Meas_noise_z'] ts = (data['Time'] - data['Time'][0]) / 1e9 plt.figure(1) for i in range(3): plt.subplot(2, 3, i+1) plt.errorbar(ts, data[meas_labels[i]], yerr=data[meas_noise_labels[i]]) plt.errorbar(ts, data[state_labels[i]], yerr=data[noise_labels[i]]) plt.ylabel(state_labels[i]) plt.xlabel('Time (seconds)') plt.legend([meas_labels[i], state_labels[i]]) plt.show()
Change estimator script based on modifications to estimator
Change estimator script based on modifications to estimator
Python
mpl-2.0
jhu-asco/aerial_autonomy,jhu-asco/aerial_autonomy,jhu-asco/aerial_autonomy,jhu-asco/aerial_autonomy,jhu-asco/aerial_autonomy,jhu-asco/aerial_autonomy
a0aa74d9e6295e34f02b4eefd76e7eb9a1e6425f
node/floor_divide.py
node/floor_divide.py
#!/usr/bin/env python from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))]
#!/usr/bin/env python from nodes import Node class FloorDiv(Node): char = "f" args = 2 results = 1 @Node.test_func([3,2], [1]) @Node.test_func([6,-3], [-2]) def func(self, a:Node.number,b:Node.number): """a/b. Rounds down, returns an int.""" return a//b @Node.test_func(["test", "e"], [["t", "e", "st"]]) def partition(self, string:str, sep:str): """Split the string at the first occurrence of sep, return a 3-list containing the part before the separator, the separator itself, and the part after the separator. If the separator is not found, return a 3-list containing the string itself, followed by two empty strings.""" return [list(string.partition(sep))] @Node.test_func(["134", 1], [["134"]]) @Node.test_func(["1234", 2], [["12", "34"]]) @Node.test_func(["1234", 3], [["1", "2", "34"]]) @Node.test_func([[4,8,15,16,23,42], 5], [[[4],[8],[15],[16],[23,42]]]) def chunk(self, inp:Node.indexable, num:Node.number): """Return inp seperated into num groups""" rtn = [] last = 0 size = len(inp)//num for i in range(size, len(inp), size): rtn.append(inp[last:i]) last = i if len(rtn) != num: rtn.append(inp[last:]) else: rtn[-1] += inp[last:] if len(rtn): if isinstance(inp, str): rtn[-1] = "".join(rtn[-1]) else: rtn[-1] = type(inp)(rtn[-1]) return [rtn]
Add a group chunk, chunks a list into N groups
Add a group chunk, chunks a list into N groups
Python
mit
muddyfish/PYKE,muddyfish/PYKE
9361af556cfa7f4fb6bb3c53b4e74e2c115cd7d7
annict/client.py
annict/client.py
# -*- coding: utf-8 -*- from operator import methodcaller import requests from furl import furl class Client(object): def __init__(self, access_token, base_url='https://api.annict.com', api_version='v1'): self.access_token = access_token self.base_url = base_url self.api_version = api_version def _request(self, http_method, path, kwargs=None): kwargs['access_token'] = self.access_token d = {} if http_method == 'post' or http_method == 'patch': d['data'] = kwargs elif http_method == 'get': d['params'] = kwargs url = furl(self.base_url) url.path.add(self.api_version).add(path) m = methodcaller(http_method, url.url, **d) response = m(requests) if not response.content: return None return response.json() def get(self, path, kwargs): return self._request('get', path, kwargs) def post(self, path, kwargs): return self._request('post', path, kwargs) def patch(self, path, kwargs): return self._request('patch', path, kwargs) def delete(self, path): return self._request('delete', path)
# -*- coding: utf-8 -*- from operator import methodcaller import requests from furl import furl class Client(object): def __init__(self, access_token, base_url='https://api.annict.com', api_version='v1'): self.access_token = access_token self.base_url = base_url self.api_version = api_version def _request(self, http_method, path, kwargs=None): kwargs['access_token'] = self.access_token d = {} if http_method == 'post' or http_method == 'patch': d['data'] = kwargs elif http_method == 'get': d['params'] = kwargs url = furl(self.base_url) url.path.add(self.api_version).add(path) m = methodcaller(http_method, url.url, **d) return m(requests) def get(self, path, kwargs): return self._request('get', path, kwargs) def post(self, path, kwargs): return self._request('post', path, kwargs) def patch(self, path, kwargs): return self._request('patch', path, kwargs) def delete(self, path): return self._request('delete', path)
Fix Client returns requests's response.
Fix Client returns requests's response.
Python
mit
kk6/python-annict
069a29351e228996a465b962b1dffed5581685de
src/gewebehaken/cli.py
src/gewebehaken/cli.py
""" Gewebehaken ~~~~~~~~~~~ Command-line interface :Copyright: 2015-2021 Jochen Kupperschmidt :License: MIT, see LICENSE for details. """ from argparse import ArgumentParser from .app import create_app DEFAULT_HOST = '127.0.0.1' DEFAULT_PORT = 5000 DEFAULT_LOG_FILENAME = 'incoming.log' def parse_args(): """Setup and apply the command line arguments parser.""" parser = ArgumentParser() parser.add_argument( '--debug', dest='debug', action='store_true', default=False, help='debug mode') parser.add_argument( '--host', dest='host', default=DEFAULT_HOST, help='the host to listen on [default: {}]'.format(DEFAULT_HOST), metavar='HOST') parser.add_argument( '--port', dest='port', type=int, default=DEFAULT_PORT, help='the port to listen on [default: {:d}]'.format(DEFAULT_PORT), metavar='PORT') return parser.parse_args() def main(): args = parse_args() app = create_app(log_filename=DEFAULT_LOG_FILENAME) app.run(host=args.host, port=args.port, debug=args.debug)
""" Gewebehaken ~~~~~~~~~~~ Command-line interface :Copyright: 2015-2021 Jochen Kupperschmidt :License: MIT, see LICENSE for details. """ from argparse import ArgumentParser from .app import create_app DEFAULT_HOST = '127.0.0.1' DEFAULT_PORT = 5000 def parse_args(): """Setup and apply the command line arguments parser.""" parser = ArgumentParser() parser.add_argument( '--debug', dest='debug', action='store_true', default=False, help='debug mode') parser.add_argument( '--host', dest='host', default=DEFAULT_HOST, help='the host to listen on [default: {}]'.format(DEFAULT_HOST), metavar='HOST') parser.add_argument( '--port', dest='port', type=int, default=DEFAULT_PORT, help='the port to listen on [default: {:d}]'.format(DEFAULT_PORT), metavar='PORT') parser.add_argument( '--logfile', dest='logfile', help='logfile to write incoming webhook requests to', metavar='LOGFILE') return parser.parse_args() def main(): args = parse_args() app = create_app(log_filename=args.logfile) app.run(host=args.host, port=args.port, debug=args.debug)
Make log filename configurable and optional
Make log filename configurable and optional
Python
mit
homeworkprod/gewebehaken
dffbc7d79c67c3629f718c7a0330f9922499640d
examples/translations/portuguese_test_1.py
examples/translations/portuguese_test_1.py
# Portuguese Language Test - Python 3 Only! from seleniumbase.translate.portuguese import CasoDeTeste class MinhaClasseDeTeste(CasoDeTeste): def test_exemplo_1(self): self.abrir_url("https://pt.wikipedia.org/wiki/") self.verificar_texto("Wikipédia") self.verificar_elemento('[title="Visitar a página principal"]') self.atualizar_texto("#searchInput", "Rio de Janeiro") self.clique("#searchButton") self.verificar_texto("Rio de Janeiro", "#firstHeading") self.verificar_elemento('img[alt*="edifícios"]') self.atualizar_texto("#searchInput", "São Paulo") self.clique("#searchButton") self.verificar_texto("São Paulo", "#firstHeading") self.verificar_elemento('img[src*="Monumento"]') self.voltar() self.verificar_verdade("Janeiro" in self.obter_url_atual()) self.avançar() # noqa self.verificar_verdade("Paulo" in self.obter_url_atual())
# Portuguese Language Test - Python 3 Only! from seleniumbase.translate.portuguese import CasoDeTeste class MinhaClasseDeTeste(CasoDeTeste): def test_exemplo_1(self): self.abrir_url("https://pt.wikipedia.org/wiki/") self.verificar_texto("Wikipédia") self.verificar_elemento('[title="Língua portuguesa"]') self.atualizar_texto("#searchInput", "Rio de Janeiro") self.clique("#searchButton") self.verificar_texto("Rio de Janeiro", "#firstHeading") self.verificar_elemento('img[alt*="edifícios"]') self.atualizar_texto("#searchInput", "São Paulo") self.clique("#searchButton") self.verificar_texto("São Paulo", "h1#firstHeading") self.verificar_elemento('img[src*="Monumento"]') self.voltar() self.verificar_verdade("Rio" in self.obter_url_atual()) self.atualizar_texto("#searchInput", "Florianópolis\n") self.verificar_texto("Florianópolis", "h1#firstHeading") self.verificar_elemento('img[alt*="Avenida Beira Mar"]')
Update the Portuguese example test
Update the Portuguese example test
Python
mit
mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase
135c84189720aa2b7c07e516c782f7fab7b4d8fe
astropy/units/format/base.py
astropy/units/format/base.py
# Licensed under a 3-clause BSD style license - see LICENSE.rst class _FormatterMeta(type): registry = {} def __new__(mcls, name, bases, members): if 'name' in members: formatter_name = members['name'].lower() else: formatter_name = members['name'] = name.lower() cls = super().__new__(mcls, name, bases, members) mcls.registry[formatter_name] = cls return cls class Base(metaclass=_FormatterMeta): """ The abstract base class of all unit formats. """ def __new__(cls, *args, **kwargs): # This __new__ is to make it clear that there is no reason to # instantiate a Formatter--if you try to you'll just get back the # class return cls @classmethod def parse(cls, s): """ Convert a string to a unit object. """ raise NotImplementedError( f"Can not parse with {cls.__name__} format") @classmethod def to_string(cls, u): """ Convert a unit object to a string. """ raise NotImplementedError( f"Can not output in {cls.__name__} format")
# Licensed under a 3-clause BSD style license - see LICENSE.rst class Base: """ The abstract base class of all unit formats. """ registry = {} def __new__(cls, *args, **kwargs): # This __new__ is to make it clear that there is no reason to # instantiate a Formatter--if you try to you'll just get back the # class return cls def __init_subclass__(cls, **kwargs): # Keep a registry of all formats. Key by the class name unless a name # is explicitly set (i.e., one *not* inherited from a superclass). if 'name' not in cls.__dict__: cls.name = cls.__name__.lower() Base.registry[cls.name] = cls super().__init_subclass__(**kwargs) @classmethod def parse(cls, s): """ Convert a string to a unit object. """ raise NotImplementedError( f"Can not parse with {cls.__name__} format") @classmethod def to_string(cls, u): """ Convert a unit object to a string. """ raise NotImplementedError( f"Can not output in {cls.__name__} format")
Remove use of metaclass for unit formats
Remove use of metaclass for unit formats
Python
bsd-3-clause
astropy/astropy,saimn/astropy,mhvk/astropy,lpsinger/astropy,saimn/astropy,lpsinger/astropy,mhvk/astropy,pllim/astropy,astropy/astropy,lpsinger/astropy,saimn/astropy,aleksandr-bakanov/astropy,pllim/astropy,astropy/astropy,pllim/astropy,lpsinger/astropy,larrybradley/astropy,mhvk/astropy,pllim/astropy,StuartLittlefair/astropy,larrybradley/astropy,StuartLittlefair/astropy,aleksandr-bakanov/astropy,saimn/astropy,aleksandr-bakanov/astropy,mhvk/astropy,StuartLittlefair/astropy,saimn/astropy,pllim/astropy,larrybradley/astropy,astropy/astropy,StuartLittlefair/astropy,astropy/astropy,larrybradley/astropy,aleksandr-bakanov/astropy,larrybradley/astropy,StuartLittlefair/astropy,lpsinger/astropy,mhvk/astropy
52ef9217f954617283be54c889a317b2432651d7
licensing/models.py
licensing/models.py
from django.db import models class License(models.Model): name = models.CharField(max_length=80, unique=True) symbols = models.CharField(max_length=5) url = models.URLField(unique=True) def __unicode__(self): return self.name def get_absolute_url(self): return self.url class Licensed(models.Model): license = models.ForeignKey(License) class Meta: abstract = True
from django.db import models class License(models.Model): name = models.CharField(max_length=80, unique=True) symbols = models.CharField(max_length=5) url = models.URLField(unique=True) def __unicode__(self): return self.name def __str__(self): return self.name def get_absolute_url(self): return self.url class Licensed(models.Model): license = models.ForeignKey(License) class Meta: abstract = True
Add __str__() method to license model
Add __str__() method to license model __unicode__() is not used in python3
Python
unlicense
editorsnotes/django-licensing,editorsnotes/django-licensing
49d831a61c5770d02609ff2df8fed3effc3869c2
avalonstar/components/games/admin.py
avalonstar/components/games/admin.py
# -*- coding: utf-8 -*- from django.contrib import admin from .models import Game, Platform class GameAdmin(admin.ModelAdmin): list_display = ['name', 'platform', 'gbid'] raw_id_fields = ['platform'] autocomplete_lookup_fields = { 'fk': ['platform'] } admin.site.register(Game, GameAdmin) class PlatformAdmin(admin.ModelAdmin): pass admin.site.register(Platform, PlatformAdmin)
# -*- coding: utf-8 -*- from django.contrib import admin from .models import Game, Platform class GameAdmin(admin.ModelAdmin): list_display = ['name', 'platform', 'gbid', 'is_abandoned', 'is_completed'] raw_id_fields = ['platform'] autocomplete_lookup_fields = { 'fk': ['platform'] } admin.site.register(Game, GameAdmin) class PlatformAdmin(admin.ModelAdmin): pass admin.site.register(Platform, PlatformAdmin)
Add booleans in for games.
Add booleans in for games.
Python
apache-2.0
bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv
ca144b8254691e9633ddedb7ad11b2c4919b8b77
app/stores/views/search.py
app/stores/views/search.py
from django.views.generic import ListView from haystack.query import SearchQuerySet from haystack.utils.geo import Point, D from ..models import Store from ..utils import caching_geo_lookup class DistanceSearchView(ListView): template_name = 'stores/store_search.html' distance = 25 def get_location(self): # TODO: geopy the location based on kwargs location = self.request.GET.get('location') lat = self.request.GET.get('lat') lng = self.request.GET.get('lng') if location: name, geo = caching_geo_lookup(location) elif lat and lng: geo = (lat, lng) else: geo = None self.location_geo = geo return Point(geo[1], geo[0]) def get_distance(self): return D(km=self.request.GET.get('distance', self.distance)) def get_queryset(self): location = self.get_location() if not location: return SearchQuerySet.none distance = self.get_distance() print location, distance return SearchQuerySet().dwithin('location', location, distance).distance('location', location).order_by('-distance') def get_context_data(self, **kwargs): ctx = super(DistanceSearchView, self).get_context_data(**kwargs) ctx.update({ 'location': self.request.GET.get('location'), 'location_geo': self.location_geo, }) return ctx
from django.views.generic import ListView from haystack.query import SearchQuerySet from haystack.utils.geo import Point, D from ..models import Store from ..utils import caching_geo_lookup class DistanceSearchView(ListView): template_name = 'stores/store_search.html' distance = 25 def get_location(self): # TODO: geopy the location based on kwargs location = self.request.GET.get('location') lat = self.request.GET.get('lat') lng = self.request.GET.get('lng') if location: name, geo = caching_geo_lookup(location) elif lat and lng: geo = (float(lat), float(lng)) else: geo = None self.location_geo = geo return Point(geo[1], geo[0]) def get_distance(self): return D(km=self.request.GET.get('distance', self.distance)) def get_queryset(self): location = self.get_location() if not location: return SearchQuerySet.none distance = self.get_distance() print location, distance return SearchQuerySet().dwithin('location', location, distance).distance('location', location).order_by('-distance') def get_context_data(self, **kwargs): ctx = super(DistanceSearchView, self).get_context_data(**kwargs) ctx.update({ 'location': self.request.GET.get('location'), 'location_geo': self.location_geo, }) return ctx
Convert the lat/lng to floats for Point.
Convert the lat/lng to floats for Point.
Python
bsd-3-clause
nikdoof/vapemap,nikdoof/vapemap
0e766eb66eba099071b6cfae49bf79492e29e648
ibmcnx/doc/DataSources.py
ibmcnx/doc/DataSources.py
###### # Check ExId (GUID) by Email through JDBC # # Author: Christoph Stoettner # Mail: [email protected] # Documentation: http://scripting101.stoeps.de # # Version: 2.0 # Date: 2014-06-04 # # License: Apache 2.0 # # Check ExId of a User in all Connections Applications import ibmcnx.functions dbs = AdminConfig.list('DataSource', AdminConfig.getid('/Cell:cnxwas1Cell01/')) print dbs dbs = dbs.split('(')[0] print dbs # dbs = ['FNOSDS', 'FNGCDDS', 'IBM_FORMS_DATA_SOURCE', 'activities', 'blogs', 'communities', 'dogear', 'files', 'forum', 'homepage', 'metrics', 'mobile', 'news', 'oauth provider', 'profiles', 'search', 'wikis'] # List of all databases to check # # for db in dbs: # t1 = ibmcnx.functions.getDSId( db ) # AdminConfig.show( t1 ) # print '\n\n' # AdminConfig.showall( t1 ) # AdminConfig.showAttribute(t1,'statementCacheSize' ) # AdminConfig.showAttribute(t1,'[statementCacheSize]' )
###### # Check ExId (GUID) by Email through JDBC # # Author: Christoph Stoettner # Mail: [email protected] # Documentation: http://scripting101.stoeps.de # # Version: 2.0 # Date: 2014-06-04 # # License: Apache 2.0 # # Check ExId of a User in all Connections Applications import ibmcnx.functions dbs = AdminConfig.list('DataSource', AdminConfig.getid('/Cell:cnxwas1Cell01/')) # print dbs dblist = [] for db in dbs: dblist.append(db) print dblist # dbs = dbs.split('(')[0] # print dbs # dbs = ['FNOSDS', 'FNGCDDS', 'IBM_FORMS_DATA_SOURCE', 'activities', 'blogs', 'communities', 'dogear', 'files', 'forum', 'homepage', 'metrics', 'mobile', 'news', 'oauth provider', 'profiles', 'search', 'wikis'] # List of all databases to check # # for db in dbs: # t1 = ibmcnx.functions.getDSId( db ) # AdminConfig.show( t1 ) # print '\n\n' # AdminConfig.showall( t1 ) # AdminConfig.showAttribute(t1,'statementCacheSize' ) # AdminConfig.showAttribute(t1,'[statementCacheSize]' )
Create documentation of DataSource Settings
8: Create documentation of DataSource Settings Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/8
Python
apache-2.0
stoeps13/ibmcnx2,stoeps13/ibmcnx2
aecff9764ef8d18b7016a6acba41e74a43e66085
clio/utils.py
clio/utils.py
import json from bson import json_util from flask.wrappers import Request, cached_property def getBoolean(string): return { '1': True, 'yes': True, 'true': True, 'on': True, '0': False, 'no': False, 'false': False, 'off': False, '': False, None: False }[string.lower()] class ExtRequest(Request): @cached_property def json(self): """If the mimetype is `application/json` this will contain the parsed JSON data. """ if self.mimetype in ('application/json','application/extjson'): if 'ext' in self.mimetype: objhook = json_util.object_hook else: objhook = None request_charset = self.mimetype_params.get('charset') if request_charset is not None: j = json.loads(self.data, encoding=request_charset, object_hook=objhook ) else: j = json.loads(self.data, object_hook=objhook) return j
import json from bson import json_util from flask.wrappers import Request, cached_property def getBoolean(string): if string is None: return False return { '1': True, 'yes': True, 'true': True, 'on': True, '0': False, 'no': False, 'false': False, 'off': False, '': False, None: False }[string.lower()] class ExtRequest(Request): @cached_property def json(self): """If the mimetype is `application/json` this will contain the parsed JSON data. """ if self.mimetype in ('application/json','application/extjson'): if 'ext' in self.mimetype: objhook = json_util.object_hook else: objhook = None request_charset = self.mimetype_params.get('charset') if request_charset is not None: j = json.loads(self.data, encoding=request_charset, object_hook=objhook ) else: j = json.loads(self.data, object_hook=objhook) return j
Add support to getBoolean function for None objects.
Add support to getBoolean function for None objects.
Python
apache-2.0
geodelic/clio,geodelic/clio
58d73429952a942d03b232242424946895ec3e8c
multi_schema/middleware.py
multi_schema/middleware.py
""" Middleware to automatically set the schema (namespace). if request.user.is_superuser, then look for a ?schema=XXX and set the schema to that. Otherwise, set the schema to the one associated with the logged in user. """ from models import Schema class SchemaMiddleware: def process_request(self, request): if request.user.is_anonymous(): return None if request.user.is_superuser and '__schema' in request.GET: request.session['schema'] = request.GET['__schema'] if request.user.is_superuser and 'schema' in request.session: Schema.objects.get(pk=request.session['schema']).activate() else: request.user.schema.schema.activate() def process_response(self, request): pass
""" Middleware to automatically set the schema (namespace). if request.user.is_superuser, then look for a ?schema=XXX and set the schema to that. Otherwise, set the schema to the one associated with the logged in user. """ from django.core.exceptions import ObjectDoesNotExist from models import Schema class SchemaMiddleware: def process_request(self, request): if request.user.is_anonymous(): return None if request.user.is_superuser: if '__schema' in request.GET: request.session['schema'] = request.GET['__schema'] if 'schema' in request.session: Schema.objects.get(pk=request.session['schema']).activate() else: try: request.user.schema.schema.activate() except ObjectDoesNotExist: pass def process_template_response(self, request, response): if request.user.is_superuser: response.context_data['schemata'] = Schema.objects.all() response.context_data['selected_schema'] = request.session['schema'] return response
Add some data into the request context. Better handling of missing Schema objects when logging in (should we raise an error?).
Add some data into the request context. Better handling of missing Schema objects when logging in (should we raise an error?).
Python
bsd-3-clause
schinckel/django-boardinghouse,schinckel/django-boardinghouse,schinckel/django-boardinghouse
b98e86ad9b3120dce9f163236b5e28f564547c27
TWLight/resources/factories.py
TWLight/resources/factories.py
# -*- coding: utf-8 -*- import factory import random from django.conf import settings from TWLight.resources.models import Partner, Stream, Video, Suggestion class PartnerFactory(factory.django.DjangoModelFactory): class Meta: model = Partner strategy = factory.CREATE_STRATEGY company_name = factory.Faker( "company", locale=random.choice(settings.FAKER_LOCALES) ) terms_of_use = factory.Faker("uri", locale=random.choice(settings.FAKER_LOCALES)) status = Partner.AVAILABLE # not the default, but usually wanted in tests class StreamFactory(factory.django.DjangoModelFactory): class Meta: model = Stream strategy = factory.CREATE_STRATEGY partner = factory.SubFactory(PartnerFactory) name = factory.Faker("bs", locale=random.choice(settings.FAKER_LOCALES)) class SuggestionFactory(factory.django.DjangoModelFactory): class Meta: model = Suggestion strategy = factory.CREATE_STRATEGY suggested_company_name = factory.Faker( "company", locale=random.choice(settings.FAKER_LOCALES) ) company_url = factory.Faker("url", locale=random.choice(settings.FAKER_LOCALES)) class VideoFactory(factory.django.DjangoModelFactory): class Meta: model = Video strategy = factory.CREATE_STRATEGY partner = factory.SubFactory(PartnerFactory)
# -*- coding: utf-8 -*- import factory import random from django.conf import settings from TWLight.resources.models import Partner, Stream, Video, Suggestion class PartnerFactory(factory.django.DjangoModelFactory): class Meta: model = Partner strategy = factory.CREATE_STRATEGY company_name = factory.Faker( "company", locale=random.choice(settings.FAKER_LOCALES) ) terms_of_use = factory.Faker("uri", locale=random.choice(settings.FAKER_LOCALES)) status = Partner.AVAILABLE # not the default, but usually wanted in tests class StreamFactory(factory.django.DjangoModelFactory): class Meta: model = Stream strategy = factory.CREATE_STRATEGY partner = factory.SubFactory(PartnerFactory) name = factory.Faker("bs", locale=random.choice(settings.FAKER_LOCALES)) class SuggestionFactory(factory.django.DjangoModelFactory): class Meta: model = Suggestion strategy = factory.CREATE_STRATEGY suggested_company_name = factory.Faker("pystr", max_chars=40) company_url = factory.Faker("url", locale=random.choice(settings.FAKER_LOCALES)) class VideoFactory(factory.django.DjangoModelFactory): class Meta: model = Video strategy = factory.CREATE_STRATEGY partner = factory.SubFactory(PartnerFactory)
Change suggested_company_name factory var to pystr
Change suggested_company_name factory var to pystr
Python
mit
WikipediaLibrary/TWLight,WikipediaLibrary/TWLight,WikipediaLibrary/TWLight,WikipediaLibrary/TWLight,WikipediaLibrary/TWLight
bf5307afe52415960d0ffc794f687b0ecebb48da
app/__init__.py
app/__init__.py
from flask import Flask from flask.ext.login import login_user, logout_user, current_user, login_required, LoginManager from flask.ext.sqlalchemy import SQLAlchemy from flask import Flask, session from flask.ext.session import Session from flask.ext.mail import Mail app = Flask(__name__) # Configuration file reading app.config.from_object('config') # Database Initialization db = SQLAlchemy(app) # Login manager init lm = LoginManager() lm.init_app(app) lm.login_view = 'login' # Session Manager Init sess = Session() sess.init_app(app) # Mail engine init mail = Mail(app) from app import views, models
from flask import Flask from flask.ext.login import login_user, logout_user, current_user, login_required, LoginManager from flask.ext.sqlalchemy import SQLAlchemy from flask import Flask, session from flask.ext.session import Session from flask.ext.mail import Mail import logging from logging.handlers import RotatingFileHandler app = Flask(__name__) # Configuration file reading app.config.from_object('config') # Database Initialization db = SQLAlchemy(app) # Login manager init lm = LoginManager() lm.init_app(app) lm.login_view = 'login' # Session Manager Init sess = Session() sess.init_app(app) # Mail engine init mail = Mail(app) ################## # Logging system # ################## # Open a file rotated every 100MB file_handler = RotatingFileHandler('tmp/cineapp.log', 'a', 100 * 1024 * 1024, 10) file_handler.setFormatter(logging.Formatter('%(asctime)s %(levelname)s: %(message)s [in %(pathname)s:%(lineno)d]')) app.logger.setLevel(logging.INFO) app.logger.addHandler(file_handler) app.logger.info('Cineapp startup') from app import views, models
Enable file logging for the application.
Enable file logging for the application.
Python
mit
ptitoliv/cineapp,ptitoliv/cineapp,ptitoliv/cineapp
04328bb0ed84180aa9e5ce7f749eafb1ab96d4fc
app/api/auth.py
app/api/auth.py
from urllib import urlencode from datetime import datetime from django.http import HttpResponseForbidden from django.contrib.auth.models import AnonymousUser from django.utils.timezone import now from api.models import AuthAPIKey, AuthAPILog class APIKeyAuthentication(object): """ Validats a request by API key passed as a GET parameter """ def is_authenticated(self, request): try: keyobj = AuthAPIKey.objects.get(key=request.GET.get('apikey', None)) except AuthAPIKey.DoesNotExist: pass else: if keyobj and keyobj.active: params = request.GET.copy() if params.get('apikey', None): del params['apikey'] if len(params): url = "%s?%s" % (request.path, urlencode(params)) else: url = request.path AuthAPILog(key=keyobj, access_datetime=now(), url=url).save() request.user = AnonymousUser() request.api_key = keyobj return True return False def challenge(self): return HttpResponseForbidden('Access Denied, use a API Key')
from urllib import urlencode from datetime import datetime from django.http import HttpResponseForbidden from django.contrib.auth.models import AnonymousUser from django.utils.timezone import now from api.models import AuthAPIKey, AuthAPILog class APIKeyAuthentication(object): """ Validats a request by API key passed as a GET parameter """ def is_authenticated(self, request): try: keyobj = AuthAPIKey.objects.get(key=request.GET.get('apikey', None)) except AuthAPIKey.DoesNotExist: pass else: if keyobj and keyobj.active: params = request.GET.copy() if params.get('apikey', None): del params['apikey'] if len(params): url = "%s?%s" % (request.path, urlencode(params)) else: url = request.path AuthAPILog.objects.create(key=keyobj, access_datetime=now(), url=url) request.user = AnonymousUser() request.api_key = keyobj return True return False def challenge(self): return HttpResponseForbidden('Access Denied, use a API Key')
Use create instead of instance and save
Use create instead of instance and save
Python
bsd-3-clause
nikdoof/test-auth
6091fccc90bb6b90c47a2e4fb7ee6821876eb1a1
synthnotes/generators/lengthgenerator.py
synthnotes/generators/lengthgenerator.py
from pkg_resources import resource_filename import pandas as pd import numpy as np class LengthGenerator(object): def __init__(self, length_file=resource_filename(__name__, 'resources/note_lengths.csv')): # print(length_file) df = pd.read_csv(length_file) notes_count = df['count'].sum() df['probability'] = df['count'] / notes_count self.note_lengths = df['note_length'].as_matrix() self.p = df['probability'].as_matrix() def generate(self, size=1): return np.random.choice(self.note_lengths, size=size, p=self.p)
from pkg_resources import resource_filename import pandas as pd import numpy as np class LengthGenerator(object): def __init__(self, length_file=resource_filename('synthnotes.resources', 'note_lengths.csv')): # print(length_file) df = pd.read_csv(length_file) notes_count = df['count'].sum() df['probability'] = df['count'] / notes_count self.note_lengths = df['note_length'].as_matrix() self.p = df['probability'].as_matrix() def generate(self, size=1): return np.random.choice(self.note_lengths, size=size, p=self.p)
Change LengthGenerator to get appropriate file path
Change LengthGenerator to get appropriate file path
Python
mit
ebegoli/SynthNotes
fc7cadecb95fa798a8e8aaeb544ad5464f13a533
nanomon/registry.py
nanomon/registry.py
from weakref import WeakValueDictionary class DuplicateEntryError(Exception): def __init__(self, name, obj, registry): self.name = name self.obj = obj self.registry = registry def __str__(self): return "Duplicate entry in '%s' registry for '%s'." % ( self.registry._registry_name, self.name) class Registry(WeakValueDictionary): def __init__(self, object_type, *args, **kwargs): self._object_type = object_type #super(Registry, self).__init__(*args, **kwargs) WeakValueDictionary.__init__(self, *args, **kwargs) def __setitem__(self, name, value): if not isinstance(value, self._object_type): raise TypeError("This registry only accepts objects of type %s." % (self._object_type.__name__)) if self.has_key(name): raise DuplicateEntryError(name, value, self) WeakValueDictionary.__setitem__(self, name, value)
from weakref import WeakValueDictionary class DuplicateEntryError(Exception): def __init__(self, name, obj, registry): self.name = name self.obj = obj self.registry = registry def __str__(self): return "Duplicate entry in '%s' registry for '%s'." % ( self.registry._object_type.__name__, self.name) class Registry(WeakValueDictionary): def __init__(self, object_type, *args, **kwargs): self._object_type = object_type WeakValueDictionary.__init__(self, *args, **kwargs) def __setitem__(self, name, value): if not isinstance(value, self._object_type): raise TypeError("This registry only accepts objects of type %s." % (self._object_type.__name__)) if self.has_key(name): raise DuplicateEntryError(name, value, self) WeakValueDictionary.__setitem__(self, name, value)
Clean up some commented out code
Clean up some commented out code
Python
bsd-2-clause
cloudtools/nymms
d9f03ad1c73cc18276666f28e9a9360c71139a0d
nib/plugins/time.py
nib/plugins/time.py
import datetime import time from nib import jinja @jinja('time') def timeformat(t=None, f='%Y-%m-%d %I:%M %p'): if t is None: t = time.gmtime() elif isinstance(t, datetime.date) or isinstance(t, datetime.datetime): t = t.timetuple() elif isinstance(t, float): t = time.gmtime(t) s = time.strftime(f, t) return s @jinja('atomtime') def atomtimeformat(t=None, f='%Y-%m-%dT%I:%M:%SZ'): return timeformat(t,f) @jinja('rsstime') def rsstimeformat(t=None, f='%a, %d %b %Y %I:%M:%S GMT'): return timeformat(t,f) @jinja('date') def dateformat(t=None, f='%Y-%m-%d'): return timeformat(t,f)
import datetime import time from nib import jinja @jinja('time') def timeformat(t=None, f='%Y-%m-%d %I:%M %p'): if t is None: t = time.gmtime() elif isinstance(t, datetime.date) or isinstance(t, datetime.datetime): t = t.timetuple() elif isinstance(t, float): t = time.gmtime(t) s = time.strftime(f, t) return s @jinja('atomtime') def atomtimeformat(t=None, f='%Y-%m-%dT%I:%M:%SZ'): return timeformat(t,f) @jinja('rsstime') def rsstimeformat(t=None, f='%a, %d %b %Y %I:%M:%S GMT'): return timeformat(t,f) @jinja('date') def dateformat(t=None, f='%Y-%m-%d'): return timeformat(t,f) @jinja('longdate') def longdateformat(t=None, f='%B %d, %Y'): return timeformat(t, f)
Add 'longdate' filter for readable dates in templates
Add 'longdate' filter for readable dates in templates
Python
mit
jreese/nib
43a515ddfbe38686672fe00d4765d3f2e1bc5346
scarlet/assets/settings.py
scarlet/assets/settings.py
from django.conf import settings # Main Assets Directory. This will be a subdirectory within MEDIA_ROOT. # Set to None to use MEDIA_ROOT directly DIRECTORY = getattr(settings, "ASSETS_DIR", 'assets') # Which size should be used as CMS thumbnail for images. CMS_THUMBNAIL_SIZE = getattr(settings, 'ASSETS_CMS_THUMBNAIL_SIZE', '80x80') # EXTRA SETTINGS # Convert Filename (UUID) HASH_FILENAME = getattr(settings, "ASSETS_HASH_FILENAME", True) # Append a qs to assets urls for cache busting USE_CACHE_BUST = getattr(settings, "ASSETS_USE_CACHE_BUST", True) ASSET_MODEL = getattr(settings, "ASSET_MODEL", "assets.Asset") ASSET_TYPES = getattr(settings, "ASSET_TYPES", None) DEFAULT_IMAGE_SIZES = { 'admin' : { 'width' : 100, 'height' : 100, 'editable': False } } IMAGE_SIZES = getattr(settings, "IMAGE_SIZES", DEFAULT_IMAGE_SIZES) IMAGE_CROPPER = '.crops.cropper' CELERY = getattr(settings, "ASSET_CELERY", None) USE_CELERY_DECORATOR = getattr(settings, "ASSET_USE_CELERY_DECORATOR", False)
from django.conf import settings # Main Assets Directory. This will be a subdirectory within MEDIA_ROOT. # Set to None to use MEDIA_ROOT directly DIRECTORY = getattr(settings, "ASSETS_DIR", 'assets') # Which size should be used as CMS thumbnail for images. CMS_THUMBNAIL_SIZE = getattr(settings, 'ASSETS_CMS_THUMBNAIL_SIZE', '80x80') # EXTRA SETTINGS # Convert Filename (UUID) HASH_FILENAME = getattr(settings, "ASSETS_HASH_FILENAME", True) # Append a qs to assets urls for cache busting USE_CACHE_BUST = getattr(settings, "ASSETS_USE_CACHE_BUST", True) ASSET_MODEL = getattr(settings, "ASSET_MODEL", "assets.Asset") ASSET_TYPES = getattr(settings, "ASSET_TYPES", None) DEFAULT_IMAGE_SIZES = { 'admin' : { 'width' : 100, 'height' : 100, 'editable': False, 'upscale': True, }, } IMAGE_SIZES = getattr(settings, "IMAGE_SIZES", DEFAULT_IMAGE_SIZES) IMAGE_CROPPER = '.crops.cropper' CELERY = getattr(settings, "ASSET_CELERY", None) USE_CELERY_DECORATOR = getattr(settings, "ASSET_USE_CELERY_DECORATOR", False)
Set upscale to True by default for admin asset
Set upscale to True by default for admin asset
Python
mit
ff0000/scarlet,ff0000/scarlet,ff0000/scarlet,ff0000/scarlet,ff0000/scarlet
b57d5ecf56640c9d0a69b565006e2240662d6b46
profile_collection/startup/11-temperature-controller.py
profile_collection/startup/11-temperature-controller.py
from ophyd import PVPositioner, EpicsSignal, EpicsSignalRO from ophyd import Component as C from ophyd.device import DeviceStatus class CS700TemperatureController(PVPositioner): setpoint = C(EpicsSignal, 'T-SP') readback = C(EpicsSignalRO, 'T-I') done = C(EpicsSignalRO, 'Cmd-Busy') stop_signal = C(EpicsSignal, 'Cmd-Cmd') def trigger(self): # There is nothing to do. Just report that we are done. # Note: This really should not necessary to do -- # future changes to PVPositioner may obviate this code. status = DeviceStatus() status._finished() return status cs700 = CS700TemperatureController('XF:28IDC-ES:1{Env:01}', name='cs700', settle_time=10) cs700.done_value = 0 cs700.read_attrs = ['setpoint', 'readback'] cs700.readback.name = 'temperautre' cs700.setpoint.name = 'temperautre_setpoint'
from ophyd import PVPositioner, EpicsSignal, EpicsSignalRO from ophyd import Component as C from ophyd.device import DeviceStatus class CS700TemperatureController(PVPositioner): setpoint = C(EpicsSignal, 'T-SP') readback = C(EpicsSignalRO, 'T-I') done = C(EpicsSignalRO, 'Cmd-Busy') stop_signal = C(EpicsSignal, 'Cmd-Cmd') def trigger(self): # There is nothing to do. Just report that we are done. # Note: This really should not necessary to do -- # future changes to PVPositioner may obviate this code. status = DeviceStatus() status._finished() return status cs700 = CS700TemperatureController('XF:28IDC-ES:1{Env:01}', name='cs700') # this functionality never worked, has now been removed, but will shortly be # coming back # settle_time=10) cs700.done_value = 0 cs700.read_attrs = ['setpoint', 'readback'] cs700.readback.name = 'temperautre' cs700.setpoint.name = 'temperautre_setpoint'
Remove settle_time kwarg from c700
Remove settle_time kwarg from c700 This kwarg has been removed from ophyd, but will be coming back (and be functional) soon. Revert these changes when that happens: ophyd 0.2.1)
Python
bsd-2-clause
NSLS-II-XPD/ipython_ophyd,NSLS-II-XPD/ipython_ophyd
28627a41918be15037ba22e930a45d022e88388d
opps/articles/adminx.py
opps/articles/adminx.py
# -*- coding: utf-8 -*- #from django.contrib import admin from .models import Post, Album, Link from opps.contrib import admin admin.site.register(Post) admin.site.register(Album) admin.site.register(Link)
# -*- coding: utf-8 -*- from django.utils.translation import ugettext_lazy as _ from .models import Post, Album, Link from opps.containers.models import ContainerSource, ContainerImage from opps.contrib import admin from opps.contrib.admin.layout import * from xadmin.plugins.inline import Inline class ImageInline(object): model = ContainerImage style = 'accordion' class SourceInline(object): model = ContainerSource style = 'accordion' class PostAdmin(object): raw_id_fields = ['main_image', 'channel', 'albums'] inlines = [ImageInline, SourceInline] style_fields = {'system': "radio-inline"} form_layout = ( Main( TabHolder( Tab(_(u'Identification'), Fieldset('site', 'title', 'slug', 'get_http_absolute_url', 'short_url'), ), Tab(_(u'Content'), Fieldset('hat', 'short_title', 'headline', 'content', 'main_image', 'main_image_caption', 'image_thumb' 'tags'), Inline(ContainerImage), Inline(ContainerSource), ), Tab(_(u'Relationships'), Fieldset('channel', 'albums'), ), )), Side( Fieldset(_(u'Publication'), 'published', 'date_available', 'show_on_root_channel', 'in_containerboxes') ) ) reversion_enable = True admin.site.register(Post, PostAdmin) admin.site.register(Album) admin.site.register(Link)
Add Inline example on post model xadmin
Add Inline example on post model xadmin
Python
mit
jeanmask/opps,opps/opps,YACOWS/opps,williamroot/opps,opps/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,opps/opps,williamroot/opps,jeanmask/opps,opps/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,YACOWS/opps