input
stringlengths 0
2.17k
| instruction
stringlengths 18
2.94k
| output
stringlengths 47
3.36k
|
---|---|---|
#!/usr/bin/python
from setuptools import setup, find_packages
setup(name="github-comments",
version='1.3',
url="https://github.com/alikins/github-comments",
description="show pull request comments in a lint like fashion",
author="Adrian Likins",
author_email="[email protected]",
packages=find_packages(),
install_requires=['argparse', 'requests'],
#scripts=["scripts/github-comments"],)
entry_points={'console_scripts': ['github-comments = github_comments.github_comments:main']})
| Add markdown and BeautifulSoup to requires.
| #!/usr/bin/python
from setuptools import setup, find_packages
setup(name="github-comments",
version='1.3',
url="https://github.com/alikins/github-comments",
description="show pull request comments in a lint like fashion",
author="Adrian Likins",
author_email="[email protected]",
packages=find_packages(),
install_requires=['argparse',
'BeautifulSoup',
'markdown',
'requests'
],
#scripts=["scripts/github-comments"],)
entry_points={'console_scripts':
['github-comments = github_comments.github_comments:main']}
)
|
#!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='TBD',
author='Kushal Pisavadia',
author_email='[email protected]',
maintainer='Kushal Pisavadia',
maintainer_email='[email protected]',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack'],
tests_require=['pytest'],
cmdclass={'test': PyTest}
)
| Use the correct module name for 'msgpack' Python module
| #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='TBD',
author='Kushal Pisavadia',
author_email='[email protected]',
maintainer='Kushal Pisavadia',
maintainer_email='[email protected]',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python'],
tests_require=['pytest'],
cmdclass={'test': PyTest}
)
|
import tweepy
import time
class TwitterBot:
def __init__(self, auth, listen_msg, response_msg):
auth = tweepy.OAuthHandler(auth['consumer_key'], auth['consumer_secret'])
auth.set_access_token(auth['access_token'], auth['access_token_secret'])
self.api = tweepy.API(auth)
self.responded_tweets = set()
self.listen, self.response = listen_msg, response_msg
def tweet(self, message, mention_id=None):
self.api.update_status(status=message, mention_id=mention_id)
def respond(self, mention_text, message):
for mention in self.api.mentions_timeline(count=1):
if mention_text in mention.text.lower():
self.tweet(message.format(mention.user.screen_name), mention.id)
self.api.create_favorite(mention.id)
print('Responded to {0}.'.format(mention.user.screen_name))
if __name__ == '__main__':
tb = TwitterBot()
tb.respond('hi', '{} hey buddy!')
| Update Tweet method for api change
| import tweepy
import time
class TwitterBot:
def __init__(self, auth, listen_msg, response_msg):
auth = tweepy.OAuthHandler(auth['consumer_key'], auth['consumer_secret'])
auth.set_access_token(auth['access_token'], auth['access_token_secret'])
self.api = tweepy.API(auth)
self.responded_tweets = set()
self.listen, self.response = listen_msg, response_msg
def tweet(self, message, mention_id=None):
self.api.update_status(status=message, in_reply_to_status_id=mention_id)
def respond(self, mention_text, message):
for mention in self.api.mentions_timeline(count=1):
if mention_text in mention.text.lower():
self.tweet(message.format(mention.user.screen_name), mention.id)
self.api.create_favorite(mention.id)
print('Responded to {0}.'.format(mention.user.screen_name))
if __name__ == '__main__':
tb = TwitterBot()
tb.respond('hi', '{} hey buddy!')
|
from django.urls import path
from django.contrib.sitemaps.views import sitemap
from story.views import StoryListView
from story.sitemap import BlogSiteMap
sitemaps = {
"blog": BlogSiteMap
}
app_name = "story"
urlpatterns = [
path('', StoryListView.as_view(), name='stories'),
path('/sitemap.xml', sitemap, {'sitemaps': sitemaps}, name='sitemap'),
]
| Remove / in the sitemap url
| from django.urls import path
from django.contrib.sitemaps.views import sitemap
from story.views import StoryListView
from story.sitemap import BlogSiteMap
sitemaps = {
"blog": BlogSiteMap
}
app_name = "story"
urlpatterns = [
path('', StoryListView.as_view(), name='stories'),
path('sitemap.xml', sitemap, {'sitemaps': sitemaps}, name='sitemap'),
]
|
#!/bin/env python
#
# glidein_ls
#
# Execute a ls command on the job directory
#
# Usage:
# glidein_ls.py <cluster>.<process> [-name <schedd_name>] [-pool <pool_name> ] [-timeout <nr secs>
#
import os
import string
import stat
import sys
sys.path.append("lib")
sys.path.append("../lib")
import glideinMonitor
def createDirMonitorFile(monitor_file_name,monitor_control_relname,
argv,condor_status):
fd=open(monitor_file_name,"w")
try:
fd.write("#!/bin/sh\n")
fd.write("outdir=`ls -lt .. | tail -1 | awk '{print $9}'`\n")
fd.write("(cd ../$outdir; if [ $? -eq 0 ]; then ls %s; else echo Internal error; fi)\n"%(string.join(argv)))
fd.write("echo Done > %s\n"%monitor_control_relname)
finally:
fd.close()
os.chmod(monitor_file_name,stat.S_IRWXU)
args=glideinMonitor.parseArgs(sys.argv[1:])
glideinMonitor.monitor(args['jid'],args['schedd_name'],args['pool_name'],
args['timeout'],
createDirMonitorFile,args['argv'])
| Change rel paths into abspaths
| #!/bin/env python
#
# condor_ls
#
# Description:
# Execute a ls command on a condor job working directory
#
# Usage:
# glidein_ls.py <cluster>.<process> [<dir>] [-name <schedd_name>] [-pool <pool_name> ] [-timeout <nr secs>]
#
# Author:
# Igor Sfiligoi (May 2007)
#
# License:
# Fermitools
#
import os,os.path
import string
import stat
import sys
sys.path.append(os.path.join(os.path[0],"lib"))
sys.path.append(os.path.join(os.path[0],"../lib"))
import glideinMonitor
def createDirMonitorFile(monitor_file_name,monitor_control_relname,
argv,condor_status):
fd=open(monitor_file_name,"w")
try:
fd.write("#!/bin/sh\n")
fd.write("outdir=`ls -lt .. | tail -1 | awk '{print $9}'`\n")
fd.write("(cd ../$outdir; if [ $? -eq 0 ]; then ls %s; else echo Internal error; fi)\n"%(string.join(argv)))
fd.write("echo Done > %s\n"%monitor_control_relname)
finally:
fd.close()
os.chmod(monitor_file_name,stat.S_IRWXU)
args=glideinMonitor.parseArgs(sys.argv[1:])
glideinMonitor.monitor(args['jid'],args['schedd_name'],args['pool_name'],
args['timeout'],
createDirMonitorFile,args['argv'])
|
from codecs import open
from os import path
from setuptools import setup
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.rst'), encoding='utf-8') as f:
readme = f.read()
with open(path.join(here, 'requirements', 'install.txt'),
encoding='utf-8') as f:
install_requires = f.read().splitlines()
setup(
name='analyzere_extras',
version='0.1.10',
description='Python extras to support visualization',
long_description=readme,
url='https://github.com/analyzere/analyzere-python-extras',
author='Analyze Re',
author_email='[email protected]',
license='MIT',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development :: Libraries :: Python Modules',
],
packages=[
'analyzere_extras',
],
install_requires=install_requires
)
| Increment version number for added ELT combiner tool
| from codecs import open
from os import path
from setuptools import setup
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.rst'), encoding='utf-8') as f:
readme = f.read()
with open(path.join(here, 'requirements', 'install.txt'),
encoding='utf-8') as f:
install_requires = f.read().splitlines()
setup(
name='analyzere_extras',
version='0.2.0',
description='Python extras to support visualization',
long_description=readme,
url='https://github.com/analyzere/analyzere-python-extras',
author='Analyze Re',
author_email='[email protected]',
license='MIT',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development :: Libraries :: Python Modules',
],
packages=[
'analyzere_extras',
],
install_requires=install_requires
)
|
import os
from setuptools import setup
with open(os.path.join(os.path.dirname(__file__), 'README.rst')) as readme:
README = readme.read()
# allow setup.py to be run from any path
os.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))
setup(
name='mezzanine-sermons',
version='0.1.0',
packages=['mezzanine_sermons'],
include_package_data=True,
license='BSD License',
description='A simple mezzanine app which facilitates the management and playing of sermons',
long_description=README,
url='https://github.com/philipsouthwell/mezzanine-sermons',
author='Philip Southwell',
author_email='[email protected]',
keywords=['django', 'mezzanine'],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Internet :: WWW/HTTP',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
)
| Change version number for release
| import os
from setuptools import setup
with open(os.path.join(os.path.dirname(__file__), 'README.rst')) as readme:
README = readme.read()
# allow setup.py to be run from any path
os.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))
setup(
name='mezzanine-sermons',
version='0.1.1',
packages=['mezzanine_sermons'],
include_package_data=True,
license='BSD License',
description='A simple mezzanine app which facilitates the management and playing of sermons',
long_description=README,
url='https://github.com/philipsouthwell/mezzanine-sermons',
author='Philip Southwell',
author_email='[email protected]',
keywords=['django', 'mezzanine'],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Internet :: WWW/HTTP',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
)
|
from setuptools import setup, find_packages
setup(
name = 'dictsheet',
version = '0.0.3',
keywords = ('dictsheet', 'spreadsheet', 'gspread'),
description = 'Dict wrapper for google spreadsheet',
license = 'MIT License',
install_requires = ['gspread>=0.4.1'],
url = 'https://github.com/previa/dictsheet',
author = 'Chandler Huang, Xander Li',
author_email = '[email protected]',
packages = find_packages(),
platforms = 'any',
)
| [Feature] Include packages in requirements.txt to install_requires().
| from setuptools import setup, find_packages
from pip.req import parse_requirements
install_reqs = parse_requirements('requirements.txt', session=False)
reqs = [str(ir.req) for ir in install_reqs]
# REFERENCE:
# http://stackoverflow.com/questions/14399534/how-can-i-reference-requirements-txt-for-the-install-requires-kwarg-in-setuptool
setup(
name = 'dictsheet',
version = '0.0.9',
keywords = ('dictsheet', 'spreadsheet', 'gspread'),
description = 'Dict wrapper for google spreadsheet',
license = 'MIT License',
install_requires = reqs,
data_files = ['requirements.txt', 'README.md', 'LICENSE.txt'],
url = 'https://github.com/previa/dictsheet',
author = 'Chandler Huang, Xander Li',
author_email = '[email protected]',
packages = find_packages(),
platforms = 'any',
)
|
from django.core.management.base import (
BaseCommand,
)
from apps.api.models import (
Convention,
Contest,
Contestant,
Performance,
)
class Command(BaseCommand):
help = "Command to denormailze data."
def handle(self, *args, **options):
vs = Convention.objects.all()
for v in vs:
v.save()
ts = Contest.objects.all()
for t in ts:
t.save()
cs = Contestant.objects.all()
for c in cs:
c.save()
ps = Performance.objects.all()
for p in ps:
p.save()
for t in ts:
t.rank()
return "Done"
| Remove ranking from denormalization command
| from django.core.management.base import (
BaseCommand,
)
from apps.api.models import (
Convention,
Contest,
Contestant,
Performance,
)
class Command(BaseCommand):
help = "Command to denormailze data."
def handle(self, *args, **options):
vs = Convention.objects.all()
for v in vs:
v.save()
ts = Contest.objects.all()
for t in ts:
t.save()
cs = Contestant.objects.all()
for c in cs:
c.save()
ps = Performance.objects.all()
for p in ps:
p.save()
return "Done"
|
"""
File: singleton.py
Purpose: Defines a class whose subclasses will act like the singleton pattern.
"""
class Singleton(object):
"""
This is a class that implements singleton for its subclasses.
The technique is based on a variant of other techniques found in:
http://stackoverflow.com/questions/6760685/creating-a-singleton-in-python
https://gist.github.com/werediver/4396488
The technique is simply to build a map of classes to their unique instances.
The first time called for some particular
class the class is mapped to the instance. On other class to the same class, the mapped instance is returned.
"""
_instances = {}
@classmethod
def instance(cls):
if cls not in cls._instances:
cls._instances[cls] = cls()
return cls._instances[cls]
| Add comments to Singleton about usage.
|
class Singleton(object):
"""
This is a class that implements singleton for its subclasses.
The technique is based on a variant of other techniques found in:
http://stackoverflow.com/questions/6760685/creating-a-singleton-in-python
https://gist.github.com/werediver/4396488
The technique is simply to build a map of classes to their unique instances.
The first time called for some particular
class the class is mapped to the instance. On other class to the same class, the mapped instance is returned.
Classes that use this must:
1) Add Singleton as a superclass.
2) Have this signature for the constructor: __init__(self, *args, **kwargs)
"""
_instances = {}
@classmethod
def instance(cls, *args, **kwargs):
if cls not in cls._instances:
cls._instances[cls] = cls(*args, **kwargs)
return cls._instances[cls]
|
from django import template
import json
from django.conf import settings
register = template.Library()
@register.simple_tag(takes_context=True)
def get_admin_context(context):
ctx = {'location':None, 'param1':None, 'param2':None, 'static':settings.STATIC_URL}
try:
context['app_list']
if len(context['app_list']) > 1:
# we are at /admin/
ctx['location'] = 'admin_index'
return json.dumps(ctx)
elif len(context['app_list']) == 1:
#We are at the app index eg: /admin/app/
ctx['location'] = 'app_index'
ctx['param1'] = context['app_list'][0]['name'].lower()
return json.dumps(ctx)
except KeyError:
pass
try:
# we are in the list view of the model eg: admin/app/model/
ctx['location']= 'model_list'
module_name = context['module_name'].lower()
# django adds an 's' to every model name in this view, so we are gonna
ctx['param1'] = module_name
return json.dumps(ctx)
except KeyError:
pass
try:
# editing a model
ctx['location']= 'model_edit'
ctx['param1'] = context['original'].__class__.__name__.lower()
ctx['param2'] = context['original'].pk
return json.dumps(ctx)
except KeyError:
pass
return json.dumps(ctx)
| Check if original object is None
| from django import template
import json
from django.conf import settings
register = template.Library()
@register.simple_tag(takes_context=True)
def get_admin_context(context):
ctx = {'location':None, 'param1':None, 'param2':None, 'static':settings.STATIC_URL}
try:
context['app_list']
if len(context['app_list']) > 1:
# we are at /admin/
ctx['location'] = 'admin_index'
return json.dumps(ctx)
elif len(context['app_list']) == 1:
#We are at the app index eg: /admin/app/
ctx['location'] = 'app_index'
ctx['param1'] = context['app_list'][0]['name'].lower()
return json.dumps(ctx)
except KeyError:
pass
try:
# we are in the list view of the model eg: admin/app/model/
ctx['location']= 'model_list'
module_name = context['module_name'].lower()
# django adds an 's' to every model name in this view, so we are gonna
ctx['param1'] = module_name
return json.dumps(ctx)
except KeyError:
pass
try:
# editing a model
ctx['location']= 'model_edit'
if context['original'] is not None:
ctx['param1'] = context['original'].__class__.__name__.lower()
ctx['param2'] = context['original'].pk
return json.dumps(ctx)
except KeyError:
pass
return json.dumps(ctx)
|
from __future__ import unicode_literals
import os
import logging
from . import helper
def readStatus(config, student):
student = student.lower()
path = config("attachment_path")
if not os.path.exists(path):
return
path = os.path.join(path, student)
if not os.path.exists(path):
return "Student ohne Abgabe"
path = os.path.join(path, "korrekturstatus.txt")
if not os.path.exists(path):
return "Unbearbeitet"
statusfile = open(path, "r")
status = statusfile.read()
statusfile.close()
return status
def writeStatus(config, student, status):
student = student.lower()
status = status.lower()
path = os.path.join(config("attachment_path"), student)
if not os.path.exists(path):
logging.error("Requested student '%s' hasn't submitted anything yet.")
return
path = os.path.join(path, "korrekturstatus.txt")
with open(path, "w") as statusfile:
statusfile.write(status)
| Move status to database instead of status file (server side)
| from __future__ import unicode_literals
import os
import logging
import sqlite3
from . import helper
def readStatus(config, student):
database = getStatusTable(config)
cursor = database.cursor()
cursor.execute("SELECT status FROM status WHERE identifier = ?", (student,))
statusRow = cursor.fetchone()[0] # just get first status
if statusRow:
return statusRow
else:
return "Unbearbeitet"
def writeStatus(config, student, status):
database = getStatusTable(config)
cursor = database.cursor()
# Check if we need to create a new row first
cursor.execute("SELECT status FROM status WHERE identifier = ?", (student,))
statusRow = cursor.fetchone()[0]
if statusRow:
cursor.execute("UPDATE status SET status = ? WHERE identifier = ?", (status, student,))
else:
cursor.execute("INSERT INTO status VALUES(?, ?)", (student, status, ))
database.commit()
def getStatusTable(config):
statusDatabasePath = config("database_path")
statusDatabase = sqlite3.connect(statusDatabasePath)
cursor = statusDatabase.cursor()
cursor.execute("""CREATE TABLE IF NOT EXISTS status
(`identifier` text UNIQUE, `status` text, PRIMARY KEY (`identifier`));""")
return statusDatabase
|
from django.conf.urls.defaults import url, patterns
urlpatterns = patterns(
'lava_scheduler_app.views',
url(r'index$', 'index'))
| Make /lava-server/scheduler/ be a valid URL
| from django.conf.urls.defaults import url, patterns
urlpatterns = patterns(
'lava_scheduler_app.views',
url(r'$', 'index'))
|
from __future__ import absolute_import
from celery.utils.log import get_task_logger
from ogn.collect.celery import app
from ogn.model import Flarm
from ogn.utils import get_ddb
logger = get_task_logger(__name__)
@app.task
def update_ddb_data():
logger.info("Update ddb data.")
app.session.query(Flarm).delete()
devices = get_ddb()
logger.info("Devices: %s"%str(devices))
app.session.bulk_save_objects(devices)
app.session.commit()
return len(devices)
# TODO: Reimplement.
def import_ddb_data(filename='custom.txt'):
flarms = get_ddb(filename)
db.session.bulk_save_objects(flarms)
session.commit()
| Reimplement ddb import from file as celery task.
| from __future__ import absolute_import
from celery.utils.log import get_task_logger
from ogn.collect.celery import app
from ogn.model import Flarm
from ogn.utils import get_ddb
logger = get_task_logger(__name__)
@app.task
def update_ddb_data():
logger.info("Update ddb data.")
app.session.query(Flarm).delete()
devices = get_ddb()
logger.debug("New Devices: %s" % str(devices))
app.session.bulk_save_objects(devices)
app.session.commit()
return len(devices)
@app.task
def import_ddb_data(filename='custom.txt'):
logger.info("Import ddb data from file.")
devices = get_ddb(filename)
app.session.bulk_save_objects(devices)
app.session.commit()
return len(devices)
|
from DataModel import DataModel
from DataModelAdapter import DataModelAdapter
from Ledger import Ledger
import pickle
DATA_FILE='ledger.pickle'
def get_ledger() :
result = None
try:
with open(DATA_FILE, 'rb') as infile:
result = pickle.load(infile)
except FileNotFoundError:
result = Ledger()
def get_model() :
model = DataModel()
model.setHeaders(list(model.root.keys()))
return model
def save_ledger(ledger) :
with open(DATA_FILE, 'wb') as outfile:
pickle.dump(ledger, outfile, pickle.HIGHEST_PROTOCOL)
| Handle EOFError on pickle load
| from DataModel import DataModel
from DataModelAdapter import DataModelAdapter
from Ledger import Ledger
import pickle
DATA_FILE='ledger.pickle'
def get_ledger() :
result = None
try:
with open(DATA_FILE, 'rb') as infile:
result = pickle.load(infile)
except FileNotFoundError:
pass
except EOFError:
pass
if not result: result = Ledger()
return result
def get_model() :
model = DataModel()
model.setHeaders(list(model.root.keys()))
return model
def save_ledger(ledger) :
with open(DATA_FILE, 'wb') as outfile:
pickle.dump(ledger, outfile, pickle.HIGHEST_PROTOCOL)
|
Add (failing) Python command line parsing tests
| import px_commandline
def test_get_command_python():
assert px_commandline.get_command("python") == "python"
assert px_commandline.get_command("/apa/Python") == "Python"
assert px_commandline.get_command("python --help") == "python"
# These are inspired by Python 2.7.11 --help output
assert px_commandline.get_command("python apa.py") == "apa.py"
assert px_commandline.get_command("python /usr/bin/hej") == "hej"
assert px_commandline.get_command("python /usr/bin/hej gris --flaska") == "hej"
assert px_commandline.get_command("python -c cmd") == "python"
assert px_commandline.get_command("python -m mod") == "mod"
assert px_commandline.get_command("python -m mod --hej gris --frukt") == "mod"
assert px_commandline.get_command("Python -") == "Python"
assert px_commandline.get_command("python -W warning:spec apa.py") == "apa.py"
assert px_commandline.get_command("python -u -t -m mod") == "mod"
# Invalid command lines
assert px_commandline.get_command("python -W") == "python"
assert px_commandline.get_command("python -c") == "python"
assert px_commandline.get_command("python -m") == "python"
|
|
from .base_settings import *
import os
INSTALLED_APPS += [
'data_aggregator.apps.DataAggregatorConfig',
'webpack_loader',
]
if os.getenv('ENV') == 'localdev':
DEBUG = True
DATA_AGGREGATOR_ACCESS_GROUP = 'u_test_group'
DATA_AGGREGATOR_THREADING_ENABLED = False
RESTCLIENTS_DAO_CACHE_CLASS = None
DATA_AGGREGATOR_THREADING_ENABLED = False
else:
DATA_AGGREGATOR_ACCESS_GROUP = os.getenv('ACCESS_GROUP', '')
DATA_AGGREGATOR_THREADING_ENABLED = True
WEBPACK_LOADER = {
'DEFAULT': {
'BUNDLE_DIR_NAME': 'data_aggregator/bundles/',
'STATS_FILE': os.path.join(BASE_DIR, 'data_aggregator', 'static', 'webpack-stats.json'),
}
}
RESTCLIENTS_CANVAS_POOL_SIZE = 25
ACADEMIC_CANVAS_ACCOUNT_ID = '84378'
| Increase rest client connection pool size to 50
| from .base_settings import *
import os
INSTALLED_APPS += [
'data_aggregator.apps.DataAggregatorConfig',
'webpack_loader',
]
if os.getenv('ENV') == 'localdev':
DEBUG = True
DATA_AGGREGATOR_ACCESS_GROUP = 'u_test_group'
DATA_AGGREGATOR_THREADING_ENABLED = False
RESTCLIENTS_DAO_CACHE_CLASS = None
DATA_AGGREGATOR_THREADING_ENABLED = False
else:
DATA_AGGREGATOR_ACCESS_GROUP = os.getenv('ACCESS_GROUP', '')
DATA_AGGREGATOR_THREADING_ENABLED = True
WEBPACK_LOADER = {
'DEFAULT': {
'BUNDLE_DIR_NAME': 'data_aggregator/bundles/',
'STATS_FILE': os.path.join(BASE_DIR, 'data_aggregator', 'static', 'webpack-stats.json'),
}
}
RESTCLIENTS_CANVAS_POOL_SIZE = 50
ACADEMIC_CANVAS_ACCOUNT_ID = '84378'
|
from __future__ import unicode_literals
import unittest
import six
import spotify
class ErrorTest(unittest.TestCase):
def test_error_has_error_code(self):
error = spotify.Error(0)
self.assertEqual(error.error_code, 0)
error = spotify.Error(1)
self.assertEqual(error.error_code, 1)
def test_error_has_error_message(self):
error = spotify.Error(0)
self.assertEqual(error.message, 'No error')
self.assertIsInstance(error.message, six.text_type)
error = spotify.Error(1)
self.assertEqual(error.message, 'Invalid library version')
def test_error_has_useful_repr(self):
error = spotify.Error(0)
self.assertEqual(repr(error), b"Error(u'No error',)")
def test_error_has_useful_str(self):
error = spotify.Error(0)
self.assertEqual(str(error), 'No error')
def test_error_has_error_constants(self):
self.assertEqual(spotify.Error.OK, 0)
self.assertEqual(spotify.Error.BAD_API_VERSION, 1)
| Make Error behavior consistent across Pythons
| from __future__ import unicode_literals
import unittest
import six
import spotify
class ErrorTest(unittest.TestCase):
def test_error_has_error_code(self):
error = spotify.Error(0)
self.assertEqual(error.error_code, 0)
error = spotify.Error(1)
self.assertEqual(error.error_code, 1)
def test_error_has_useful_repr(self):
error = spotify.Error(0)
self.assertIn('No error', repr(error))
def test_error_has_useful_string_representation(self):
error = spotify.Error(0)
self.assertEqual('%s' % error, 'No error')
self.assertIsInstance('%s' % error, six.text_type)
error = spotify.Error(1)
self.assertEqual('%s' % error, 'Invalid library version')
def test_error_has_error_constants(self):
self.assertEqual(spotify.Error.OK, 0)
self.assertEqual(spotify.Error.BAD_API_VERSION, 1)
|
"""Add new Indexes for faster searching
Revision ID: 3097d57f3f0b
Revises: 4fe230f7a26e
Create Date: 2021-06-19 20:18:55.332165
"""
# revision identifiers, used by Alembic.
revision = '3097d57f3f0b'
down_revision = '4fe230f7a26e'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index(
'ix_root_authority_id',
'certificates',
['root_authority_id'],
unique=False,
postgresql_where=sqlalchemy.text("root_authority_id IS NOT NULL"))
op.create_index(
'certificate_associations_certificate_id_idx',
'certificate_associations',
['certificate_id'],
unique=False)
op.create_index(
'ix_certificates_serial',
'certificates',
['serial'],
unique=False)
def downgrade():
op.drop_index(
'ix_root_authority_id',
table_name='certificates')
op.drop_index(
'certificate_associations_certificate_id_idx',
table_name='certificate_associations')
op.drop_index(
'ix_certificates_serial',
table_name='certificates')
| Fix sqlalchemy import alias in DB migration file | """Add new Indexes for faster searching
Revision ID: 3097d57f3f0b
Revises: 4fe230f7a26e
Create Date: 2021-06-19 20:18:55.332165
"""
# revision identifiers, used by Alembic.
revision = '3097d57f3f0b'
down_revision = '4fe230f7a26e'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index(
'ix_root_authority_id',
'certificates',
['root_authority_id'],
unique=False,
postgresql_where=sa.text("root_authority_id IS NOT NULL"))
op.create_index(
'certificate_associations_certificate_id_idx',
'certificate_associations',
['certificate_id'],
unique=False)
op.create_index(
'ix_certificates_serial',
'certificates',
['serial'],
unique=False)
def downgrade():
op.drop_index(
'ix_root_authority_id',
table_name='certificates')
op.drop_index(
'certificate_associations_certificate_id_idx',
table_name='certificate_associations')
op.drop_index(
'ix_certificates_serial',
table_name='certificates')
|
"""
Manager interface.
"""
class Manager(object):
"""
Manager interface.
"""
_INTERFACE_NAME = 'org.storage.stratis1.Manager'
def __init__(self, dbus_object):
"""
Initializer.
:param dbus_object: the dbus object
"""
self._dbus_object = dbus_object
def CreatePool(self, pool_name, devices, num_devices):
"""
Create a pool.
:param str pool_name: the pool name
:param devices: the component devices
:type devices: sequence of str
"""
return self._dbus_object.CreatePool(
pool_name,
devices,
num_devices,
dbus_interface=self._INTERFACE_NAME,
)
def DestroyPool(self, pool_name):
"""
Destroy a pool.
:param str pool_name: the name of the pool
"""
return self._dbus_object.DestroyPool(
pool_name,
dbus_interface=self._INTERFACE_NAME
)
def ListPools(self):
"""
List all pools.
"""
return self._dbus_object.ListPools(dbus_interface=self._INTERFACE_NAME)
| Use Properties interface to get Manager properties.
Signed-off-by: mulhern <[email protected]>
| """
Manager interface.
"""
from ._properties import Properties
class Manager(object):
"""
Manager interface.
"""
_INTERFACE_NAME = 'org.storage.stratis1.Manager'
def __init__(self, dbus_object):
"""
Initializer.
:param dbus_object: the dbus object
"""
self._dbus_object = dbus_object
def CreatePool(self, pool_name, devices, num_devices):
"""
Create a pool.
:param str pool_name: the pool name
:param devices: the component devices
:type devices: sequence of str
"""
return self._dbus_object.CreatePool(
pool_name,
devices,
num_devices,
dbus_interface=self._INTERFACE_NAME,
)
def DestroyPool(self, pool_name):
"""
Destroy a pool.
:param str pool_name: the name of the pool
"""
return self._dbus_object.DestroyPool(
pool_name,
dbus_interface=self._INTERFACE_NAME
)
def ListPools(self):
"""
List all pools.
"""
return self._dbus_object.ListPools(dbus_interface=self._INTERFACE_NAME)
@property
def Version(self):
"""
Stratisd Version getter.
:rtype: String
"""
return Properties(self._dbus_object).Get(
self._INTERFACE_NAME,
'Version'
)
@property
def LogLevel(self):
"""
Stratisd LogLevel getter.
:rtype: String
"""
return Properties(self._dbus_object).Get(
self._INTERFACE_NAME,
'LogLevel'
)
@LogLevel.setter
def LogLevel(self, value):
"""
Stratisd LogLevel setter.
:param str value: the value to set
"""
return Properties(self._dbus_object).Set(
self._INTERFACE_NAME,
'LogLevel',
value
)
|
class Config:
DEBUG = False
TESTING = False
REDIS = {
'host': 'localhost',
'port': 6379,
'db': 0,
'password': None
}
REDIS_KEY_EXPIRE = 604800 # a week in seconds
IMAGE_DIR = './resources'
LOG_FILE = 'app.log'
class DevelopmentConfig(Config):
DEBUG = True
def load_config(app, flags):
if 'dev' in flags:
app.config.from_object('config.DevelopmentConfig')
else:
app.config.from_object('config.Config')
def rotating_handler(filename):
if filename.startswith('~'):
filename = filename.replace('~', os.path.expanduser('~'))
handler = RotatingFileHandler(filename, maxBytes=5242880, backupCount=2)
formatter = Formatter('%(asctime)s [%(levelname)s] %(message)s', '%Y-%m-%d %H:%M:%S')
handler.setFormatter(formatter)
return handler
def setup_logging(app):
if not app.debug:
from logging import Formatter
from logging.handlers import RotatingFileHandler
import logging
filename = app.config['LOG_FILE']
handler = rotating_handler(filename)
handler.setLevel(logging.INFO)
app.logger.addHandler(handler)
| Move imports to top of file
| from logging import Formatter
from logging.handlers import RotatingFileHandler
import logging
class Config:
DEBUG = False
TESTING = False
REDIS = {
'host': 'localhost',
'port': 6379,
'db': 0,
'password': None
}
REDIS_KEY_EXPIRE = 604800 # a week in seconds
IMAGE_DIR = './resources'
LOG_FILE = 'app.log'
class DevelopmentConfig(Config):
DEBUG = True
def load_config(app, flags):
if 'dev' in flags:
app.config.from_object('config.DevelopmentConfig')
else:
app.config.from_object('config.Config')
def rotating_handler(filename):
if filename.startswith('~'):
filename = filename.replace('~', os.path.expanduser('~'))
handler = RotatingFileHandler(filename, maxBytes=5242880, backupCount=2)
formatter = Formatter('%(asctime)s [%(levelname)s] %(message)s', '%Y-%m-%d %H:%M:%S')
handler.setFormatter(formatter)
return handler
def setup_logging(app):
if not app.debug:
filename = app.config['LOG_FILE']
handler = rotating_handler(filename)
handler.setLevel(logging.INFO)
app.logger.addHandler(handler)
|
from __future__ import unicode_literals
from django.db import models
from django.utils.encoding import python_2_unicode_compatible
from bazaar.listings.models import Publishing, Store
@python_2_unicode_compatible
class Order(models.Model):
ORDER_PENDING = 0
ORDER_COMPLETED = 1
ORDER_STATUS_CHOICES = (
(ORDER_PENDING, "Pending"),
(ORDER_COMPLETED, "Completed"),
)
external_id = models.CharField(max_length=256)
store = models.ForeignKey(Store)
publishing = models.ForeignKey(Publishing, null=True, blank=True)
quantity = models.IntegerField(default=1)
status = models.CharField(max_length=50, choices=ORDER_STATUS_CHOICES, default=ORDER_PENDING)
def __str__(self):
return "Order %s from %s" % (self.external_id, self.store)
| Order status changed to integer
| from __future__ import unicode_literals
from django.db import models
from django.utils.encoding import python_2_unicode_compatible
from bazaar.listings.models import Publishing, Store
@python_2_unicode_compatible
class Order(models.Model):
ORDER_PENDING = 0
ORDER_COMPLETED = 1
ORDER_STATUS_CHOICES = (
(ORDER_PENDING, "Pending"),
(ORDER_COMPLETED, "Completed"),
)
external_id = models.CharField(max_length=256)
store = models.ForeignKey(Store)
publishing = models.ForeignKey(Publishing, null=True, blank=True)
quantity = models.IntegerField(default=1)
status = models.IntegerField(max_length=50, choices=ORDER_STATUS_CHOICES, default=ORDER_PENDING)
def __str__(self):
return "Order %s from %s" % (self.external_id, self.store)
|
# -*- coding: utf-8 -*-
'''
Classification of Salt exit codes. These are intended to augment
universal exit codes (found in Python's `os` module with the `EX_`
prefix or in `sysexits.h`).
'''
# Too many situations use "exit 1" - try not to use it when something
# else is more appropriate.
EX_GENERIC = 1
# Salt SSH "Thin" deployment failures
EX_THIN_PYTHON_OLD = 10
EX_THIN_DEPLOY = 11
EX_THIN_CHECKSUM = 12
EX_MOD_DEPLOY = 13
# The os.EX_* exit codes are Unix only so in the interest of cross-platform
# compatiblility define them explicitly here.
#
# These constants are documented here:
# https://docs.python.org/2/library/os.html#os.EX_OK
EX_OK = 0
EX_NOUSER = 67
EX_UNAVAILABLE = 69
EX_CANTCREAT = 73
EX_SOFTWARE = 70
EX_USAGE = 64
| Add Salt specific exit code
| # -*- coding: utf-8 -*-
'''
Classification of Salt exit codes. These are intended to augment
universal exit codes (found in Python's `os` module with the `EX_`
prefix or in `sysexits.h`).
'''
# Too many situations use "exit 1" - try not to use it when something
# else is more appropriate.
EX_GENERIC = 1
# Salt SSH "Thin" deployment failures
EX_THIN_PYTHON_OLD = 10
EX_THIN_DEPLOY = 11
EX_THIN_CHECKSUM = 12
EX_MOD_DEPLOY = 13
# The os.EX_* exit codes are Unix only so in the interest of cross-platform
# compatiblility define them explicitly here.
#
# These constants are documented here:
# https://docs.python.org/2/library/os.html#os.EX_OK
EX_OK = 0
EX_NOUSER = 67
EX_UNAVAILABLE = 69
EX_CANTCREAT = 73
EX_SOFTWARE = 70
EX_USAGE = 64
# The Salt specific exit codes are defined below:
# SALT_BUILD_FAIL is used when salt fails to build something, like a container
SALT_BUILD_FAIL = 101
|
# -*- coding: utf-8 -*-
"""Example Dropbox local settings file. Copy this file to local.py and change
these settings.
"""
# Get an app key and secret at https://www.dropbox.com/developers/apps
DROPBOX_KEY = 'changeme'
DROPBOX_SECRET = 'changeme' | Fix coding style for travis ci build.
| # -*- coding: utf-8 -*-
"""Example Dropbox local settings file. Copy this file to local.py and change
these settings.
"""
# Get an app key and secret at https://www.dropbox.com/developers/apps
DROPBOX_KEY = 'changeme'
DROPBOX_SECRET = 'changeme'
|
Add some necessary things to run the app
| #!/usr/bin/env python
# coding=utf-8
#
# Copyright 2015 cc98.org
import sys
reload(sys)
sys.setdefaultencoding("utf-8")
import os.path
import re
import memcache
import torndb
import tornado.httpserver
import tornado.ioloop
import tornado.options
import tornado.web
from tornado.options import define, options
from jinja2 import Environment, FileSystemLoader
# Define a new command line option
define("port", default = 80, type = int, help = "run on the given port")
define("mysql_host", default = "mysql_host", help = "community database host")
define("mysql_database", default = "mysql_database", help = "community database name")
define("mysql_user", default = "mysql_db_user", help = "community database user")
define("mysql_pwd", default = "mysql_db_pwd", help = "community database pwd")
class Application(tornado.web.Application):
def __init__(self):
settings = dict(
)
handlers = [
(),
]
tornado.web.Application.__init__(self, handlers, settings)
# Have one global connection to the blog DB across all the handlers
self.db = torndb.Connection(
)
# Have one global session controller
# Have one global memcache controller
self.mc = memcache.Client(["127.0.0.1:11211"]])
def main():
tornado.options.parse_command_line()
httpserver = tornado.httpserver.HTTPServer(Application())
httpserver.listen(options.port)
tornado.ioloop.IOLoop.instance().start()
if __name__ = "__main__":
main()
|
|
# Copyright (c) 2020, CNRS
# Authors: Guilhem Saurel <[email protected]>
import os
import subprocess
import time
import hpp.corbaserver
try:
from subprocess import DEVNULL, run
except ImportError: # Python2 fallback
DEVNULL = os.open(os.devnull, os.O_RDWR)
def run(*args):
subprocess.Popen(*args).wait()
class ServerManager:
"""A context to ensure a server is running."""
def __init__(self, server="hppcorbaserver"):
self.server = server
run(["killall", self.server])
def __enter__(self):
"""Run the server in background
stdout and stderr outputs of the child process are redirected to devnull.
preexec_fn is used to ignore ctrl-c signal send to the main script
(otherwise they are forwarded to the child process)
"""
self.process = subprocess.Popen(
self.server, stdout=DEVNULL, stderr=DEVNULL, preexec_fn=os.setpgrp
)
# give it some time to start
time.sleep(3)
def __exit__(self, exc_type, exc_value, exc_traceback):
tool = hpp.corbaserver.tools.Tools()
tool.shutdown()
self.process.communicate()
| Add sleep after shutting down hppcorbaserver in ServerManager
Co-authored-by: Joseph Mirabel <[email protected]> | # Copyright (c) 2020, CNRS
# Authors: Guilhem Saurel <[email protected]>
import os
import subprocess
import time
import hpp.corbaserver
try:
from subprocess import DEVNULL, run
except ImportError: # Python2 fallback
DEVNULL = os.open(os.devnull, os.O_RDWR)
def run(*args):
subprocess.Popen(*args).wait()
class ServerManager:
"""A context to ensure a server is running."""
def __init__(self, server="hppcorbaserver"):
self.server = server
run(["killall", self.server])
def __enter__(self):
"""Run the server in background
stdout and stderr outputs of the child process are redirected to devnull.
preexec_fn is used to ignore ctrl-c signal send to the main script
(otherwise they are forwarded to the child process)
"""
self.process = subprocess.Popen(
self.server, stdout=DEVNULL, stderr=DEVNULL, preexec_fn=os.setpgrp
)
# give it some time to start
time.sleep(3)
def __exit__(self, exc_type, exc_value, exc_traceback):
tool = hpp.corbaserver.tools.Tools()
tool.shutdown()
# Give some time to HPP to properly shutdown.
time.sleep(1)
# Once HPP process is stopped, this removes the defunct process.
self.process.communicate()
|
import pytest
import socket
import struct
from threading import Thread
from server import app
@pytest.fixture(autouse=True, scope="module")
def server(request):
t = Thread(target=app.serve_forever)
t.start()
def fin():
app.shutdown()
app.server_close()
t.join()
request.addfinalizer(fin)
return app
@pytest.yield_fixture
def sock(server):
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.connect(server.socket.getsockname())
yield sock
sock.close()
@pytest.fixture
def mbap():
return struct.pack('>HHHB', 0, 0, 6, 1)
| Fix import error in test suite so tests pass in Python 3.3+.
| import pytest
import socket
import struct
from threading import Thread
from .server import app
@pytest.fixture(autouse=True, scope="module")
def server(request):
t = Thread(target=app.serve_forever)
t.start()
def fin():
app.shutdown()
app.server_close()
t.join()
request.addfinalizer(fin)
return app
@pytest.yield_fixture
def sock(server):
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.connect(server.socket.getsockname())
yield sock
sock.close()
@pytest.fixture
def mbap():
return struct.pack('>HHHB', 0, 0, 6, 1)
|
#!/usr/bin/env python
import os
from tempdir import TempDir
import pytest
boto = pytest.importorskip('boto')
from simplekv.net.botostore import BotoStore
from basic_store import BasicStore
from url_store import UrlStore
from bucket_manager import boto_credentials, boto_bucket
@pytest.fixture(params=boto_credentials,
ids=[c['access_key'] for c in boto_credentials])
def credentials(request):
return request.param
@pytest.yield_fixture()
def bucket(credentials):
with boto_bucket(**credentials) as bucket:
yield bucket
class TestBotoStorage(BasicStore, UrlStore):
@pytest.fixture(params=['', '/test-prefix'])
def prefix(self, request):
return request.param
@pytest.fixture
def store(self, bucket, prefix):
return BotoStore(bucket, prefix)
def test_get_filename_nonexistant(self, store, key):
# NOTE: boto misbehaves here and tries to erase the target file
# the parent tests use /dev/null, which you really should not try
# to os.remove!
with TempDir() as tmpdir:
with pytest.raises(KeyError):
store.get_file(key, os.path.join(tmpdir, 'a'))
| Fix another boto /dev/null testing error.
| #!/usr/bin/env python
import os
from tempdir import TempDir
import pytest
boto = pytest.importorskip('boto')
from simplekv.net.botostore import BotoStore
from basic_store import BasicStore
from url_store import UrlStore
from bucket_manager import boto_credentials, boto_bucket
@pytest.fixture(params=boto_credentials,
ids=[c['access_key'] for c in boto_credentials])
def credentials(request):
return request.param
@pytest.yield_fixture()
def bucket(credentials):
with boto_bucket(**credentials) as bucket:
yield bucket
class TestBotoStorage(BasicStore, UrlStore):
@pytest.fixture(params=['', '/test-prefix'])
def prefix(self, request):
return request.param
@pytest.fixture
def store(self, bucket, prefix):
return BotoStore(bucket, prefix)
def test_get_filename_nonexistant(self, store, key):
# NOTE: boto misbehaves here and tries to erase the target file
# the parent tests use /dev/null, which you really should not try
# to os.remove!
with TempDir() as tmpdir:
with pytest.raises(KeyError):
store.get_file(key, os.path.join(tmpdir, 'a'))
def test_key_error_on_nonexistant_get_filename(self, store, key):
with TempDir() as tmpdir:
with pytest.raises(KeyError):
store.get_file(key, os.path.join(tmpdir, 'a'))
|
import re
import sqlite3
def do_karma(bot, user, channel, karma):
if karma[1] == '++':
k = 1
else:
k = -1
conn = sqlite3.connect('karma.db')
c = conn.cursor()
t = (karma[0],)
c.execute('select * from karma where word=?', t)
res = c.fetchone()
if res != None:
u = k + res[2]
q = (u,karma[0],)
c.execute('update karma set karma = ? where word=?', q)
else:
u = k
q = (karma[0],u,)
c.execute('insert into karma (word, karma) VALUES (?,?)',q)
conn.commit()
return bot.say(channel, "Karma for %s is now %s" % (karma[0], u))
def handle_privmsg(bot, user, reply, msg):
"""Grab karma changes from the messages and handle them"""
m = re.findall('([a-zA-Z0-9.-_]*)(\+\+|\-\-)', msg)
if len(m) == 0: return None
for k in m:
do_karma(bot, user, reply, k)
return
| Add .tolower() when adding to DB to avoid potential issues
| import re
import sqlite3
def do_karma(bot, user, channel, karma):
if karma[1] == '++':
k = 1
else:
k = -1
conn = sqlite3.connect('karma.db')
c = conn.cursor()
t = (karma[0],)
c.execute('select * from karma where word=?', t)
res = c.fetchone()
if res != None:
u = k + res[2]
q = (u,karma[0].lower(),)
c.execute('update karma set karma = ? where word=?', q)
else:
u = k
q = (karma[0].lower(),u,)
c.execute('insert into karma (word, karma) VALUES (?,?)',q)
conn.commit()
return bot.say(channel, "Karma for %s is now %s" % (karma[0], u))
def handle_privmsg(bot, user, reply, msg):
"""Grab karma changes from the messages and handle them"""
m = re.findall('([a-zA-Z0-9.-_]*)(\+\+|\-\-)', msg)
if len(m) == 0: return None
for k in m:
do_karma(bot, user, reply, k)
return
|
"""Python script Exporter class"""
#-----------------------------------------------------------------------------
# Copyright (c) 2013, the IPython Development Team.
#
# Distributed under the terms of the Modified BSD License.
#
# The full license is in the file COPYING.txt, distributed with this software.
#-----------------------------------------------------------------------------
#-----------------------------------------------------------------------------
# Imports
#-----------------------------------------------------------------------------
from IPython.utils.traitlets import Unicode
from .templateexporter import TemplateExporter
#-----------------------------------------------------------------------------
# Classes
#-----------------------------------------------------------------------------
class PythonExporter(TemplateExporter):
"""
Exports a Python code file.
"""
file_extension = Unicode(
'py', config=True,
help="Extension of the file that should be written to disk")
def _raw_mimetype_default(self):
return 'application/x-python'
| Add MIME types to nbconvert exporters
| """Python script Exporter class"""
#-----------------------------------------------------------------------------
# Copyright (c) 2013, the IPython Development Team.
#
# Distributed under the terms of the Modified BSD License.
#
# The full license is in the file COPYING.txt, distributed with this software.
#-----------------------------------------------------------------------------
#-----------------------------------------------------------------------------
# Imports
#-----------------------------------------------------------------------------
from IPython.utils.traitlets import Unicode
from .templateexporter import TemplateExporter
#-----------------------------------------------------------------------------
# Classes
#-----------------------------------------------------------------------------
class PythonExporter(TemplateExporter):
"""
Exports a Python code file.
"""
file_extension = Unicode(
'py', config=True,
help="Extension of the file that should be written to disk")
def _raw_mimetype_default(self):
return 'application/x-python'
mime_type = Unicode('text/x-python', config=True,
help="MIME type of the result file, for HTTP response headers."
)
|
# -*- coding: utf-8 -*-
import os
import sphinx_rtd_theme
import sys
REPO_DIR = os.path.dirname(os.path.dirname(__file__))
sys.path.append(REPO_DIR)
project = 'Ichnaea'
copyright = '2013-2019, Mozilla'
# The short X.Y version.
version = '2.0'
# The full version, including alpha/beta/rc tags.
release = '2.0'
autoclass_content = 'class'
exclude_patterns = ['build/html/README.rst', '.DS_Store', 'Thumbs.db']
html_static_path = []
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
master_doc = 'index'
modindex_common_prefix = ['ichnaea.']
pygments_style = 'sphinx'
source_suffix = '.rst'
templates_path = ['_templates']
extensions = [
'sphinx.ext.linkcode',
'everett.sphinxext',
]
def linkcode_resolve(domain, info):
if domain != 'py':
return None
if not info['module']:
return None
filename = info['module'].replace('.', '/')
return "https://github.com/mozilla/ichnaea/tree/master/%s.py" % filename
| Add mock for shapely module
Adding a mock for the shapely module allows ReadTheDocs to build the
docs even though Shapely isn't installed.
| # -*- coding: utf-8 -*-
import os
import sphinx_rtd_theme
import sys
from unittest import mock
# Add repository root so we can import ichnaea things
REPO_DIR = os.path.dirname(os.path.dirname(__file__))
sys.path.append(REPO_DIR)
# Fake the shapely module so things will import
sys.modules['shapely'] = mock.MagicMock()
project = 'Ichnaea'
copyright = '2013-2019, Mozilla'
# The short X.Y version.
version = '2.0'
# The full version, including alpha/beta/rc tags.
release = '2.0'
autoclass_content = 'class'
exclude_patterns = ['build/html/README.rst', '.DS_Store', 'Thumbs.db']
html_static_path = []
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
master_doc = 'index'
modindex_common_prefix = ['ichnaea.']
pygments_style = 'sphinx'
source_suffix = '.rst'
templates_path = ['_templates']
extensions = [
'sphinx.ext.linkcode',
'everett.sphinxext',
]
def linkcode_resolve(domain, info):
if domain != 'py':
return None
if not info['module']:
return None
filename = info['module'].replace('.', '/')
return "https://github.com/mozilla/ichnaea/tree/master/%s.py" % filename
|
"""Linear referencing
"""
from shapely.topology import Delegating
class LinearRefBase(Delegating):
def _validate_line(self, ob):
super(LinearRefBase, self)._validate(ob)
try:
assert ob.geom_type in ['LineString', 'MultiLineString']
except AssertionError:
raise TypeError("Only linear types support this operation")
class ProjectOp(LinearRefBase):
def __call__(self, this, other):
self._validate_line(this)
self._validate(other)
return self.fn(this._geom, other._geom)
class InterpolateOp(LinearRefBase):
def __call__(self, this, distance):
self._validate_line(this)
return self.fn(this._geom, distance)
| Allow linear referencing on rings.
Closes #286.
Eliminating the assert is good for optimization reasons, too.
| """Linear referencing
"""
from shapely.topology import Delegating
class LinearRefBase(Delegating):
def _validate_line(self, ob):
super(LinearRefBase, self)._validate(ob)
if not ob.geom_type in ['LinearRing', 'LineString', 'MultiLineString']:
raise TypeError("Only linear types support this operation")
class ProjectOp(LinearRefBase):
def __call__(self, this, other):
self._validate_line(this)
self._validate(other)
return self.fn(this._geom, other._geom)
class InterpolateOp(LinearRefBase):
def __call__(self, this, distance):
self._validate_line(this)
return self.fn(this._geom, distance)
|
# -*- coding: utf-8 -*-
'''
Manage PagerDuty users.
Example:
.. code-block:: yaml
ensure bruce test user 1:
pagerduty.user_present:
- name: 'Bruce TestUser1'
- email: [email protected]
- requester_id: P1GV5NT
'''
def __virtual__():
'''
Only load if the pygerduty module is available in __salt__
'''
return 'pagerduty_user' if 'pagerduty_util.get_resource' in __salt__ else False
def present(profile='pagerduty', subdomain=None, api_key=None, **kwargs):
'''
Ensure pagerduty user exists.
Arguments match those suppored by
https://developer.pagerduty.com/documentation/rest/users/create.
'''
return __salt__['pagerduty_util.resource_present']('users',
['email', 'name', 'id'],
None,
profile,
subdomain,
api_key,
**kwargs)
def absent(profile='pagerduty', subdomain=None, api_key=None, **kwargs):
'''
Ensure pagerduty user does not exist.
Name can be pagerduty id, email address, or user name.
'''
return __salt__['pagerduty_util.resource_absent']('users',
['email', 'name', 'id'],
profile,
subdomain,
api_key,
**kwargs)
| Fix typo an suppored -> supported
| # -*- coding: utf-8 -*-
'''
Manage PagerDuty users.
Example:
.. code-block:: yaml
ensure bruce test user 1:
pagerduty.user_present:
- name: 'Bruce TestUser1'
- email: [email protected]
- requester_id: P1GV5NT
'''
def __virtual__():
'''
Only load if the pygerduty module is available in __salt__
'''
return 'pagerduty_user' if 'pagerduty_util.get_resource' in __salt__ else False
def present(profile='pagerduty', subdomain=None, api_key=None, **kwargs):
'''
Ensure pagerduty user exists.
Arguments match those supported by
https://developer.pagerduty.com/documentation/rest/users/create.
'''
return __salt__['pagerduty_util.resource_present']('users',
['email', 'name', 'id'],
None,
profile,
subdomain,
api_key,
**kwargs)
def absent(profile='pagerduty', subdomain=None, api_key=None, **kwargs):
'''
Ensure pagerduty user does not exist.
Name can be pagerduty id, email address, or user name.
'''
return __salt__['pagerduty_util.resource_absent']('users',
['email', 'name', 'id'],
profile,
subdomain,
api_key,
**kwargs)
|
import os
PROJECT_DIR = os.path.dirname(os.path.abspath(__file__))
SCREENSHOT_PATH = None
OPENGL = {
"version": (3, 3),
}
WINDOW = {
"class": "demosys.context.pyqt.Window",
"size": (1920, 1080),
"aspect_ratio": 16 / 9,
"fullscreen": False,
"resizable": False,
"title": "Examples",
"vsync": True,
"cursor": True,
"samples": 4,
}
HEADLESS_DURATION = 100.0
ROCKET = {
"mode": "editor",
"rps": 24,
"project": None,
"files": None,
}
| Reduce default resoution for example project
| import os
PROJECT_DIR = os.path.dirname(os.path.abspath(__file__))
SCREENSHOT_PATH = None
OPENGL = {
"version": (3, 3),
}
WINDOW = {
"class": "demosys.context.pyqt.Window",
"size": (1280, 720),
"aspect_ratio": 16 / 9,
"fullscreen": False,
"resizable": False,
"title": "Examples",
"vsync": True,
"cursor": True,
"samples": 4,
}
HEADLESS_DURATION = 100.0
ROCKET = {
"mode": "editor",
"rps": 24,
"project": None,
"files": None,
}
|
self.gauge( ... ) # Sample a gauge metric
self.increment( ... ) # Increment a counter metric
self.decrement( ... ) # Decrement a counter metric
self.histogram( ... ) # Sample a histogram metric
self.rate( ... ) # Sample a point, with the rate calculated at the end of the check
self.count( ... ) # Sample a raw count metric
self.monotonic_count( ... ) # Sample an increasing counter metric
| Revert "Document AgentCheck count and monotonic_count methods"
This reverts commit e731c3a4a8590f5cddd23fd2f9af265749f08a38.
| self.gauge( ... ) # Sample a gauge metric
self.increment( ... ) # Increment a counter metric
self.decrement( ... ) # Decrement a counter metric
self.histogram( ... ) # Sample a histogram metric
self.rate( ... ) # Sample a point, with the rate calculated at the end of the check
|
from django.apps import AppConfig
class ModuleConfig(AppConfig):
name = 'adhocracy4.modules'
label = 'a4modules'
| Fix plural in modules app config
| from django.apps import AppConfig
class ModulesConfig(AppConfig):
name = 'adhocracy4.modules'
label = 'a4modules'
|
# TODO(niboshi): Currently this file is just a placeholder for future migration of routine tests from test_array.py.
# In order to do that, we need to somehow share check logics/test parameters with test_array.py.
| Add dummy test for flake8
| # TODO(niboshi): Currently this file is just a placeholder for future migration of routine tests from test_array.py.
# In order to do that, we need to somehow share check logics/test parameters with test_array.py.
def test_dummy():
pass
|
# Copyright 2014 Dave Kludt
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from datetime import datetime
from dateutil import tz
from dateutil.relativedelta import relativedelta
UTC = tz.tzutc()
class Region:
def __init__(self, data):
self.name = data.get('name').title()
self.abbreviation = data.get('abbreviation').upper()
self.active = bool(data.get('active'))
class Account:
def __init__(self, data):
self.account_number = data.get('account_number')
self.cache_expiration = self.set_expiration()
self.host_servers = data.get('host_servers')
self.public_zones = data.get('public_zones')
self.region = data.get('region').lower()
self.servers = data.get('servers')
self.lookup_type = data.get('lookup_type')
def set_expiration(self):
return datetime.now(UTC) + relativedelta(days=1)
| Update model for CBS host and volume information
| # Copyright 2014 Dave Kludt
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from datetime import datetime
from dateutil import tz
from dateutil.relativedelta import relativedelta
UTC = tz.tzutc()
class Region:
def __init__(self, data):
self.name = data.get('name').title()
self.abbreviation = data.get('abbreviation').upper()
self.active = bool(data.get('active'))
class Account:
def __init__(self, data):
self.account_number = data.get('account_number')
self.cache_expiration = self.set_expiration()
self.host_servers = data.get('host_servers')
self.public_zones = data.get('public_zones')
self.region = data.get('region').lower()
self.servers = data.get('servers')
self.volumes = data.get('volumes')
self.cbs_hosts = data.get('cbs_hosts')
self.lookup_type = data.get('lookup_type')
def set_expiration(self):
return datetime.now(UTC) + relativedelta(days=1)
|
class ControllerError(StandardError): pass
class InterpreterError(StandardError): pass
| Use Exception instead of StandardError
Python 3 doesn't have StandardError...
| class CytoplasmError(Exception): pass
class ControllerError(CytoplasmError): pass
class InterpreterError(CytoplasmError): pass
|
'''
Holds static data components, like the palette
'''
def msg(msg, logfile='console_log.txt'):
'''
Send a message to a logfile, defaults to console_log.txt.
This is useful to replace a print statement since curses does put
a bit of a damper on this
'''
with open(logfile, 'a+') as fp_:
fp_.write(str(msg))
def get_palette(theme='std'):
'''
Return the preferred palette theme
Themes:
std
The standard theme used by the console
'''
if theme == 'bright':
return [
('banner', 'white', 'dark blue')
]
else:
return [
('banner', 'white', 'dark blue')
]
| Add convenience function to load in some test data
| '''
Holds static data components, like the palette
'''
import pprint
def tree_seed():
return {'jids': [
{'_|-76789876543456787654': [{'localhost': {'return': True}},
{'otherhost': {'return': True}}],},
{'_|-76789876543456787655': [{'localhost': {'return': True}},
{'otherhost': {'return': True}}],},
],
}
def msg(msg, logfile='console_log.txt'):
'''
Send a message to a logfile, defaults to console_log.txt.
This is useful to replace a print statement since curses does put
a bit of a damper on this
'''
with open(logfile, 'a+') as fp_:
fp_.write('{0}\n'.format(pprint.pformat(msg)))
def get_palette(theme='std'):
'''
Return the preferred palette theme
Themes:
std
The standard theme used by the console
'''
if theme == 'bright':
return [
('banner', 'white', 'dark blue')
]
else:
return [
('banner', 'white', 'dark blue')
]
|
#! /usr/bin/env python
"""Test script for the dumbdbm module
Original by Roger E. Masse
"""
# XXX This test is a disgrace. It doesn't test that it works.
import dumbdbm as dbm
from dumbdbm import error
from test_support import verbose
filename = '/tmp/delete_me'
d = dbm.open(filename, 'c')
d['a'] = 'b'
d['12345678910'] = '019237410982340912840198242'
d.keys()
if d.has_key('a'):
if verbose:
print 'Test dbm keys: ', d.keys()
d.close()
d = dbm.open(filename, 'r')
d.close()
d = dbm.open(filename, 'w')
d.close()
d = dbm.open(filename, 'n')
d.close()
import os
def rm(fn):
try:
os.unlink(fn)
except os.error:
pass
rm(filename + '.dir')
rm(filename + '.dat')
rm(filename + '.bak')
| Use a saner test filename, to work on Windows.
| #! /usr/bin/env python
"""Test script for the dumbdbm module
Original by Roger E. Masse
"""
# XXX This test is a disgrace. It doesn't test that it works.
import dumbdbm as dbm
from dumbdbm import error
from test_support import verbose, TESTFN as filename
d = dbm.open(filename, 'c')
d['a'] = 'b'
d['12345678910'] = '019237410982340912840198242'
d.keys()
if d.has_key('a'):
if verbose:
print 'Test dbm keys: ', d.keys()
d.close()
d = dbm.open(filename, 'r')
d.close()
d = dbm.open(filename, 'w')
d.close()
d = dbm.open(filename, 'n')
d.close()
import os
def rm(fn):
try:
os.unlink(fn)
except os.error:
pass
rm(filename + '.dir')
rm(filename + '.dat')
rm(filename + '.bak')
|
from twisted.plugin import IPlugin
from txircd.module_interface import IModuleData, ModuleData
from zope.interface import implements
class MultiPrefix(ModuleData):
implements(IPlugin, IModuleData)
name = "MultiPrefix"
def actions(self):
return [ ("channelstatuses", 2, self.allStatuses),
("capabilitylist", 10, self.addCapability) ]
def load(self):
if "cap-add" in self.ircd.moduleFunctionCache:
self.ircd.moduleFunctionCache["cap-add"]("multi-prefix")
def unload(self):
if "cap-add" in self.ircd.moduleFunctionCache:
self.ircd.moduleFunctionCache["cap-add"]("multi-prefix")
def addCapability(self, capList):
capList.append("multi-prefix")
def allStatuses(self, channel, user, requestingUser):
if "capabilities" not in requestingUser.cache or "multi-prefix" not in requestingUser.cache["capabilities"]:
return None
if user not in channel.users:
return ""
statusList = []
for status in channel.users[user]["status"]:
statusList.append(self.ircd.channelStatuses[status][0])
return "".join(statusList)
multiPrefix = MultiPrefix() | Reduce undoings of multi-prefix on users
| from twisted.plugin import IPlugin
from txircd.module_interface import IModuleData, ModuleData
from zope.interface import implements
class MultiPrefix(ModuleData):
implements(IPlugin, IModuleData)
name = "MultiPrefix"
def actions(self):
return [ ("channelstatuses", 2, self.allStatuses),
("capabilitylist", 10, self.addCapability) ]
def load(self):
if "unloading-multi-prefix" in self.ircd.dataCache:
del self.ircd.dataCache["unloading-multi-prefix"]
return
if "cap-add" in self.ircd.functionCache:
self.ircd.functionCache["cap-add"]("multi-prefix")
def unload(self):
self.ircd.dataCache["unloading-multi-prefix"] = True
def fullUnload(self):
del self.ircd.dataCache["unloading-multi-prefix"]
if "cap-del" in self.ircd.functionCache:
self.ircd.functionCache["cap-del"]("multi-prefix")
def addCapability(self, capList):
capList.append("multi-prefix")
def allStatuses(self, channel, user, requestingUser):
if "capabilities" not in requestingUser.cache or "multi-prefix" not in requestingUser.cache["capabilities"]:
return None
if user not in channel.users:
return ""
statusList = []
for status in channel.users[user]["status"]:
statusList.append(self.ircd.channelStatuses[status][0])
return "".join(statusList)
multiPrefix = MultiPrefix() |
import sys, random, time
boardSize = (10,10)
while True: foo, bar, baz, globals()['board'] = None if globals().get('board') is None else [
(
[sys.stdout.write('X' if cell else ' ') for cell in row],
sys.stdout.write('\n')
) for row in board
], time.sleep(1), sys.stdout.write('==============\n'), [
[random.random() < 0.5 for i in range(boardSize[0])] for j in range(boardSize[1])
] if 'board' not in globals() else [
map(
lambda z: (z[1] in (2,3) and board[y][z[0]]) or z[1]==3,
[
(
x,
sum(
[
int(
y0 in range(len(board)) and x0 in range(len(board[y0])) and board[y0][x0]
) for x0,y0 in (
(x - 1, y - 1),
(x, y -1),
(x + 1, y - 1),
(x - 1, y),
(x + 1, y),
(x - 1, y + 1),
(x, y + 1),
(x + 1, y + 1)
)
]
)
) for x in range(len(board[y]))]
) for y in range(len(board))
]
| Make separator line match width of board
| import sys, random, time
boardSize = (10,10)
while True: foo, bar, baz, globals()['board'] = None if globals().get('board') is None else [
(
[sys.stdout.write('X' if cell else ' ') for cell in row],
sys.stdout.write('\n')
) for row in board
], time.sleep(1), sys.stdout.write('=' * boardSize[0] +'\n'), [
[random.random() < 0.5 for i in range(boardSize[0])] for j in range(boardSize[1])
] if 'board' not in globals() else [
map(
lambda z: (z[1] in (2,3) and board[y][z[0]]) or z[1]==3,
[
(
x,
sum(
[
int(
y0 in range(len(board)) and x0 in range(len(board[y0])) and board[y0][x0]
) for x0,y0 in (
(x - 1, y - 1),
(x, y -1),
(x + 1, y - 1),
(x - 1, y),
(x + 1, y),
(x - 1, y + 1),
(x, y + 1),
(x + 1, y + 1)
)
]
)
) for x in range(len(board[y]))]
) for y in range(len(board))
]
|
from .. import idx
import os
def test__count_dimensions():
yield check__count_dimensions, 9, 0
yield check__count_dimensions, [1, 2], 1
yield check__count_dimensions, [[1, 2], [3, 6, 2]], 2
yield check__count_dimensions, [[[1,2], [2]]], 3
def check__count_dimensions(lst, i):
assert idx._count_dimensions(lst) == i
# these two are equivalent according to the format on http://yann.lecun.com/exdb/mnist/
_somelist = [[1, 2], [3, 4]]
_somebytes = b'\x00\x00\x0C\x02' + b'\x01\x02\x03\x04'
def test_list_to_idx():
data = idx.list_to_idx(_somelist, 'i')
assert data == _somebytes
def test_idx_to_list():
lst = idx.idx_to_list(_somebytes)
assert lst == _somelist
| Fix issue where idx test uses wrong bytes object
Forgot to include the sizes of each dimension
| from .. import idx
import os
def test__count_dimensions():
yield check__count_dimensions, 9, 0
yield check__count_dimensions, [1, 2], 1
yield check__count_dimensions, [[1, 2], [3, 6, 2]], 2
yield check__count_dimensions, [[[1,2], [2]]], 3
def check__count_dimensions(lst, i):
assert idx._count_dimensions(lst) == i
# these two are equivalent according to the format on http://yann.lecun.com/exdb/mnist/
_somelist = [[1, 2], [3, 4]]
def _get_somebytes():
header = b'\x00\x00\x0C\x02'
dimensionsizes = b'\x00\x00\x00\x02' + b'\x00\x00\x00\x02'
data = b'\x00\x00\x00\x01' + b'\x00\x00\x00\x02'
data += b'\x00\x00\x00\x03' + b'\x00\x00\x00\x04'
return header + dimensionsizes + data
_somebytes = _get_somebytes()
def test_list_to_idx():
data = idx.list_to_idx(_somelist, 'i')
print(data, _somebytes)
assert data == _somebytes
def test_idx_to_list():
lst = idx.idx_to_list(_somebytes)
assert lst == _somelist
|
# pylint: disable=W0613
from django.dispatch import receiver
from django.db.models.signals import post_save
from voting import constants
from source.models import Source
from voting.models import VotingRound
from source import constants as source_constants
from contracts.models import Contract
@receiver(signal=post_save, sender=Source)
def create_voting_round(instance, created, **kwargs):
"""
Creates a voting round after new Source is created.
"""
if created:
voting_round = VotingRound(source=instance)
voting_round.save()
@receiver(signal=post_save, sender=VotingRound)
def process_voting_round(instance, created, **kwargs):
"""
Edits Source according to decision made in voting round.
If source already has valid contract then we can switch directly
to running state.
"""
if not created:
source = instance.source
if instance.state == constants.VOTE_APPROVE:
if source.contract_set.valid():
source.state = source_constants.STATE_RUNNING
source.save()
return
else:
contract = Contract(source=source)
contract.publisher = source.publisher
contract.save()
contract.sources.add(source)
source.state = constants.VOTE_TO_SOURCE[instance.state]
source.save()
| Fix process_voting_round to reflect contract model
| # pylint: disable=W0613
from django.dispatch import receiver
from django.db.models.signals import post_save
from voting import constants
from source.models import Source
from voting.models import VotingRound
from source import constants as source_constants
from contracts.models import Contract
@receiver(signal=post_save, sender=Source)
def create_voting_round(instance, created, **kwargs):
"""
Creates a voting round after new Source is created.
"""
if created:
voting_round = VotingRound(source=instance)
voting_round.save()
@receiver(signal=post_save, sender=VotingRound)
def process_voting_round(instance, created, **kwargs):
"""
Edits Source according to decision made in voting round.
If source already has valid contract then we can switch directly
to running state.
"""
if not created:
source = instance.source
if instance.state == constants.VOTE_APPROVE:
if source.contract_set.valid():
source.state = source_constants.STATE_RUNNING
source.save()
return
else:
contract = Contract()
contract.publisher = source.publisher
contract.save()
contract.sources.add(source)
source.state = constants.VOTE_TO_SOURCE[instance.state]
source.save()
|
from django.contrib import admin
from .models import Language, CultureCode
class LanguageAdmin(admin.ModelAdmin):
list_display = ('name_en', 'name_native', 'iso_639_1', 'iso_639_2T', 'iso_639_2B', 'iso_639_2T',
'iso_639_3', 'iso_639_6', 'notes')
list_display_links = ('name_en',)
class CultureCodeAdmin(admin.ModelAdmin):
list_display = ('code', 'language', 'country')
list_display_links = ('code',)
admin.site.register(Language, LanguageAdmin)
admin.site.register(CultureCode, CultureCodeAdmin)
| Define `search_fields` for Admin classes
This enables the search box on the admin change list page [1], and can
be used by other apps like django-autocomplete-light [2].
1: https://docs.djangoproject.com/en/1.7/ref/contrib/admin/#django.contrib.admin.ModelAdmin.search_fields
2: https://github.com/yourlabs/django-autocomplete-light/pull/361
| from django.contrib import admin
from .models import Language, CultureCode
class LanguageAdmin(admin.ModelAdmin):
list_display = ('name_en', 'name_native', 'iso_639_1', 'iso_639_2T', 'iso_639_2B', 'iso_639_2T',
'iso_639_3', 'iso_639_6', 'notes')
list_display_links = ('name_en',)
search_fields = ('name_en', 'name_native')
class CultureCodeAdmin(admin.ModelAdmin):
list_display = ('code', 'language', 'country')
list_display_links = ('code',)
search_fields = ('code', 'language', 'country')
admin.site.register(Language, LanguageAdmin)
admin.site.register(CultureCode, CultureCodeAdmin)
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
urlpatterns = patterns(
'',
url(r'^$', 'beethoven.views.index', name='index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^', include('allauth.urls')),
url(r'^', include('courses.urls', namespace='courses')),
url(r'^', include('quizzes.urls', namespace='quizzes')),
url(r'^api/', include('quizzes.api_urls', namespace='quizzes_api'))
)
| Fix static file 404 error
| from django.conf.urls import patterns, include, url
from django.contrib import admin
from beethoven import settings
urlpatterns = patterns(
'',
url(r'^$', 'beethoven.views.index', name='index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^', include('allauth.urls')),
url(r'^', include('courses.urls', namespace='courses')),
url(r'^', include('quizzes.urls', namespace='quizzes')),
url(r'^api/', include('quizzes.api_urls', namespace='quizzes_api'))
)
if not settings.PRODUCTION:
urlpatterns += patterns(
'',
(r'^static/(?P<path>.*)$',
'django.views.static.serve',
{'document_root': settings.STATIC_ROOT})
)
|
# -*- mode: python; indent-tabs-mode: nil -*-
class Connection(object):
"""Interface for receiver connections.
A receiver connection is something that can send messages (filter requests,
multilateration results) to a particular receiver. A single connection
may handle only a single receiver, or may multiplex multiple receivers.
This is a duck-typed interface, implementations are not required to inherit
this class as long as they provide methods with equivalent signatures.
"""
def request_traffic(self, receiver, icao_set):
"""Request that a receiver starts sending traffic for the given
set of aircraft.
receiver: the handle of the concerned receiver
icao_set: a set of ICAO addresses (as ints) to start sending
"""
pass
def suppress_traffic(self, receiver, icao_set):
"""Request that a receiver stops sending traffic for the given
set of aircraft.
receiver: the handle of the concerned receiver
icao_set: a set of ICAO addresses (as ints) to stop sending
"""
pass
def report_mlat_position(self, receiver,
icao, utc, ecef, ecef_cov, nstations):
"""Report a multilaterated position result.
receiver: the handle of the concerned receiver
icao: the ICAO address of the aircraft (as an int)
utc: the approximate validity time of the position
ecef: an (x,y,z) tuple giving the position in ECEF coordinates
ecef_cov: a 3x3 matrix giving the covariance matrix of ecef
nstations: the number of stations that contributed to the result
"""
pass
| Raise NotImplemented if methods aren't overridden
| # -*- mode: python; indent-tabs-mode: nil -*-
class Connection(object):
"""Interface for receiver connections.
A receiver connection is something that can send messages (filter requests,
multilateration results) to a particular receiver. A single connection
may handle only a single receiver, or may multiplex multiple receivers.
This is a duck-typed interface, implementations are not required to inherit
this class as long as they provide methods with equivalent signatures.
"""
def request_traffic(self, receiver, icao_set):
"""Request that a receiver starts sending traffic for the given
set of aircraft.
receiver: the handle of the concerned receiver
icao_set: a set of ICAO addresses (as ints) to start sending
"""
raise NotImplementedError
def suppress_traffic(self, receiver, icao_set):
"""Request that a receiver stops sending traffic for the given
set of aircraft.
receiver: the handle of the concerned receiver
icao_set: a set of ICAO addresses (as ints) to stop sending
"""
raise NotImplementedError
def report_mlat_position(self, receiver,
icao, utc, ecef, ecef_cov, nstations):
"""Report a multilaterated position result.
receiver: the handle of the concerned receiver
icao: the ICAO address of the aircraft (as an int)
utc: the approximate validity time of the position
ecef: an (x,y,z) tuple giving the position in ECEF coordinates
ecef_cov: a 3x3 matrix giving the covariance matrix of ecef
nstations: the number of stations that contributed to the result
"""
raise NotImplementedError
|
import matplotlib.pyplot as plt; plt.rcdefaults()
import numpy as np
def horizontal_bar(sorted_streaks, sort):
"""
Render a horizontal bar chart of streaks.
Values have already been sorted by sort.
"""
# Only extract those users & streaks for streaks that are non-zero:
users, streaks = zip(*[(user, streak.get(sort)) for user, streak
in sorted_streaks if streak.get(sort) > 0][::-1])
title = 'Top Contributors by {} Streak'.format(sort.title())
figure = plt.figure()
y_pos = np.arange(len(users)) # y-location of bars
print('y_pos', y_pos)
plt.barh(y_pos, streaks, facecolor='#ff9999', edgecolor='grey', align='center')
plt.yticks(y_pos, users)
plt.xlim([0, max(streaks) + 0.5]) # x-limits a bit wider at right
plt.subplots_adjust(left=0.2) # Wider left margin
plt.title(title)
for format in ('png', 'svg'):
figure.savefig('temp/top_{}.{}'.format(sort, format), format=format)
| Remove weird matplot lib defaults thing that did nothing
| import matplotlib.pyplot as plt
import numpy as np
def horizontal_bar(sorted_streaks, sort):
"""
Render a horizontal bar chart of streaks.
Values have already been sorted by sort.
"""
# Only extract those users & streaks for streaks that are non-zero:
users, streaks = zip(*[(user, streak.get(sort)) for user, streak
in sorted_streaks if streak.get(sort) > 0][::-1])
title = 'Top Contributors by {} Streak'.format(sort.title())
figure = plt.figure()
y_pos = np.arange(len(users)) # y-location of bars
print('y_pos', y_pos)
plt.barh(y_pos, streaks, facecolor='#ff9999', edgecolor='grey', align='center')
plt.yticks(y_pos, users)
plt.xlim([0, max(streaks) + 0.5]) # x-limits a bit wider at right
plt.subplots_adjust(left=0.2) # Wider left margin
plt.title(title)
for format in ('png', 'svg'):
figure.savefig('temp/top_{}.{}'.format(sort, format), format=format)
|
import os
import sys
if not (2, 6) <= sys.version_info < (3,):
sys.exit(u'Mopidy requires Python >= 2.6, < 3')
VERSION = (0, 4, 0)
def is_in_git_repo():
git_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '../.git'))
return os.path.exists(git_dir)
def get_git_version():
if not is_in_git_repo():
return None
git_version = os.popen('git describe').read().strip()
if git_version.startswith('v'):
git_version = git_version[1:]
return git_version
def get_plain_version():
return '.'.join(map(str, VERSION))
def get_version():
if is_in_git_repo():
return get_git_version()
else:
return get_plain_version()
class MopidyException(Exception):
def __init__(self, message, *args, **kwargs):
super(MopidyException, self).__init__(message, *args, **kwargs)
self._message = message
@property
def message(self):
"""Reimplement message field that was deprecated in Python 2.6"""
return self._message
@message.setter
def message(self, message):
self._message = message
class SettingsError(MopidyException):
pass
class OptionalDependencyError(MopidyException):
pass
from mopidy import settings as default_settings_module
from mopidy.utils.settings import SettingsProxy
settings = SettingsProxy(default_settings_module)
| Use subprocess instead of os.popen
| import sys
if not (2, 6) <= sys.version_info < (3,):
sys.exit(u'Mopidy requires Python >= 2.6, < 3')
from subprocess import PIPE, Popen
VERSION = (0, 4, 0)
def get_git_version():
process = Popen(['git', 'describe'], stdout=PIPE)
if process.wait() != 0:
raise Exception|('Execution of "git describe" failed')
version = process.stdout.read().strip()
if version.startswith('v'):
version = version[1:]
return version
def get_plain_version():
return '.'.join(map(str, VERSION))
def get_version():
try:
return get_git_version()
except Exception:
return get_plain_version()
class MopidyException(Exception):
def __init__(self, message, *args, **kwargs):
super(MopidyException, self).__init__(message, *args, **kwargs)
self._message = message
@property
def message(self):
"""Reimplement message field that was deprecated in Python 2.6"""
return self._message
@message.setter
def message(self, message):
self._message = message
class SettingsError(MopidyException):
pass
class OptionalDependencyError(MopidyException):
pass
from mopidy import settings as default_settings_module
from mopidy.utils.settings import SettingsProxy
settings = SettingsProxy(default_settings_module)
|
from flask import Flask, url_for, Response, json, request
class MonitorApp(object):
def __init__(self, monitor):
self.app = Flask(__name__)
self.app.monitor = monitor
self.setup()
def setup(self):
@self.app.route('/anomaly', methods = ['POST'])
def api_anomaly():
data = request.json
if request.headers['Content-Type'] == 'application/json':
success = self.app.monitor.process_anomaly_data(data)
return handle_response(success)
else:
return Response("Unsupported media type\n" + data, status=415)
@self.app.route('/monitor', methods = ['POST'])
def api_monitor():
data = request.json
if request.headers['Content-Type'] == 'application/json':
success = self.app.monitor.process_monitor_flows(data)
return handle_response(success)
else:
return Response("Unsupported media type\n" + data, status=415)
def handle_response(self, success):
if success:
return Response("OK\n" + data, status=status)
else:
return Response("BAD REQUEST\n" + data, status=status)
| Fix status codes of handled responses
| from flask import Flask, url_for, Response, json, request
class MonitorApp(object):
def __init__(self, monitor):
self.app = Flask(__name__)
self.app.monitor = monitor
self.setup()
def setup(self):
@self.app.route('/anomaly', methods = ['POST'])
def api_anomaly():
data = request.json
if request.headers['Content-Type'] == 'application/json':
success = self.app.monitor.process_anomaly_data(data)
return self.handle_response(success, data)
else:
return Response("Unsupported media type\n" + data, status=415)
@self.app.route('/monitor', methods = ['POST'])
def api_monitor():
data = request.json
if request.headers['Content-Type'] == 'application/json':
success = self.app.monitor.process_monitor_flows(data)
return self.handle_response(success, data)
else:
return Response("Unsupported media type\n" + data, status=415)
def handle_response(self, success, data):
json_data = json.dumps(data)
if success:
return Response("OK\n" + json_data, status=200)
else:
return Response("BAD REQUEST\n" + json_data, status=400)
|
from django import forms
from django.forms import ModelForm
from django.forms.models import inlineformset_factory
from common.models import Repository, Collaboration, User
slug_errors = {
'invalid' : "Use only letters, numbers, underscores, and hyphens",
}
class NewRepositoryForm(forms.Form):
repo_name = forms.SlugField(max_length=100, error_messages=slug_errors)
class RepositoryForm(ModelForm):
class Meta:
model = Repository
fields = ['description', 'long_description', 'is_public']
class CollaborationForm(ModelForm):
user = forms.CharField()
class Meta:
model = Collaboration
exclude = ('repository', 'user')
def __init__(self, **kwargs):
super(CollaborationForm, self).__init__(**kwargs)
if 'instance' in kwargs:
print kwargs['instance']
self.fields['user'] = forms.CharField(initial=kwargs['instance'].user.username)
def save(self, **kwargs):
username = self.cleaned_data['user']
user = User.objects.get(username=username)
self.instance.user = user
return super(CollaborationForm, self).save(**kwargs)
CollaborationFormSet = inlineformset_factory(Repository, Repository.collaborators.through, form=CollaborationForm)
| Fix IntegrityError and DoesNotExist 500s
|
from django import forms
from django.forms import ModelForm
from django.forms.models import inlineformset_factory
from common.models import Repository, Collaboration, User
slug_errors = {
'invalid' : "Use only letters, numbers, underscores, and hyphens",
}
class NewRepositoryForm(forms.Form):
repo_name = forms.SlugField(max_length=100, error_messages=slug_errors)
class RepositoryForm(ModelForm):
class Meta:
model = Repository
fields = ['description', 'long_description', 'is_public']
class CollaborationForm(ModelForm):
user = forms.CharField()
class Meta:
model = Collaboration
exclude = ('repository', 'user')
def __init__(self, **kwargs):
super(CollaborationForm, self).__init__(**kwargs)
if 'instance' in kwargs:
self.fields['user'] = forms.CharField(initial=kwargs['instance'].user.username)
def clean(self):
cleaned_data = super(CollaborationForm, self).clean()
self.instance.full_clean()
return cleaned_data
def clean_user(self):
username = self.cleaned_data['user']
user = None
try:
user = User.objects.get(username=username)
self.instance.user = user
except User.DoesNotExist:
raise forms.ValidationError("User %(username_s does not exist",
params={'username':username})
CollaborationFormSet = inlineformset_factory(Repository, Repository.collaborators.through, form=CollaborationForm)
|
# Copyright 2013 Donald Stufft
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, division, print_function
from __future__ import unicode_literals
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__title__ = "twine"
__summary__ = "Collection of utilities for interacting with PyPI"
__uri__ = "https://github.com/pypa/twine"
__version__ = "1.8.1"
__author__ = "Donald Stufft and individual contributors"
__email__ = "[email protected]"
__license__ = "Apache License, Version 2.0"
__copyright__ = "Copyright 2013 Donald Stufft"
| Allow star imports from twine
Unicode literals on Python 2 prevent people from being able to use
from twine import *
Closes gh-209
(cherry picked from commit c2cd72d0f4ff4d380845333fbfaaf2c92d6a5674)
| # Copyright 2013 Donald Stufft
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, division, print_function
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__title__ = "twine"
__summary__ = "Collection of utilities for interacting with PyPI"
__uri__ = "https://github.com/pypa/twine"
__version__ = "1.8.1"
__author__ = "Donald Stufft and individual contributors"
__email__ = "[email protected]"
__license__ = "Apache License, Version 2.0"
__copyright__ = "Copyright 2013 Donald Stufft"
|
from django.conf.urls import url
from . import views
app_name = 'places'
urlpatterns = [
url(r'^$', views.IndexView.as_view(), name='index'),
url(r'^(?P<slug>[-\w]+)/$', views.PlaceView.as_view(), name='place'),
]
| Move places urlpatterns to Django 2.0 preferred method
| from django.urls import include, path
from . import views
app_name = 'places'
urlpatterns = [
path('', views.IndexView.as_view(), name='index'),
path('<slug:slug>/', views.PlaceView.as_view(), name='place'),
]
|
#! /usr/bin/env python
import os
from cyclus_tools import run_cyclus
from tests_list import sim_files
def main():
"""Creates reference databases. Assumes that cyclus is included into PATH.
"""
cwd = os.getcwd()
# Run cyclus
run_cyclus("cyclus", cwd, sim_files)
if __name__ == "__main__": main()
| Correct import statement after renaming test_lists.py to testcases
| #! /usr/bin/env python
import os
from cyclus_tools import run_cyclus
from testcases import sim_files
def main():
"""Creates reference databases. Assumes that cyclus is included into PATH.
"""
cwd = os.getcwd()
# Run cyclus
run_cyclus("cyclus", cwd, sim_files)
if __name__ == "__main__": main()
|
"""Python Library Boilerplate contains all the boilerplate you need to create a Python package."""
__author__ = 'Michael Joseph'
__email__ = '[email protected]'
__url__ = 'https://github.com/michaeljoseph/pymoji'
__version__ = '0.0.1'
def pymoji():
return 'Hello World!'
| Return the emoji and format it
| """Emits HTML from emoji"""
__author__ = 'Michael Joseph'
__email__ = '[email protected]'
__url__ = 'https://github.com/michaeljoseph/pymoji'
__version__ = '0.0.1'
from .emoji import emoji
def pymoji(text):
if text[0] <> text[:-1] and text[0] <> ':':
text = ':%s:' % text
return emoji(text)
|
from __future__ import absolute_import
from celery.signals import setup_logging
import orchestrator.logger
__version__ = '0.3.5'
__author__ = 'sukrit'
orchestrator.logger.init_logging()
setup_logging.connect(orchestrator.logger.init_celery_logging)
| Prepare for next dev version to incorporate encofing fixes in flask-hyperschema library
| from __future__ import absolute_import
from celery.signals import setup_logging
import orchestrator.logger
__version__ = '0.3.6'
__author__ = 'sukrit'
orchestrator.logger.init_logging()
setup_logging.connect(orchestrator.logger.init_celery_logging)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*- #
from __future__ import unicode_literals
# This file is only used if you use `make publish` or
# explicitly specify it as your config file.
import os
import sys
sys.path.append(os.curdir)
from pelicanconf import *
SITEURL = ''
RELATIVE_URLS = False
FEED_ALL_ATOM = 'feeds/all.atom.xml'
CATEGORY_FEED_ATOM = 'feeds/%s.atom.xml'
DELETE_OUTPUT_DIRECTORY = True
# Following items are often useful when publishing
#DISQUS_SITENAME = ""
#GOOGLE_ANALYTICS = ""
| Stop generate feeds when publishing
| #!/usr/bin/env python3
# This file is only used if you use `make publish` or
# explicitly specify it as your config file.
import os
import sys
sys.path.append(os.curdir)
from pelicanconf import *
SITEURL = ''
RELATIVE_URLS = False
DELETE_OUTPUT_DIRECTORY = True
# Following items are often useful when publishing
#DISQUS_SITENAME = ""
#GOOGLE_ANALYTICS = ""
|
from flask import Flask, abort, jsonify
from flask_caching import Cache
from flask_cors import CORS
import main
import slack
app = Flask(__name__)
cache = Cache(app, config={"CACHE_TYPE": "simple"})
cors = CORS(app, resources={r"/*": {"origins": "*"}})
app.register_blueprint(slack.blueprint, url_prefix="/api/slack")
@app.route("/api/")
@cache.cached(timeout=10800)
def list_entities():
return jsonify({"entities": ["restaurant"]})
@app.route("/api/restaurant/")
@cache.cached(timeout=10800)
def list_restaurants():
return jsonify({"restaurants": main.list_restaurants()})
@app.route("/api/restaurant/<name>/")
@cache.cached(timeout=10800)
def get_restaurant(name):
data = dict(main.get_restaurant(name))
if not data:
abort(status=404)
data["menu"] = [{"dish": entry} for entry in data["menu"]]
return jsonify({"restaurant": data})
| Remove trailing slashes, add origin url to responses
| import flask
import flask_caching
import flask_cors
import main
import slack
app = flask.Flask(__name__)
cache = flask_caching.Cache(app, config={"CACHE_TYPE": "simple"})
cors = flask_cors.CORS(app, resources={r"/*": {"origins": "*"}})
app.register_blueprint(slack.blueprint, url_prefix="/api/slack")
@app.route("/api")
@cache.cached(timeout=10800)
def list_entities():
return flask.jsonify({"entities": ["restaurant"],
"url": flask.url_for("list_entities", _external=True)})
@app.route("/api/restaurant")
@cache.cached(timeout=10800)
def list_restaurants():
return flask.jsonify({"restaurants": main.list_restaurants(),
"url": flask.url_for("list_restaurants", _external=True)})
@app.route("/api/restaurant/<name>")
@cache.cached(timeout=10800)
def get_restaurant(name):
data = dict(main.get_restaurant(name))
if not data:
abort(status=404)
data["menu"] = [{"dish": entry} for entry in data["menu"]]
return flask.jsonify({"restaurant": data,
"url": flask.url_for("get_restaurant", name=name, _external=True)})
|
"""
Usage:
./manage.py import_tfl_stops < data/tfl/bus-stops.csv
"""
import requests
from titlecase import titlecase
from django.core.exceptions import ObjectDoesNotExist, MultipleObjectsReturned
from busstops.management.import_from_csv import ImportFromCSVCommand
from busstops.models import StopPoint
class Command(ImportFromCSVCommand):
@staticmethod
def get_name(atco_code):
"""
Given a stop's ATCO code, returns the best-formatted version of its common name from the
TfL API
"""
data = requests.get('https://api.tfl.gov.uk/StopPoint/%s' % atco_code).json()
return data.get('commonName')
def handle_row(self, row):
if row['Naptan_Atco'] in (None, '', 'NONE'):
return None
try:
stop = StopPoint.objects.get(pk=row['Naptan_Atco'])
except ObjectDoesNotExist:
try:
stop = StopPoint.objects.get(pk__contains=row['Naptan_Atco'])
except (ObjectDoesNotExist, MultipleObjectsReturned) as e:
print e, row
return None
if row['Heading'] != '':
stop.heading = row['Heading']
stop.common_name = self.get_name(stop.atco_code)
stop.tfl = True
if stop.street.isupper():
stop.street = titlecase(stop.street)
if stop.landmark.isupper():
stop.landmark = titlecase(stop.landmark)
stop.save()
| Work around null TfL common names
| """
Usage:
./manage.py import_tfl_stops < data/tfl/bus-stops.csv
"""
import requests
from titlecase import titlecase
from django.core.exceptions import ObjectDoesNotExist, MultipleObjectsReturned
from busstops.management.import_from_csv import ImportFromCSVCommand
from busstops.models import StopPoint
class Command(ImportFromCSVCommand):
@staticmethod
def get_name(atco_code):
"""
Given a stop's ATCO code, returns the best-formatted version of its common name from the
TfL API
"""
data = requests.get('https://api.tfl.gov.uk/StopPoint/%s' % atco_code).json()
return data.get('commonName')
def handle_row(self, row):
if row['Naptan_Atco'] in (None, '', 'NONE'):
return None
try:
stop = StopPoint.objects.get(pk=row['Naptan_Atco'])
except ObjectDoesNotExist:
try:
stop = StopPoint.objects.get(pk__contains=row['Naptan_Atco'])
except (ObjectDoesNotExist, MultipleObjectsReturned) as e:
print e, row
return None
if row['Heading'] != '':
stop.heading = row['Heading']
stop.common_name = self.get_name(stop.atco_code) or stop.common_name
stop.tfl = True
if stop.street.isupper():
stop.street = titlecase(stop.street)
if stop.landmark.isupper():
stop.landmark = titlecase(stop.landmark)
stop.save()
|
#!/usr/bin/env python
#-------------------------------------------------------------------------
# Copyright (c) Microsoft. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#--------------------------------------------------------------------------
from distutils.core import setup
setup(name='ptvsd',
version='2.2.0rc2',
description='Python Tools for Visual Studio remote debugging server',
license='Apache License 2.0',
author='Microsoft Corporation',
author_email='[email protected]',
url='https://aka.ms/ptvs',
classifiers=[
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'License :: OSI Approved :: Apache Software License'],
packages=['ptvsd']
)
| Update ptvsd version number for 2.2 release.
| #!/usr/bin/env python
#-------------------------------------------------------------------------
# Copyright (c) Microsoft. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#--------------------------------------------------------------------------
from distutils.core import setup
setup(name='ptvsd',
version='2.2.0',
description='Python Tools for Visual Studio remote debugging server',
license='Apache License 2.0',
author='Microsoft Corporation',
author_email='[email protected]',
url='https://aka.ms/ptvs',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'License :: OSI Approved :: Apache Software License'],
packages=['ptvsd']
)
|
from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from formtools.wizard.views import SessionWizardView
from ansible.models import Github
def index(request):
return HttpResponse("200")
class PlaybookWizard(SessionWizardView):
def get_form_initial(self, step):
initial = {}
if step == '1':
prev_data = self.storage.get_step_data('0')
initial['name'] = prev_data['0-repository']
return self.initial_dict.get(step, initial)
return self.initial_dict.get(step, {})
def done(self, form_list, **kwargs):
form_data = {}
for form in form_list:
form.save()
return HttpResponseRedirect('/ansible')
| Save form data to DB on each step
| from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from formtools.wizard.views import SessionWizardView
from ansible.models import Github, Playbook
import sys
def index(request):
return HttpResponse("200")
class PlaybookWizard(SessionWizardView):
def get_form_initial(self, step):
initial = {}
if step == '1':
prev_data = self.storage.get_step_data('0')
initial['name'] = prev_data['0-repository']
return self.initial_dict.get(step, initial)
return self.initial_dict.get(step, {})
def get_form_step_data(self, form):
data = {}
if self.get_form_prefix() == '0':
github = Github()
github.repository = form.data.dict()['0-repository']
github.username = form.data.dict()['0-username']
github.save()
if self.get_form_prefix() == '1':
playbook = Playbook()
playbook.name = form.data.dict()['1-name']
playbook.inventory = form.data.dict()['1-inventory']
playbook.user = form.data.dict()['1-user']
playbook.save()
return form.data
def done(self, form_list, **kwargs):
return HttpResponseRedirect('/ansible')
|
#!/usr/bin/env python3
from src.TwircBot import TwircBot
import sys
try:
bot = TwircBot(sys.argv[1])
except IndexError:
bot = TwircBot()
bot.print_config()
bot.start()
| Add extremely basic template for command modules
| #!/usr/bin/env python3
from src.TwircBot import TwircBot
from src.CommandModule import CommandModule
import sys
try:
bot = TwircBot(sys.argv[1])
except IndexError:
bot = TwircBot()
module = CommandModule()
bot.print_config()
# bot.start()
|
# Based on fix_intern.py. Original copyright:
# Copyright 2006 Georg Brandl.
# Licensed to PSF under a Contributor Agreement.
"""Fixer for intern().
intern(s) -> sys.intern(s)"""
# Local imports
from lib2to3 import pytree
from lib2to3 import fixer_base
from lib2to3.fixer_util import Name, Attr, touch_import
class FixReload(fixer_base.BaseFix):
BM_compatible = True
order = "pre"
PATTERN = """
power< 'reload'
after=any*
>
"""
def transform(self, node, results):
touch_import('imp', u'reload', node)
return node
| Fix 2to3 fixers to work with Python 3.
| # Based on fix_intern.py. Original copyright:
# Copyright 2006 Georg Brandl.
# Licensed to PSF under a Contributor Agreement.
"""Fixer for intern().
intern(s) -> sys.intern(s)"""
# Local imports
from lib2to3 import pytree
from lib2to3 import fixer_base
from lib2to3.fixer_util import Name, Attr, touch_import
class FixReload(fixer_base.BaseFix):
BM_compatible = True
order = "pre"
PATTERN = """
power< 'reload'
after=any*
>
"""
def transform(self, node, results):
touch_import('imp', 'reload', node)
return node
|
# Copyright 2019 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import json
import sys
import argparse
parser = argparse.ArgumentParser(description='Fetches a field from a single API in the catalog')
parser.add_argument('file', help='File to load')
parser.add_argument('id', help='ID of API to fetch')
parser.add_argument('field', help='Field to find and output')
args = parser.parse_args()
filename = sys.argv[1]
file = open(filename, "r")
catalog = json.load(file)
query = [api.get(args.field) for api in catalog["apis"] if api["id"] == args.id]
if len(query) != 1:
raise Exception(f"API {args.id} not found (or has duplicate definitions)")
elif not query[0]:
raise Exception(f"API {args.id} has no field {args.field}")
else:
print(query[0])
| Allow a default value to be specified when fetching a field value
| # Copyright 2019 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import json
import sys
import argparse
parser = argparse.ArgumentParser(description='Fetches a field from a single API in the catalog')
parser.add_argument('file', help='File to load')
parser.add_argument('id', help='ID of API to fetch')
parser.add_argument('field', help='Field to find and output')
parser.add_argument('--default', help='Default value to output if field is not present')
args = parser.parse_args()
filename = sys.argv[1]
file = open(filename, "r")
catalog = json.load(file)
query = [api.get(args.field) for api in catalog["apis"] if api["id"] == args.id]
if len(query) != 1:
raise Exception(f"API {args.id} not found (or has duplicate definitions)")
elif not query[0] and args.default:
print(args.default)
elif not query[0]:
raise Exception(f"API {args.id} has no field {args.field}")
else:
print(query[0])
|
import pytest
from fmn.rules.cache import cache
@pytest.fixture(autouse=True, scope="session")
def configured_cache():
cache.configure()
| Fix intermittent failures of test_guard_http_exception
Signed-off-by: Ryan Lerch <[email protected]>
| import pytest
from fmn.rules.cache import cache
@pytest.fixture(autouse=True)
def configured_cache():
if not cache.region.is_configured:
cache.configure()
yield
cache.region.invalidate()
|
import functools
import os
from nose.tools import istest, assert_equal
import spur
def test(func):
@functools.wraps(func)
def run_test():
for shell in _create_shells():
yield func, shell
def _create_shells():
return [
spur.LocalShell(),
_create_ssh_shell()
]
def _create_ssh_shell():
return spur.SshShell(
hostname=os.environ.get("TEST_SSH_HOSTNAME", "127.0.0.1"),
username=os.environ["TEST_SSH_USERNAME"],
password=os.environ["TEST_SSH_PASSWORD"],
port=int(os.environ.get("TEST_SSH_PORT"))
)
return istest(run_test)
@test
def output_of_run_is_stored(shell):
result = shell.run(["echo", "hello"])
assert_equal("hello\n", result.output)
@test
def cwd_of_run_can_be_set(shell):
result = shell.run(["pwd"], cwd="/")
assert_equal("/\n", result.output)
@test
def environment_variables_can_be_added_for_run(shell):
result = shell.run(["sh", "-c", "echo $NAME"], update_env={"NAME": "Bob"})
assert_equal("Bob\n", result.output)
| Add test for output that doesn't end in a newline
| import functools
import os
from nose.tools import istest, assert_equal
import spur
def test(func):
@functools.wraps(func)
def run_test():
for shell in _create_shells():
yield func, shell
def _create_shells():
return [
spur.LocalShell(),
_create_ssh_shell()
]
def _create_ssh_shell():
return spur.SshShell(
hostname=os.environ.get("TEST_SSH_HOSTNAME", "127.0.0.1"),
username=os.environ["TEST_SSH_USERNAME"],
password=os.environ["TEST_SSH_PASSWORD"],
port=int(os.environ.get("TEST_SSH_PORT"))
)
return istest(run_test)
@test
def output_of_run_is_stored(shell):
result = shell.run(["echo", "hello"])
assert_equal("hello\n", result.output)
@test
def output_is_not_truncated_when_not_ending_in_a_newline(shell):
result = shell.run(["echo", "-n", "hello"])
assert_equal("hello", result.output)
@test
def cwd_of_run_can_be_set(shell):
result = shell.run(["pwd"], cwd="/")
assert_equal("/\n", result.output)
@test
def environment_variables_can_be_added_for_run(shell):
result = shell.run(["sh", "-c", "echo $NAME"], update_env={"NAME": "Bob"})
assert_equal("Bob\n", result.output)
|
from unittest import TestCase
from django.test.utils import setup_test_template_loader, override_settings
from django.template import Context
from django.template.loader import get_template
TEMPLATES = {
'basetag': '''{% load damn %}{% assets %}''',
'test2': '''
<!doctype html>{% load damn %}
<html>
<head>
{% assets %}
</head>
<body>
{% asset 'js/jquery.js' %}
</body>
</html>
''',
}
DAMN_PROCESSORS = {
'js': {
'class': 'damn.processors.ScriptProcessor',
},
}
class TagTests(TestCase):
def setUp(self):
setup_test_template_loader(TEMPLATES)
@override_settings(
DAMN_PROCESSORS=DAMN_PROCESSORS,
)
def test_simple(self):
t = get_template('basetag')
t.render()
@override_settings(
DAMN_PROCESSORS=DAMN_PROCESSORS,
)
def test_one(self):
t = get_template('test2')
o = t.render(Context())
self.assertContains(o, '<script src="/static/js/jquery.js"></script>')
| Use TestCase from Django
Set STATIC_URL
|
#from unittest import TestCase
from django.test import TestCase
from django.test.utils import setup_test_template_loader, override_settings
from django.template import Context
from django.template.loader import get_template
TEMPLATES = {
'basetag': '''{% load damn %}{% assets %}''',
'test2': '''
<!doctype html>{% load damn %}
<html>
<head>
{% assets %}
</head>
<body>
{% asset 'js/jquery.js' %}
</body>
</html>
''',
}
DAMN_PROCESSORS = {
'js': {
'processor': 'damn.processors.ScriptProcessor',
},
}
class TagTests(TestCase):
def setUp(self):
setup_test_template_loader(TEMPLATES)
@override_settings(
DAMN_PROCESSORS=DAMN_PROCESSORS,
STATIC_URL = '/',
)
def test_simple(self):
t = get_template('basetag')
t.render()
@override_settings(
DAMN_PROCESSORS=DAMN_PROCESSORS,
STATIC_URL = '/',
)
def test_one(self):
t = get_template('test2')
o = t.render(Context())
self.assertTrue('<script src="/static/js/jquery.js"></script>' in o)
|
#!/usr/bin/env python
from os import listdir
import re
#reads in old makefile from folder
#parses for compiler arguments
#creates cmake lists file with parsed arguments as parent-scope variables
def readAndMake(folder):
inStream = open(folder+"/Makefile", "r")
oldMake = inStream.readlines()
inStream.close()
makeVars = {'CFLAGS': '-fno-builtin -Wno-int-to-pointer-cast '}
for line in oldMake:
line = re.sub('[\r\n\t]', '', line)
if "=" in line:
var, val = line.split('=', 1)
var = var.strip()
val = val.strip()
if var == "CFLAGS":
makeVars[var] += val.replace('-Werror', '')
else:
makeVars[var] = val
newMake = ""
for var, val in makeVars.iteritems():
newMake += 'set( {} "{}" )\n'.format(var, val)
newMake += 'buildCB(${CFLAGS})'
outStream = open(folder + "/CMakeLists.txt", "w")
outStream.write(newMake)
outStream.close()
#write makeFiles for all folders in path
def doAll(path):
dirs = listdir(path)
for folder in dirs:
folder = path + "/" + folder
#print folder
if "00" in folder:
print folder
readAndMake(folder)
if __name__ == '__main__':
path = "../cqe-challenges"
doAll(path) ##path should be folder containing multiple challenge binaries and nothing else.
| Add -Wno-writable-strings to clean up output
| #!/usr/bin/env python
from os import listdir
import re
#reads in old makefile from folder
#parses for compiler arguments
#creates cmake lists file with parsed arguments as parent-scope variables
def readAndMake(folder):
inStream = open(folder+"/Makefile", "r")
oldMake = inStream.readlines()
inStream.close()
makeVars = {'CFLAGS': '-fno-builtin -Wno-int-to-pointer-cast -Wno-writable-strings '}
for line in oldMake:
line = re.sub('[\r\n\t]', '', line)
if "=" in line:
var, val = line.split('=', 1)
var = var.strip()
val = val.strip()
if var == "CFLAGS":
makeVars[var] += val.replace('-Werror', '')
else:
makeVars[var] = val
newMake = ""
for var, val in makeVars.iteritems():
newMake += 'set( {} "{}" )\n'.format(var, val)
newMake += 'buildCB(${CFLAGS})'
outStream = open(folder + "/CMakeLists.txt", "w")
outStream.write(newMake)
outStream.close()
#write makeFiles for all folders in path
def doAll(path):
dirs = listdir(path)
for folder in dirs:
folder = path + "/" + folder
#print folder
if "00" in folder:
print folder
readAndMake(folder)
if __name__ == '__main__':
path = "../cqe-challenges"
doAll(path) ##path should be folder containing multiple challenge binaries and nothing else.
|
import unittest
import os
os.environ['OGN_CONFIG_MODULE'] = 'config/test.py'
from ogn_python import db # noqa: E402
class TestBaseDB(unittest.TestCase):
@classmethod
def setUpClass(cls):
db.session.execute('CREATE EXTENSION IF NOT EXISTS postgis;')
db.session.commit()
db.create_all()
def setUp(self):
pass
def tearDown(self):
db.session.execute("""
DELETE FROM aircraft_beacons;
DELETE FROM receiver_beacons;
DELETE FROM takeoff_landings;
DELETE FROM logbook;
DELETE FROM receiver_coverages;
DELETE FROM device_stats;
DELETE FROM receiver_stats;
DELETE FROM receivers;
DELETE FROM devices;
""")
if __name__ == '__main__':
unittest.main()
| Drop db before each test
| import unittest
import os
os.environ['OGN_CONFIG_MODULE'] = 'config/test.py'
from ogn_python import db # noqa: E402
class TestBaseDB(unittest.TestCase):
@classmethod
def setUpClass(cls):
db.drop_all()
db.session.execute('CREATE EXTENSION IF NOT EXISTS postgis;')
db.session.commit()
db.create_all()
def setUp(self):
pass
def tearDown(self):
db.session.execute("""
DELETE FROM aircraft_beacons;
DELETE FROM receiver_beacons;
DELETE FROM takeoff_landings;
DELETE FROM logbook;
DELETE FROM receiver_coverages;
DELETE FROM device_stats;
DELETE FROM receiver_stats;
DELETE FROM receivers;
DELETE FROM devices;
""")
if __name__ == '__main__':
unittest.main()
|
#!/usr/bin/env python3
"""
Unit tests at Windows environments required to invoke from py module,
because of multiprocessing:
https://py.rtfd.io/en/latest/faq.html?highlight=cmdline#issues-with-py-test-multiprocess-and-setuptools
"""
import sys
import py
if __name__ == "__main__":
sys.exit(py.test.cmdline.main())
| Add PYTEST_MD_REPORT_COLOR environment variable setting
| #!/usr/bin/env python3
"""
Unit tests at Windows environments required to invoke from py module,
because of multiprocessing:
https://py.rtfd.io/en/latest/faq.html?highlight=cmdline#issues-with-py-test-multiprocess-and-setuptools
"""
import os
import sys
import py
if __name__ == "__main__":
os.environ["PYTEST_MD_REPORT_COLOR"] = "text"
sys.exit(py.test.cmdline.main())
|
from django.conf.urls import patterns, url
from django.views.generic import TemplateView
from studentvoice import views
urlpatterns = patterns('',
url(r'^$', views.home, name='home'),
url(r'^create/$', views.create, name='create'),
url(r'^search/', views.search, name='search'),
url(r'^(?P<voice_id>\d+)/$', views.show, name='show'),
url(r'^vote/$', views.vote, name='vote'),
url(r'^(?P<voice_id>\d+)/report/$', views.report, name='report'),
url(r'^(?P<voice_id>\d+)/create_comment/$', views.create_comment, name='create_comment'),
url(r'^delete/(?P<voice_id>\d+)/$', views.delete, name='delete'),
url(r'^(?P<voice_id>\d+)/edit/$', views.edit, name='edit'),
url(r'^(?P<voice_id>\d+)/respond/$', views.respond, name='respond'),
url(r'^(?P<voice_id>\d+)/respond/edit/$', views.edit_response, name='edit_response'),
)
| Add the about page to url.py
| from django.conf.urls import patterns, url
from django.views.generic import TemplateView
from studentvoice import views
urlpatterns = patterns('',
url(r'^$', views.home, name='home'),
url(r'^about/$', views.about, name='about'),
url(r'^create/$', views.create, name='create'),
url(r'^search/', views.search, name='search'),
url(r'^(?P<voice_id>\d+)/$', views.show, name='show'),
url(r'^vote/$', views.vote, name='vote'),
url(r'^(?P<voice_id>\d+)/report/$', views.report, name='report'),
url(r'^(?P<voice_id>\d+)/create_comment/$', views.create_comment, name='create_comment'),
url(r'^delete/(?P<voice_id>\d+)/$', views.delete, name='delete'),
url(r'^(?P<voice_id>\d+)/edit/$', views.edit, name='edit'),
url(r'^(?P<voice_id>\d+)/respond/$', views.respond, name='respond'),
url(r'^(?P<voice_id>\d+)/respond/edit/$', views.edit_response, name='edit_response'),
)
|
class PluginLoader:
def load(self, directory):
ret = []
for plugin in directory.children:
plugin_file = plugin.find(name=plugin.name, type='py')
if not plugin_file:
continue
plugin_class = plugin_file.classes[plugin.name]
self._check_attributes(plugin_class)
plugin_instance = plugin_class()
ret.append(plugin_instance)
return ret
def _check_attributes(self, klass):
self._check_aliases(klass)
self._check_matches(klass)
self._check_priority(klass)
def _check_aliases(self, klass):
self._check_attribute(klass, 'aliases', '')
def _check_matches(self, klass):
def matches(self, expression):
if isinstance(self.aliases, str):
return expression == self.aliases
return expression in self.aliases
self._check_attribute(klass, 'matches', matches)
def _check_priority(self, klass):
self._check_attribute(klass, 'priority', 'normal')
if klass.priority not in ('low', 'normal', 'high'):
klass.priority = 'normal'
def _check_attribute(self, klass, attribute, value):
if not hasattr(klass, attribute):
setattr(klass, attribute, value)
| Make plugin loader more robust
| class PluginLoader:
def load(self, directory):
ret = []
for plugin in directory.children:
plugin_file = plugin.find(name=plugin.name, type='py')
if not plugin_file:
continue
plugin_class = plugin_file.classes.get(plugin.name)
if not plugin_class:
print 'Plugin file is missing proper class!', plugin.name, plugin_file.classes
continue
self._check_attributes(plugin_class)
plugin_instance = plugin_class()
ret.append(plugin_instance)
return ret
def _check_attributes(self, klass):
self._check_aliases(klass)
self._check_matches(klass)
self._check_priority(klass)
def _check_aliases(self, klass):
self._check_attribute(klass, 'aliases', '')
def _check_matches(self, klass):
def matches(self, expression):
if isinstance(self.aliases, str):
return expression == self.aliases
return expression in self.aliases
self._check_attribute(klass, 'matches', matches)
def _check_priority(self, klass):
self._check_attribute(klass, 'priority', 'normal')
if klass.priority not in ('low', 'normal', 'high'):
klass.priority = 'normal'
def _check_attribute(self, klass, attribute, value):
if not hasattr(klass, attribute):
setattr(klass, attribute, value)
|
# -*- coding: utf-8 -*-
# Copyright 2017 GIG Technology NV
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# @@license_version:1.3@@
NAMESPACE = u'tff_backend'
KEY_ALGORITHM = u'ed25519'
KEY_NAME = u'threefold'
THREEFOLD_APP_ID = u'em-be-threefold-token'
FULL_CURRENCY_NAMES = {
'USD': 'dollar',
'EUR': 'euro',
'YEN': 'yen',
'UAE': 'dirham',
'GBP': 'pound',
}
CURRENCY_RATES = {
'USD': 5.0,
'EUR': 4.2,
'YEN': 543.6,
'UAE': 18.6,
'GBP': 3.83
}
| Add BTC to possible currencies
| # -*- coding: utf-8 -*-
# Copyright 2017 GIG Technology NV
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# @@license_version:1.3@@
NAMESPACE = u'tff_backend'
KEY_ALGORITHM = u'ed25519'
KEY_NAME = u'threefold'
THREEFOLD_APP_ID = u'em-be-threefold-token'
FULL_CURRENCY_NAMES = {
'USD': 'dollar',
'EUR': 'euro',
'YEN': 'yen',
'UAE': 'dirham',
'GBP': 'pound',
'BTC': 'bitcoin',
}
CURRENCY_RATES = {
'USD': 5.0,
'EUR': 4.2,
'YEN': 543.6,
'UAE': 18.6,
'GBP': 3.83,
'BTC': .0011,
}
|
from math import gcd
def solve(number: int=20) -> str:
if number <= 0:
raise TypeError
lcd = 1
for i in range(1, number + 1):
lcd = (lcd * i) // gcd(lcd, i)
return str(lcd)
| Use ValueError for wrong input in 5
| from math import gcd
def solve(number: int=20) -> str:
if number <= 0:
raise ValueError
lcd = 1
for i in range(1, number + 1):
lcd = (lcd * i) // gcd(lcd, i)
return str(lcd)
|
# Function to return a list of pvs from a given file
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
| Add utility function to write pvs to file
| # Function to return a list of pvs from a given file
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
def get_pv_names(mode):
''' Given a certain ring mode as a string, return all available pvs '''
ap.machines.load(mode)
result = set()
elements = ap.getElements('*')
for element in elements:
pvs = element.pv()
if(len(pvs) > 0):
pv_name = pvs[0].split(':')[0]
result.add(pv_name)
return result
def get_pvs_from_file(filepath):
''' Return a list of pvs from a given file '''
with open(filepath) as f:
contents = f.read().splitlines()
return contents
def write_pvs_to_file(filename, data):
''' Write given pvs to file '''
f = open(filename, 'w')
for element in data:
f.write(element, '\n')
f.close()
|
import os
from django.core.files.base import ContentFile
from imagekit.lib import Image, StringIO
from .models import Photo
import pickle
def get_image_file():
"""
See also:
http://en.wikipedia.org/wiki/Lenna
http://sipi.usc.edu/database/database.php?volume=misc&image=12
"""
path = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'assets', 'lenna-800x600-white-border.jpg')
tmp = StringIO()
tmp.write(open(path, 'r+b').read())
tmp.seek(0)
return tmp
def create_image():
return Image.open(get_image_file())
def create_instance(model_class, image_name):
instance = model_class()
img = get_image_file()
file = ContentFile(img.read())
instance.original_image = file
instance.original_image.save(image_name, file)
instance.save()
img.close()
return instance
def create_photo(name):
return create_instance(Photo, name)
def pickleback(obj):
pickled = StringIO()
pickle.dump(obj, pickled)
pickled.seek(0)
return pickle.load(pickled)
| Add util for generating named image file
| import os
from django.core.files.base import ContentFile
from imagekit.lib import Image, StringIO
from tempfile import NamedTemporaryFile
from .models import Photo
import pickle
def _get_image_file(file_factory):
"""
See also:
http://en.wikipedia.org/wiki/Lenna
http://sipi.usc.edu/database/database.php?volume=misc&image=12
"""
path = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'assets', 'lenna-800x600-white-border.jpg')
tmp = file_factory()
tmp.write(open(path, 'r+b').read())
tmp.seek(0)
return tmp
def get_image_file():
return _get_image_file(StringIO)
def get_named_image_file():
return _get_image_file(NamedTemporaryFile)
def create_image():
return Image.open(get_image_file())
def create_instance(model_class, image_name):
instance = model_class()
img = get_image_file()
file = ContentFile(img.read())
instance.original_image = file
instance.original_image.save(image_name, file)
instance.save()
img.close()
return instance
def create_photo(name):
return create_instance(Photo, name)
def pickleback(obj):
pickled = StringIO()
pickle.dump(obj, pickled)
pickled.seek(0)
return pickle.load(pickled)
|
from matasano.util.converters import base64_to_bytes
from Crypto.Cipher import AES
import base64
if __name__ == "__main__":
chal_file = open("matasano/data/c7.txt", 'r');
key = "YELLOW SUBMARINE"
# Instantiate the cipher
cipher = AES.new(key, AES.MODE_ECB)
# Covert from base64 to bytes and encode
ct = base64_to_bytes(chal_file.read()).encode('latin-1')
# Perform the decryption
pt = cipher.decrypt(ct)
print(pt.decode())
| Switch to using base64 builtin decoder for simplicity.
| from matasano.util.converters import base64_to_bytes
from Crypto.Cipher import AES
import base64
if __name__ == "__main__":
chal_file = open("matasano/data/c7.txt", 'r');
key = "YELLOW SUBMARINE"
# Instantiate the cipher
cipher = AES.new(key, AES.MODE_ECB)
# Covert from base64 to bytes and encode
ct = base64.b64decode(chal_file.read())
# Perform the decryption
pt = cipher.decrypt(ct)
print(pt.decode())
|
# EXTRACT NODE INFO PY
import argparse, os, pickle, sys
from Node import Node
from utils import abort
parser = argparse.ArgumentParser(description='Parse all log files')
parser.add_argument('directory',
help='The experiment directory (EXPID)')
args = parser.parse_args()
node_pkl = args.directory + "/node-info.pkl"
try:
with open(node_pkl, 'rb') as fp:
data = pickle.load(fp)
except IOError as e:
abort(e, os.EX_IOERR, "Could not read: " + node_pkl)
# print(data)
for item in data.values():
print(item.str_table())
# print(len(data))
| Replace abort() with fail() again
|
# EXTRACT NODE INFO PY
import argparse, os, pickle, sys
from Node import Node
from utils import fail
parser = argparse.ArgumentParser(description='Parse all log files')
parser.add_argument('directory',
help='The experiment directory (EXPID)')
args = parser.parse_args()
node_pkl = args.directory + "/node-info.pkl"
try:
with open(node_pkl, 'rb') as fp:
data = pickle.load(fp)
except IOError as e:
fail(e, os.EX_IOERR, "Could not read: " + node_pkl)
# print(data)
for item in data.values():
print(item.str_table())
# print(len(data))
|
import pytest
from radar.validation.reset_password import ResetPasswordValidation
from radar.validation.core import ValidationError
from radar.tests.validation.helpers import validation_runner
def test_valid():
obj = valid({
'token': '12345',
'username': 'hello',
'password': 'password',
})
assert obj['token'] == '12345'
assert obj['username'] == 'hello'
assert obj['password'] == 'password'
def test_token_missing():
invalid({
'username': 'hello',
'password': 'password',
})
def test_username_missing():
invalid({
'token': '12345',
'password': 'password',
})
def test_password_missing():
invalid({
'token': '12345',
'username': 'hello',
})
def invalid(obj, **kwargs):
with pytest.raises(ValidationError) as e:
valid(obj, **kwargs)
return e
def valid(obj, **kwargs):
return validation_runner(dict, ResetPasswordValidation, obj, **kwargs)
| Use stronger password in reset password test
| import pytest
from radar.validation.reset_password import ResetPasswordValidation
from radar.validation.core import ValidationError
from radar.tests.validation.helpers import validation_runner
def test_valid():
obj = valid({
'token': '12345',
'username': 'hello',
'password': '2irPtfNUURf8G',
})
assert obj['token'] == '12345'
assert obj['username'] == 'hello'
assert obj['password'] == '2irPtfNUURf8G'
def test_token_missing():
invalid({
'username': 'hello',
'password': 'password',
})
def test_username_missing():
invalid({
'token': '12345',
'password': 'password',
})
def test_password_missing():
invalid({
'token': '12345',
'username': 'hello',
})
def test_weak_password():
invalid({
'token': '12345',
'username': 'hello',
'password': 'password',
})
def invalid(obj, **kwargs):
with pytest.raises(ValidationError) as e:
valid(obj, **kwargs)
return e
def valid(obj, **kwargs):
return validation_runner(dict, ResetPasswordValidation, obj, **kwargs)
|
if 'loaded' in locals():
import imp
imp.reload(blendergltf)
from .blendergltf import *
else:
loaded = True
from .blendergltf import * | Add experimental support to run module as Blender addon
| bl_info = {
"name": "glTF format",
"author": "Daniel Stokes",
"version": (0, 1, 0),
"blender": (2, 76, 0),
"location": "File > Import-Export",
"description": "Export glTF",
"warning": "",
"wiki_url": ""
"",
"support": 'TESTING',
"category": "Import-Export"}
# Treat as module
if '.' in __name__:
if 'loaded' in locals():
import imp
imp.reload(blendergltf)
from .blendergltf import *
else:
loaded = True
from .blendergltf import *
# Treat as addon
else:
if "bpy" in locals():
import importlib
importlib.reload(blendergltf)
import json
import bpy
from bpy.props import (
StringProperty,
)
from bpy_extras.io_utils import (
ExportHelper,
)
from . import blendergltf
class ExportGLTF(bpy.types.Operator, ExportHelper):
"""Save a Khronos glTF File"""
bl_idname = "export_scene.gltf"
bl_label = 'Export glTF'
filename_ext = ".gltf"
filter_glob = StringProperty(
default="*.gltf",
options={'HIDDEN'},
)
check_extension = True
def execute(self, context):
scene = {
'camera': bpy.data.cameras,
'lamps': bpy.data.lamps,
'images': bpy.data.images,
'materials': bpy.data.materials,
'meshes': bpy.data.meshes,
'objects': bpy.data.objects,
'scenes': bpy.data.scenes,
'textures': bpy.data.textures,
}
gltf = blendergltf.export_gltf(scene)
with open(self.filepath, 'w') as fout:
json.dump(gltf, fout, indent=4)
return {'FINISHED'}
def menu_func_export(self, context):
self.layout.operator(ExportGLTF.bl_idname, text="glTF (.gltf)")
def register():
bpy.utils.register_module(__name__)
bpy.types.INFO_MT_file_export.append(menu_func_export)
def unregister():
bpy.utils.unregister_module(__name__)
bpy.types.INFO_MT_file_export.remove(menu_func_export) |
#!/usr/bin/env python
#
#
# Generate linker script to only expose symbols of the public API
#
import sys
import re
if __name__ == '__main__':
funcs = list()
last_line = ''
for line in sys.stdin:
m = re.match(r'^(\S+.*\s+\**)?(rd_kafka_\S+)\s*\(', line)
if m:
sym = m.group(2)
m2 = re.match(r'(RD_UNUSED|__attribute__\(\(unused\)\))', line)
if not m2:
funcs.append(sym)
last_line = ''
else:
last_line = line
print('# Automatically generated by lds-gen.py - DO NOT EDIT')
print('{\n global:')
if len(funcs) == 0:
print(' *;')
else:
for f in sorted(funcs):
print(' %s;' % f)
print('};')
| Stop exporting internal symbols from the shared libraries.
| #!/usr/bin/env python
#
#
# Generate linker script to only expose symbols of the public API
#
import sys
import re
if __name__ == '__main__':
funcs = list()
last_line = ''
for line in sys.stdin:
m = re.match(r'^(\S+.*\s+\**)?(rd_kafka_\S+)\s*\(', line)
if m:
sym = m.group(2)
m2 = re.match(r'(RD_UNUSED|__attribute__\(\(unused\)\))', line)
if not m2:
funcs.append(sym)
last_line = ''
else:
last_line = line
print('# Automatically generated by lds-gen.py - DO NOT EDIT')
print('{\n global:')
if len(funcs) == 0:
print(' *;')
else:
for f in sorted(funcs):
print(' %s;' % f)
print('local:\n *;')
print('};')
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
class Migration(migrations.Migration):
dependencies = [
('companyprofile', '0001_squashed_0003_company_image'),
]
operations = [
migrations.CreateModel(
name='CareerOpportunity',
fields=[
('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),
('title', models.CharField(max_length=100, verbose_name='tittel')),
('ingress', models.CharField(max_length=250, verbose_name='ingress')),
('description', models.TextField(verbose_name='beskrivelse')),
('start', models.DateTimeField(verbose_name='aktiv fra')),
('end', models.DateTimeField(verbose_name='aktiv til')),
('featured', models.BooleanField(default=False, verbose_name='fremhevet')),
('company', models.ForeignKey(related_name='company', to='companyprofile.Company')),
],
options={
'verbose_name': 'karrieremulighet',
'verbose_name_plural': 'karrieremuligheter',
'permissions': (('view_careeropportunity', 'View CareerOpportunity'),),
},
bases=(models.Model,),
),
]
| Revert "Change careeropportunity migration dep"
This reverts commit 60fdfab7e3b557e46276c225ff159f5773930525.
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
class Migration(migrations.Migration):
dependencies = [
('companyprofile', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CareerOpportunity',
fields=[
('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),
('title', models.CharField(max_length=100, verbose_name='tittel')),
('ingress', models.CharField(max_length=250, verbose_name='ingress')),
('description', models.TextField(verbose_name='beskrivelse')),
('start', models.DateTimeField(verbose_name='aktiv fra')),
('end', models.DateTimeField(verbose_name='aktiv til')),
('featured', models.BooleanField(default=False, verbose_name='fremhevet')),
('company', models.ForeignKey(related_name='company', to='companyprofile.Company')),
],
options={
'verbose_name': 'karrieremulighet',
'verbose_name_plural': 'karrieremuligheter',
'permissions': (('view_careeropportunity', 'View CareerOpportunity'),),
},
bases=(models.Model,),
),
]
|
from fluent_pages.pagetypes.flatpage.models import FlatPage
from fluent_pages.pagetypes.fluentpage.models import FluentPage
from haystack import indexes
class FluentPageIndex(indexes.SearchIndex, indexes.Indexable):
"""
Search index for a fluent page.
"""
text = indexes.CharField(document=True, use_template=True)
author = indexes.CharField(model_attr='author')
publication_date = indexes.DateTimeField(model_attr='publication_date', null=True)
@staticmethod
def get_model():
"""
Get the model for the search index.
"""
return FluentPage
def index_queryset(self, using=None):
"""
Queryset appropriate for this object to allow search for.
"""
return self.get_model().objects.published()
class FlatPageIndex(FluentPageIndex):
"""
Search index for a flat page.
As everything except the model is the same as for a FluentPageIndex
we shall subclass it and overwrite the one part we need.
"""
@staticmethod
def get_model():
"""
Get the model for the search index.
"""
return FlatPage
| Add setting to turn of search indexes.
| from fluent_pages.pagetypes.flatpage.models import FlatPage
from fluent_pages.pagetypes.fluentpage.models import FluentPage
from haystack import indexes
from django.conf import settings
# Optional search indexes which can be used with the default FluentPage and FlatPage models.
if getattr(settings, 'ICEKIT_USE_SEARCH_INDEXES', True):
class FluentPageIndex(indexes.SearchIndex, indexes.Indexable):
"""
Search index for a fluent page.
"""
text = indexes.CharField(document=True, use_template=True)
author = indexes.CharField(model_attr='author')
publication_date = indexes.DateTimeField(model_attr='publication_date', null=True)
@staticmethod
def get_model():
"""
Get the model for the search index.
"""
return FluentPage
def index_queryset(self, using=None):
"""
Queryset appropriate for this object to allow search for.
"""
return self.get_model().objects.published()
class FlatPageIndex(FluentPageIndex):
"""
Search index for a flat page.
As everything except the model is the same as for a FluentPageIndex
we shall subclass it and overwrite the one part we need.
"""
@staticmethod
def get_model():
"""
Get the model for the search index.
"""
return FlatPage
|
# Examplary development configuration for the "CozyLAN" demo site
DEBUG = True
SECRET_KEY = b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00'
SESSION_COOKIE_SECURE = False
SQLALCHEMY_DATABASE_URI = 'postgresql+psycopg2://byceps:[email protected]/byceps'
REDIS_URL = 'redis://127.0.0.1:6379/0'
APP_MODE = 'site'
SITE_ID = 'cozylan'
MAIL_DEBUG = True
MAIL_DEFAULT_SENDER = 'CozyLAN <[email protected]>'
MAIL_SUPPRESS_SEND = False
MAIL_TRANSPORT = 'logging'
DEBUG_TOOLBAR_ENABLED = True
STYLE_GUIDE_ENABLED = True
| Remove default email sender from CozyLAN config
Brand-specific emails should always use the sender from the
brand-specific email configuration.
| # Examplary development configuration for the "CozyLAN" demo site
DEBUG = True
SECRET_KEY = b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00'
SESSION_COOKIE_SECURE = False
SQLALCHEMY_DATABASE_URI = 'postgresql+psycopg2://byceps:[email protected]/byceps'
REDIS_URL = 'redis://127.0.0.1:6379/0'
APP_MODE = 'site'
SITE_ID = 'cozylan'
MAIL_DEBUG = True
MAIL_SUPPRESS_SEND = False
MAIL_TRANSPORT = 'logging'
DEBUG_TOOLBAR_ENABLED = True
STYLE_GUIDE_ENABLED = True
|
from pkg_resources import resource_string
version = resource_string(__name__, 'version.txt').strip()
__version__ = version
| Add decode to version read from pkg_resources.
| from pkg_resources import resource_string
version = resource_string(__name__, 'version.txt').strip()
__version__ = version.decode('utf-8')
|
# Copyright (c) 2020, CNRS
# Authors: Guilhem Saurel <[email protected]>
import os
import subprocess
import time
try:
from subprocess import DEVNULL, run
except ImportError: # Python2 fallback
DEVNULL = os.open(os.devnull, os.O_RDWR)
def run(*args):
subprocess.Popen(*args).wait()
class ServerManager:
"""A context to ensure a server is running."""
def __init__(self, server="hppcorbaserver"):
self.server = server
run(["killall", self.server])
def __enter__(self):
"""Run the server in background
stdout and stderr outputs of the child process are redirected to devnull.
preexec_fn is used to ignore ctrl-c signal send to the main script
(otherwise they are forwarded to the child process)
"""
self.process = subprocess.Popen(
self.server, stdout=DEVNULL, stderr=DEVNULL, preexec_fn=os.setpgrp
)
# give it some time to start
time.sleep(3)
def __exit__(self, exc_type, exc_value, exc_traceback):
self.process.kill()
| Fix how hppcorbaserver is killed in ServerManager
| # Copyright (c) 2020, CNRS
# Authors: Guilhem Saurel <[email protected]>
import os
import subprocess
import time
import hpp.corbaserver
try:
from subprocess import DEVNULL, run
except ImportError: # Python2 fallback
DEVNULL = os.open(os.devnull, os.O_RDWR)
def run(*args):
subprocess.Popen(*args).wait()
class ServerManager:
"""A context to ensure a server is running."""
def __init__(self, server="hppcorbaserver"):
self.server = server
run(["killall", self.server])
def __enter__(self):
"""Run the server in background
stdout and stderr outputs of the child process are redirected to devnull.
preexec_fn is used to ignore ctrl-c signal send to the main script
(otherwise they are forwarded to the child process)
"""
self.process = subprocess.Popen(
self.server, stdout=DEVNULL, stderr=DEVNULL, preexec_fn=os.setpgrp
)
# give it some time to start
time.sleep(3)
def __exit__(self, exc_type, exc_value, exc_traceback):
tool = hpp.corbaserver.tools.Tools()
tool.shutdown()
self.process.communicate()
|
import django.dispatch
cache_read = django.dispatch.Signal(providing_args=["func", "hit"])
cache_invalidated = django.dispatch.Signal(providing_args=["obj_dict"])
| Stop using Signal(providing_args) deprected in Django 4.0
Closes #393
| import django.dispatch
cache_read = django.dispatch.Signal() # args: func, hit
cache_invalidated = django.dispatch.Signal() # args: obj_dict
|
from jinja2 import Environment, FileSystemLoader
from stayput import Templater
class Jinja2Templater(Templater):
def __init__(self, site, *args, **kwargs):
self.site = site
self.env = Environment(loader=FileSystemLoader(site.templates_path))
def template(self, item):
return self.env.from_string(item.contents).render(site=self.site, item=item)
| Update for stayput master and ensure forward compatibility
| from jinja2 import Environment, FileSystemLoader
from stayput import Templater
class Jinja2Templater(Templater):
def __init__(self, site, *args, **kwargs):
self.site = site
self.env = Environment(loader=FileSystemLoader(site.templates_path))
def template(self, item, site, *args, **kwargs):
return self.env.from_string(item.contents).render(site=self.site, item=item)
|
from . import changeling
| Add docstring to linters package
| """
Linters for verifying the correctness of certain character types
The `commands.lint` function can lint all basic files, but special character
types sometimes need extra checks. The linters in this package encapsulate that
logic.
All linter packages have a single main entry point `lint` which accepts a
character information dict. Various keyword arguments are used for options and
supporting data.
"""
from . import changeling
|
import json
from exporters.writers.base_writer import BaseWriter, ItemsLimitReached
class ConsoleWriter(BaseWriter):
"""
It is just a writer with testing purposes. It prints every item in console.
"""
def __init__(self, options):
super(ConsoleWriter, self).__init__(options)
self.logger.info('ConsoleWriter has been initiated')
self.pretty_print = self.options.get('pretty_print', False)
def write_batch(self, batch):
for item in batch:
formatted_item = item.formatted
if self.pretty_print:
formatted_item = self._format(formatted_item)
print formatted_item
self._increment_written_items()
if self.items_limit and self.items_limit == self.stats['items_count']:
raise ItemsLimitReached('Finishing job after items_limit reached: {} items written.'.format(self.stats['items_count']))
self.logger.debug('Wrote items')
def _format(self, item):
try:
return json.dumps(json.loads(item), indent=2)
except:
return item
| Use sort_keys=True for the ConsoleWritter pretty printing
| import json
from exporters.writers.base_writer import BaseWriter, ItemsLimitReached
class ConsoleWriter(BaseWriter):
"""
It is just a writer with testing purposes. It prints every item in console.
"""
def __init__(self, options):
super(ConsoleWriter, self).__init__(options)
self.logger.info('ConsoleWriter has been initiated')
self.pretty_print = self.options.get('pretty_print', False)
def write_batch(self, batch):
for item in batch:
formatted_item = item.formatted
if self.pretty_print:
formatted_item = self._format(formatted_item)
print formatted_item
self._increment_written_items()
if self.items_limit and self.items_limit == self.stats['items_count']:
raise ItemsLimitReached('Finishing job after items_limit reached: {} items written.'.format(self.stats['items_count']))
self.logger.debug('Wrote items')
def _format(self, item):
try:
return json.dumps(json.loads(item), indent=2, sort_keys=True)
except:
return item
|
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
("admin_interface", "0024_remove_theme_css"),
]
operations = [
migrations.AddField(
model_name="theme",
name="language_chooser_control",
field=models.CharField(
choices=[
("default-select", "Default Select"),
("minimal-select", "Minimal Select"),
],
default="select",
max_length=20,
verbose_name="control",
),
),
]
| Fix `language_choose_control` migration default value.
Co-Authored-By: Éric <34afff4eac2f9b94bd269db558876db6be315161@users.noreply.github.com>
| from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
("admin_interface", "0024_remove_theme_css"),
]
operations = [
migrations.AddField(
model_name="theme",
name="language_chooser_control",
field=models.CharField(
choices=[
("default-select", "Default Select"),
("minimal-select", "Minimal Select"),
],
default="default-select",
max_length=20,
verbose_name="control",
),
),
]
|
import functools
ITERABLE_TYPES = (
list,
set,
tuple,
)
try:
from numpy import ndarray
ITERABLE_TYPES = ITERABLE_TYPES + (ndarray,)
except ImportError:
pass
def isiterable(v):
return isinstance(v, ITERABLE_TYPES)
def batchable(func=None, batch_size=100):
def do_batch(*args, **kwargs):
_batch_size = kwargs.pop('batch_size', batch_size)
if isiterable(args[0]):
_self = None
to_batch = args[0]
args = args[1:]
else:
_self = args[0]
to_batch = args[1]
args = args[2:]
if not isiterable(to_batch):
to_batch = [to_batch]
for _batch in [
to_batch[i:i+_batch_size]
for i in xrange(0, len(to_batch), _batch_size)
]:
if _self is None:
func(_batch, *args, **kwargs)
else:
func(_self, _batch, *args, **kwargs)
if func is None:
return functools.partial(batchable, batch_size=batch_size)
return do_batch
| Fix passing sets to batchable methods
Sets don't support indexing, so convert them to lists.
| import functools
ITERABLE_TYPES = (
list,
set,
tuple,
)
try:
from numpy import ndarray
ITERABLE_TYPES = ITERABLE_TYPES + (ndarray,)
except ImportError:
pass
def isiterable(v):
return isinstance(v, ITERABLE_TYPES)
def batchable(func=None, batch_size=100):
def do_batch(*args, **kwargs):
_batch_size = kwargs.pop('batch_size', batch_size)
if isiterable(args[0]):
_self = None
to_batch = args[0]
args = args[1:]
else:
_self = args[0]
to_batch = args[1]
args = args[2:]
if not isiterable(to_batch):
to_batch = [to_batch]
if isinstance(to_batch, set):
to_batch = list(to_batch)
for _batch in [
to_batch[i:i+_batch_size]
for i in xrange(0, len(to_batch), _batch_size)
]:
if _self is None:
func(_batch, *args, **kwargs)
else:
func(_self, _batch, *args, **kwargs)
if func is None:
return functools.partial(batchable, batch_size=batch_size)
return do_batch
|
from devito.tools import Tag
__all__ = ['DOMAIN', 'OWNED', 'HALO', 'NOPAD', 'FULL',
'LEFT', 'RIGHT', 'CENTER']
class DataRegion(Tag):
pass
DOMAIN = DataRegion('domain')
OWNED = DataRegion('owned') # within DOMAIN
HALO = DataRegion('halo')
NOPAD = DataRegion('nopad') # == DOMAIN+HALO
FULL = DataRegion('full') # == DOMAIN+HALO+PADDING
class DataSide(Tag):
pass
LEFT = DataSide('left')
RIGHT = DataSide('right')
CENTER = DataSide('center')
| data: Add static value to LEFT, CENTER, RIGHT
| from devito.tools import Tag
__all__ = ['DOMAIN', 'OWNED', 'HALO', 'NOPAD', 'FULL',
'LEFT', 'RIGHT', 'CENTER']
class DataRegion(Tag):
pass
DOMAIN = DataRegion('domain')
OWNED = DataRegion('owned') # within DOMAIN
HALO = DataRegion('halo')
NOPAD = DataRegion('nopad') # == DOMAIN+HALO
FULL = DataRegion('full') # == DOMAIN+HALO+PADDING
class DataSide(Tag):
pass
LEFT = DataSide('left', -1)
CENTER = DataSide('center', 0)
RIGHT = DataSide('right', 1)
|
from .standard_deviation import standard_deviation
from .mean import mean
def coefficient_of_variation(data):
"""
The `coefficient_of_variation`_ is the ratio of the standard deviation to the mean
.. _`coefficient of variation`: https://en.wikipedia.org/wiki/Coefficient_of_variation
Args:
data: A list of numerical objects.
Returns:
A float object.
Examples:
>>> coefficient_of_variation([1, 2, 3])
0.5
>>> coefficient_of_variation([1, 2, 3, 4])
0.5163977794943222
>>> coefficient_of_variation([-1, 0, 1, 2, 3, 4])
1.247219128924647
"""
return standard_deviation(data) / mean(data)
| Add sample param to CV function
Boolean param to make possible to calculate coefficient of variation
for population (default is sample).
| from .standard_deviation import standard_deviation
from .mean import mean
def coefficient_of_variation(data, sample = True):
"""
The `coefficient of variation`_ is the ratio of the standard deviation to the mean.
.. _`coefficient of variation`: https://en.wikipedia.org/wiki/Coefficient_of_variation
Args:
data: A list of numerical objects.
Returns:
A float object.
Examples:
>>> coefficient_of_variation([1, 2, 3])
0.5
>>> ss.coefficient_of_variation([1, 2, 3], False)
0.408248290463863
>>> coefficient_of_variation([1, 2, 3, 4])
0.5163977794943222
>>> coefficient_of_variation([-1, 0, 1, 2, 3, 4])
1.247219128924647
"""
return standard_deviation(data, sample) / mean(data)
|
from allauth.socialaccount.models import SocialAccount
from django.core.management.base import BaseCommand
from django.db import IntegrityError
from social_django.models import UserSocialAuth
class Command(BaseCommand):
help = 'Migrate allauth social logins to social auth'
def handle(self, *args, **options):
self.stdout.write(self.style.SUCCESS('Going through all SocialAccount objects...'))
# Retrieve existing objects
providers = {}
for usa in UserSocialAuth.objects.all():
provider = providers.setdefault(usa.provider, {})
provider[usa.user_id] = usa
for sa in SocialAccount.objects.all():
provider = providers.setdefault(sa.provider, {})
if sa.user_id in provider:
continue
provider[sa.user_id] = UserSocialAuth.objects.create(
user=sa.user,
provider=sa.provider,
uid=sa.uid,
extra_data=sa.extra_data,
)
self.stdout.write(self.style.SUCCESS('Added. (provider: {}, uid: {})'.format(sa.provider, sa.uid)))
self.stdout.write(self.style.SUCCESS('Done.'))
| Implement app secret printing to social_auth migration tool
| from allauth.socialaccount.models import SocialAccount, SocialApp
from django.core.management.base import BaseCommand
from django.db import IntegrityError
from social_django.models import UserSocialAuth
class Command(BaseCommand):
help = 'Migrate allauth social logins to social auth'
def add_arguments(self, parser):
parser.add_argument('--apps', action='store_true', dest='apps',
help='Print social app keys and secrets')
parser.add_argument('--accounts', action='store_true', dest='accounts',
help='Migrate accounts')
def migrate_accounts(self):
self.stdout.write(self.style.SUCCESS('Going through all SocialAccount objects...'))
# Retrieve existing objects
providers = {}
for usa in UserSocialAuth.objects.all():
provider = providers.setdefault(usa.provider, {})
provider[usa.user_id] = usa
for sa in SocialAccount.objects.all():
provider = providers.setdefault(sa.provider, {})
if sa.user_id in provider:
continue
provider[sa.user_id] = UserSocialAuth.objects.create(
user=sa.user,
provider=sa.provider,
uid=sa.uid,
extra_data=sa.extra_data,
)
self.stdout.write(self.style.SUCCESS('Added. (provider: {}, uid: {})'.format(sa.provider, sa.uid)))
self.stdout.write(self.style.SUCCESS('Done.'))
def migrate_apps(self):
for app in SocialApp.objects.all():
app_id = app.provider.upper()
print("SOCIAL_AUTH_%s_KEY = '%s'" % (app_id, app.client_id))
print("SOCIAL_AUTH_%s_SECRET = '%s'" % (app_id, app.secret))
print()
def handle(self, *args, **options):
if options['apps']:
self.migrate_apps()
if options['accounts']:
self.migrate_accounts()
|
from django import template
register = template.Library()
@register.assignment_tag(takes_context=True)
def build_absolute_uri(context, url):
if url:
return context['request'].build_absolute_uri(url)
| Use simple_tag decorators as those tags can now also be used for assignments
| from django import template
register = template.Library()
@register.simple_tag(takes_context=True)
def build_absolute_uri(context, url):
if url:
return context['request'].build_absolute_uri(url)
|
__author__ = 'Brian Wickman'
__version__ = '0.5.2'
__license__ = 'MIT'
from pystachio.typing import (
Type,
TypeCheck,
TypeFactory)
from pystachio.base import Environment
from pystachio.parsing import MustacheParser
from pystachio.naming import Namable, Ref
from pystachio.basic import (
Float,
Integer,
String)
from pystachio.container import (
List,
Map)
from pystachio.composite import (
Default,
Empty,
Provided,
Required,
Struct)
| Add check for minimum Python version
| __author__ = 'Brian Wickman'
__version__ = '0.5.2'
__license__ = 'MIT'
import sys
if sys.version_info < (2, 6, 5):
raise ImportError("pystachio requires Python >= 2.6.5")
from pystachio.typing import (
Type,
TypeCheck,
TypeFactory)
from pystachio.base import Environment
from pystachio.parsing import MustacheParser
from pystachio.naming import Namable, Ref
from pystachio.basic import (
Float,
Integer,
String)
from pystachio.container import (
List,
Map)
from pystachio.composite import (
Default,
Empty,
Provided,
Required,
Struct)
|
def test_right_arrows(page):
page.goto("index.html")
while(True):
# Keeps going to the next page until there is no right arrow
right_arrow = page.query_selector("//*[@id='relations-next']/a")
if(right_arrow):
page.click("//*[@id='relations-next']/a")
page.wait_for_load_state()
else:
break
# TODO make a similar test but going from de last page
# to the previous one until it gets to the first one
| Implement assertions and a for instead of a while loop
| def get_menu_titles(page) -> list:
page.goto("index.html")
page.wait_for_load_state()
menu_list = page.query_selector_all("//*[@class='toctree-wrapper compound']/ul/li/a")
menu_titles = []
for i in menu_list:
menu_item = i.as_element().inner_text()
menu_titles.append(menu_item)
return menu_titles
def test_check_titles(page):
menu_list = get_menu_titles(page)
page.goto("index.html")
page.wait_for_load_state()
for menu_item in menu_list:
right_arrow = page.query_selector("//*[@id='relations-next']/a")
if(right_arrow):
page.click("//*[@id='relations-next']/a")
page.wait_for_load_state()
page_title = page.title()
page_title = page_title.split(" — ")[0]
assert page_title == menu_item
else:
break
|
#!/usr/bin/env python
import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
f = tempfile.NamedTemporaryFile()
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
| Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file | #!/usr/bin/env python
import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
f = tempfile.NamedTemporaryFile(suffix=".wav")
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
|