commit stringlengths 40 40 | old_file stringlengths 4 118 | new_file stringlengths 4 118 | old_contents stringlengths 0 2.94k | new_contents stringlengths 1 4.43k | subject stringlengths 15 444 | message stringlengths 16 3.45k | lang stringclasses 1
value | license stringclasses 13
values | repos stringlengths 5 43.2k | prompt stringlengths 17 4.58k | response stringlengths 1 4.43k | prompt_tagged stringlengths 58 4.62k | response_tagged stringlengths 1 4.43k | text stringlengths 132 7.29k | text_tagged stringlengths 173 7.33k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ce54b2e002a46f3abecccbd7cdd591c1ec4c7074 | beetsplug/embedcoverart.py | beetsplug/embedcoverart.py | from beets.plugins import BeetsPlugin
from beets import mediafile
import os, logging
from email.mime.image import MIMEImage
log = logging.getLogger('beets')
log.addHandler(logging.StreamHandler())
class EmbedAlbumartPlugin(BeetsPlugin):
'''Allows albumart to be embedded into the actual files'''
def __init... | Add initial version of the embed coverart plugin. | Add initial version of the embed coverart plugin.
This plugin allows users to embed the cover into the audio file.
Probaly still has a few bugs but it should work in most cases right
now.
| Python | mit | google-code-export/beets,google-code-export/beets,google-code-export/beets | Add initial version of the embed coverart plugin.
This plugin allows users to embed the cover into the audio file.
Probaly still has a few bugs but it should work in most cases right
now. | from beets.plugins import BeetsPlugin
from beets import mediafile
import os, logging
from email.mime.image import MIMEImage
log = logging.getLogger('beets')
log.addHandler(logging.StreamHandler())
class EmbedAlbumartPlugin(BeetsPlugin):
'''Allows albumart to be embedded into the actual files'''
def __init... | <commit_before><commit_msg>Add initial version of the embed coverart plugin.
This plugin allows users to embed the cover into the audio file.
Probaly still has a few bugs but it should work in most cases right
now.<commit_after> | from beets.plugins import BeetsPlugin
from beets import mediafile
import os, logging
from email.mime.image import MIMEImage
log = logging.getLogger('beets')
log.addHandler(logging.StreamHandler())
class EmbedAlbumartPlugin(BeetsPlugin):
'''Allows albumart to be embedded into the actual files'''
def __init... | Add initial version of the embed coverart plugin.
This plugin allows users to embed the cover into the audio file.
Probaly still has a few bugs but it should work in most cases right
now.from beets.plugins import BeetsPlugin
from beets import mediafile
import os, logging
from email.mime.image import MIMEImage
log =... | <commit_before><commit_msg>Add initial version of the embed coverart plugin.
This plugin allows users to embed the cover into the audio file.
Probaly still has a few bugs but it should work in most cases right
now.<commit_after>from beets.plugins import BeetsPlugin
from beets import mediafile
import os, logging
from... | |
c62dfb75477f5a641a2a9fd2de16d96cc2638e5f | migrations/versions/ebcc92fc4d27_remove_tag_table.py | migrations/versions/ebcc92fc4d27_remove_tag_table.py | """Remove tag table
Revision ID: ebcc92fc4d27
Revises: 444c69da7c45
Create Date: 2017-05-08 01:01:48.865909
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'ebcc92fc4d27'
down_revision = '444c69da7c45'
branch_labels = None
depends_on = None
def upgrade():
... | Remove tags field from user Model | [Chore] Migrate: Remove tags field from user Model
| Python | mit | andela-wcyn/bucketlist | [Chore] Migrate: Remove tags field from user Model | """Remove tag table
Revision ID: ebcc92fc4d27
Revises: 444c69da7c45
Create Date: 2017-05-08 01:01:48.865909
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'ebcc92fc4d27'
down_revision = '444c69da7c45'
branch_labels = None
depends_on = None
def upgrade():
... | <commit_before><commit_msg>[Chore] Migrate: Remove tags field from user Model<commit_after> | """Remove tag table
Revision ID: ebcc92fc4d27
Revises: 444c69da7c45
Create Date: 2017-05-08 01:01:48.865909
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'ebcc92fc4d27'
down_revision = '444c69da7c45'
branch_labels = None
depends_on = None
def upgrade():
... | [Chore] Migrate: Remove tags field from user Model"""Remove tag table
Revision ID: ebcc92fc4d27
Revises: 444c69da7c45
Create Date: 2017-05-08 01:01:48.865909
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'ebcc92fc4d27'
down_revision = '444c69da7c45'
branch_la... | <commit_before><commit_msg>[Chore] Migrate: Remove tags field from user Model<commit_after>"""Remove tag table
Revision ID: ebcc92fc4d27
Revises: 444c69da7c45
Create Date: 2017-05-08 01:01:48.865909
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'ebcc92fc4d27'... | |
299882e55ea82d8c442b51b243caf63707befe4c | xoinvader/tests/test_utils.py | xoinvader/tests/test_utils.py | import sys
import unittest
import pprint
from xoinvader.utils import create_logger
from xoinvader.utils import InfiniteList
from xoinvader.utils import Point
class TestUtils(unittest.TestCase):
def test_create_logger(self):
logger = create_logger("test", "test.log")
self.assertTrue(logger)
d... | Add tests for utils module. | Add tests for utils module.
| Python | mit | pkulev/xoinvader,pankshok/xoinvader | Add tests for utils module. | import sys
import unittest
import pprint
from xoinvader.utils import create_logger
from xoinvader.utils import InfiniteList
from xoinvader.utils import Point
class TestUtils(unittest.TestCase):
def test_create_logger(self):
logger = create_logger("test", "test.log")
self.assertTrue(logger)
d... | <commit_before><commit_msg>Add tests for utils module.<commit_after> | import sys
import unittest
import pprint
from xoinvader.utils import create_logger
from xoinvader.utils import InfiniteList
from xoinvader.utils import Point
class TestUtils(unittest.TestCase):
def test_create_logger(self):
logger = create_logger("test", "test.log")
self.assertTrue(logger)
d... | Add tests for utils module.import sys
import unittest
import pprint
from xoinvader.utils import create_logger
from xoinvader.utils import InfiniteList
from xoinvader.utils import Point
class TestUtils(unittest.TestCase):
def test_create_logger(self):
logger = create_logger("test", "test.log")
sel... | <commit_before><commit_msg>Add tests for utils module.<commit_after>import sys
import unittest
import pprint
from xoinvader.utils import create_logger
from xoinvader.utils import InfiniteList
from xoinvader.utils import Point
class TestUtils(unittest.TestCase):
def test_create_logger(self):
logger = crea... | |
9ac9f94f3b1409ae3a47c8a9e890f578a69b020f | test/requests/test_forgot_password.py | test/requests/test_forgot_password.py | import requests
from wqflask import user_manager
from utility.elasticsearch_tools import get_user_by_unique_column
from parameterized import parameterized
from parametrized_test import ParametrizedTest
passwork_reset_link = ''
forgot_password_page = None
class TestForgotPassword(ParametrizedTest):
def setUp(self... | Add tests for Forgot Password feature | Add tests for Forgot Password feature
* Add tests to ensure the "Forgot Password" feature works appropriately.
| Python | agpl-3.0 | zsloan/genenetwork2,pjotrp/genenetwork2,DannyArends/genenetwork2,DannyArends/genenetwork2,DannyArends/genenetwork2,zsloan/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,pjotrp/genenetwork2,DannyArends/genenetwork2,genenetwork/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2,zsloan/genenetwork2,genenetwor... | Add tests for Forgot Password feature
* Add tests to ensure the "Forgot Password" feature works appropriately. | import requests
from wqflask import user_manager
from utility.elasticsearch_tools import get_user_by_unique_column
from parameterized import parameterized
from parametrized_test import ParametrizedTest
passwork_reset_link = ''
forgot_password_page = None
class TestForgotPassword(ParametrizedTest):
def setUp(self... | <commit_before><commit_msg>Add tests for Forgot Password feature
* Add tests to ensure the "Forgot Password" feature works appropriately.<commit_after> | import requests
from wqflask import user_manager
from utility.elasticsearch_tools import get_user_by_unique_column
from parameterized import parameterized
from parametrized_test import ParametrizedTest
passwork_reset_link = ''
forgot_password_page = None
class TestForgotPassword(ParametrizedTest):
def setUp(self... | Add tests for Forgot Password feature
* Add tests to ensure the "Forgot Password" feature works appropriately.import requests
from wqflask import user_manager
from utility.elasticsearch_tools import get_user_by_unique_column
from parameterized import parameterized
from parametrized_test import ParametrizedTest
passwo... | <commit_before><commit_msg>Add tests for Forgot Password feature
* Add tests to ensure the "Forgot Password" feature works appropriately.<commit_after>import requests
from wqflask import user_manager
from utility.elasticsearch_tools import get_user_by_unique_column
from parameterized import parameterized
from parametr... | |
42b50470458d91cd0882f927ed83ba267ab78764 | zinnia_wymeditor/__init__.py | zinnia_wymeditor/__init__.py | """WYMeditor for Django-blog-zinnia"""
__version__ = '1.0'
__license__ = 'BSD License'
__author__ = 'Fantomas42'
__email__ = '[email protected]'
__url__ = 'https://github.com/django-blog-zinnia/zinnia-wysiwyg-wymeditor'
| Create zinnia_wymeditor module with his metadatas | Create zinnia_wymeditor module with his metadatas
| Python | bsd-3-clause | layar/zinnia-wysiwyg-wymeditor,django-blog-zinnia/zinnia-wysiwyg-wymeditor,django-blog-zinnia/zinnia-wysiwyg-wymeditor,layar/zinnia-wysiwyg-wymeditor,django-blog-zinnia/zinnia-wysiwyg-wymeditor,layar/zinnia-wysiwyg-wymeditor,django-blog-zinnia/zinnia-wysiwyg-wymeditor,layar/zinnia-wysiwyg-wymeditor | Create zinnia_wymeditor module with his metadatas | """WYMeditor for Django-blog-zinnia"""
__version__ = '1.0'
__license__ = 'BSD License'
__author__ = 'Fantomas42'
__email__ = '[email protected]'
__url__ = 'https://github.com/django-blog-zinnia/zinnia-wysiwyg-wymeditor'
| <commit_before><commit_msg>Create zinnia_wymeditor module with his metadatas<commit_after> | """WYMeditor for Django-blog-zinnia"""
__version__ = '1.0'
__license__ = 'BSD License'
__author__ = 'Fantomas42'
__email__ = '[email protected]'
__url__ = 'https://github.com/django-blog-zinnia/zinnia-wysiwyg-wymeditor'
| Create zinnia_wymeditor module with his metadatas"""WYMeditor for Django-blog-zinnia"""
__version__ = '1.0'
__license__ = 'BSD License'
__author__ = 'Fantomas42'
__email__ = '[email protected]'
__url__ = 'https://github.com/django-blog-zinnia/zinnia-wysiwyg-wymeditor'
| <commit_before><commit_msg>Create zinnia_wymeditor module with his metadatas<commit_after>"""WYMeditor for Django-blog-zinnia"""
__version__ = '1.0'
__license__ = 'BSD License'
__author__ = 'Fantomas42'
__email__ = '[email protected]'
__url__ = 'https://github.com/django-blog-zinnia/zinnia-wysiwyg-wymeditor'
| |
435cc2548b38d92f8ffdc4bee8845f5a58d655ce | ocrsite/ocrlab/management/commands/runfilescript.py | ocrsite/ocrlab/management/commands/runfilescript.py | """
Run a script that takes a file input/output.
"""
import os
import sys
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from django.contrib.auth.models import User
from django.core.exceptions import ImproperlyConfigured
from django.utils import simplejson as json
... | Add a management command that runs a given script taking an input file and an output file as args. | Add a management command that runs a given script taking an input file and an output file as args.
| Python | mit | mikesname/python-ocrlab | Add a management command that runs a given script taking an input file and an output file as args. | """
Run a script that takes a file input/output.
"""
import os
import sys
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from django.contrib.auth.models import User
from django.core.exceptions import ImproperlyConfigured
from django.utils import simplejson as json
... | <commit_before><commit_msg>Add a management command that runs a given script taking an input file and an output file as args.<commit_after> | """
Run a script that takes a file input/output.
"""
import os
import sys
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from django.contrib.auth.models import User
from django.core.exceptions import ImproperlyConfigured
from django.utils import simplejson as json
... | Add a management command that runs a given script taking an input file and an output file as args."""
Run a script that takes a file input/output.
"""
import os
import sys
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from django.contrib.auth.models import User
fro... | <commit_before><commit_msg>Add a management command that runs a given script taking an input file and an output file as args.<commit_after>"""
Run a script that takes a file input/output.
"""
import os
import sys
from optparse import make_option
from django.core.management.base import BaseCommand, CommandError
from d... | |
862753353a09400d0d99960ef2cd6d70fb9b4a7e | tests/top_destinations_tests.py | tests/top_destinations_tests.py | import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicTopD... | Add tests for top destinations | Add tests for top destinations
| Python | mit | Jamil/sabre_dev_studio | Add tests for top destinations | import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicTopD... | <commit_before><commit_msg>Add tests for top destinations<commit_after> | import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicTopD... | Add tests for top destinationsimport unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": --... | <commit_before><commit_msg>Add tests for top destinations<commit_after>import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_clie... | |
6825c853892c263a3c193490ee1f620b303aa79d | translate_labels.py | translate_labels.py | """Find known body parts in sentences with predicted label 'Lichaamsdeel'.
Extended body parts are saved to new text files.
Usage: python classify_body_parts.py <json file with body part mapping> <dir
with input texts> <dir for output texts>
"""
import os
import codecs
import argparse
import json
from count_labels im... | Add script to translate labels | Add script to translate labels
| Python | apache-2.0 | NLeSC/embodied-emotions-scripts,NLeSC/embodied-emotions-scripts | Add script to translate labels | """Find known body parts in sentences with predicted label 'Lichaamsdeel'.
Extended body parts are saved to new text files.
Usage: python classify_body_parts.py <json file with body part mapping> <dir
with input texts> <dir for output texts>
"""
import os
import codecs
import argparse
import json
from count_labels im... | <commit_before><commit_msg>Add script to translate labels<commit_after> | """Find known body parts in sentences with predicted label 'Lichaamsdeel'.
Extended body parts are saved to new text files.
Usage: python classify_body_parts.py <json file with body part mapping> <dir
with input texts> <dir for output texts>
"""
import os
import codecs
import argparse
import json
from count_labels im... | Add script to translate labels"""Find known body parts in sentences with predicted label 'Lichaamsdeel'.
Extended body parts are saved to new text files.
Usage: python classify_body_parts.py <json file with body part mapping> <dir
with input texts> <dir for output texts>
"""
import os
import codecs
import argparse
im... | <commit_before><commit_msg>Add script to translate labels<commit_after>"""Find known body parts in sentences with predicted label 'Lichaamsdeel'.
Extended body parts are saved to new text files.
Usage: python classify_body_parts.py <json file with body part mapping> <dir
with input texts> <dir for output texts>
"""
i... | |
14bc31730b783b1a6649c97a51d1ef9ae2053b42 | tests/asttools/test_compiler.py | tests/asttools/test_compiler.py | """Test suite for asttools.compiler."""
from __future__ import division
from __future__ import absolute_import
from __future__ import print_function
from __future__ import unicode_literals
import pytest
from pycc.asttools import parse
from pycc.asttools import compiler
source = """
x = True
for y in range(10):
... | Add simple tests for asttools.compiler module | Add simple tests for asttools.compiler module
These tests don't do anything except ensure that exceptions aren't
thrown during compiling.
Signed-off-by: Kevin Conway <[email protected]>
| Python | apache-2.0 | kevinconway/pycc,kevinconway/pycc | Add simple tests for asttools.compiler module
These tests don't do anything except ensure that exceptions aren't
thrown during compiling.
Signed-off-by: Kevin Conway <[email protected]> | """Test suite for asttools.compiler."""
from __future__ import division
from __future__ import absolute_import
from __future__ import print_function
from __future__ import unicode_literals
import pytest
from pycc.asttools import parse
from pycc.asttools import compiler
source = """
x = True
for y in range(10):
... | <commit_before><commit_msg>Add simple tests for asttools.compiler module
These tests don't do anything except ensure that exceptions aren't
thrown during compiling.
Signed-off-by: Kevin Conway <[email protected]><commit_after> | """Test suite for asttools.compiler."""
from __future__ import division
from __future__ import absolute_import
from __future__ import print_function
from __future__ import unicode_literals
import pytest
from pycc.asttools import parse
from pycc.asttools import compiler
source = """
x = True
for y in range(10):
... | Add simple tests for asttools.compiler module
These tests don't do anything except ensure that exceptions aren't
thrown during compiling.
Signed-off-by: Kevin Conway <[email protected]>"""Test suite for asttools.compiler."""
from __future__ import division
from __future__ import abso... | <commit_before><commit_msg>Add simple tests for asttools.compiler module
These tests don't do anything except ensure that exceptions aren't
thrown during compiling.
Signed-off-by: Kevin Conway <[email protected]><commit_after>"""Test suite for asttools.compiler."""
from __future__ im... | |
b7274e91bc2fe6a8059ce7e65e02c20d000322b1 | tests/thread/mutate_instance.py | tests/thread/mutate_instance.py | # test concurrent mutating access to a shared user instance
#
# MIT license; Copyright (c) 2016 Damien P. George on behalf of Pycom Ltd
import _thread
# the shared user class and instance
class User:
def __init__(self):
self.a = 'A'
self.b = 'B'
self.c = 'C'
user = User()
# main thread fu... | Add test for concurrent mutating of user instance. | tests/thread: Add test for concurrent mutating of user instance.
| Python | mit | mhoffma/micropython,ryannathans/micropython,deshipu/micropython,HenrikSolver/micropython,jmarcelino/pycom-micropython,adafruit/micropython,bvernoux/micropython,TDAbboud/micropython,oopy/micropython,emfcamp/micropython,swegener/micropython,Timmenem/micropython,bvernoux/micropython,alex-robbins/micropython,lowRISC/microp... | tests/thread: Add test for concurrent mutating of user instance. | # test concurrent mutating access to a shared user instance
#
# MIT license; Copyright (c) 2016 Damien P. George on behalf of Pycom Ltd
import _thread
# the shared user class and instance
class User:
def __init__(self):
self.a = 'A'
self.b = 'B'
self.c = 'C'
user = User()
# main thread fu... | <commit_before><commit_msg>tests/thread: Add test for concurrent mutating of user instance.<commit_after> | # test concurrent mutating access to a shared user instance
#
# MIT license; Copyright (c) 2016 Damien P. George on behalf of Pycom Ltd
import _thread
# the shared user class and instance
class User:
def __init__(self):
self.a = 'A'
self.b = 'B'
self.c = 'C'
user = User()
# main thread fu... | tests/thread: Add test for concurrent mutating of user instance.# test concurrent mutating access to a shared user instance
#
# MIT license; Copyright (c) 2016 Damien P. George on behalf of Pycom Ltd
import _thread
# the shared user class and instance
class User:
def __init__(self):
self.a = 'A'
s... | <commit_before><commit_msg>tests/thread: Add test for concurrent mutating of user instance.<commit_after># test concurrent mutating access to a shared user instance
#
# MIT license; Copyright (c) 2016 Damien P. George on behalf of Pycom Ltd
import _thread
# the shared user class and instance
class User:
def __ini... | |
05084740393180994ea0b330c7f91a148a5d7d86 | util/checkZscale.py | util/checkZscale.py | # Copyright 2014 Open Connectome Project (http://openconnecto.me)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by app... | Check the zscale and cubedims for a given database | [util] Check the zscale and cubedims for a given database
| Python | apache-2.0 | openconnectome/open-connectome,openconnectome/open-connectome,neurodata/ndstore,openconnectome/open-connectome,openconnectome/open-connectome,neurodata/ndstore,neurodata/ndstore,openconnectome/open-connectome,neurodata/ndstore,openconnectome/open-connectome | [util] Check the zscale and cubedims for a given database | # Copyright 2014 Open Connectome Project (http://openconnecto.me)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by app... | <commit_before><commit_msg>[util] Check the zscale and cubedims for a given database<commit_after> | # Copyright 2014 Open Connectome Project (http://openconnecto.me)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by app... | [util] Check the zscale and cubedims for a given database# Copyright 2014 Open Connectome Project (http://openconnecto.me)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.a... | <commit_before><commit_msg>[util] Check the zscale and cubedims for a given database<commit_after># Copyright 2014 Open Connectome Project (http://openconnecto.me)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a co... | |
59f40f5b7683d60264f5c38a5e0da7810fa67fd8 | cross-multinomial/cross_multinomial_predict.py | cross-multinomial/cross_multinomial_predict.py | import sys
import numpy as np
import numpy.random as npr
from calculate_phist import read_counts
from calculate_phist import normalize_haplotypes
from scipy.special import gammaln
import matplotlib.pyplot as plt
def log_factorial(n):
return gammaln(n+1)
def log_multinomial(xs, ps):
n = np.sum(xs)
log_prob = log_fa... | Add prediction script for cross multinomial | Add prediction script for cross multinomial
| Python | apache-2.0 | rnowling/pop-gen-models | Add prediction script for cross multinomial | import sys
import numpy as np
import numpy.random as npr
from calculate_phist import read_counts
from calculate_phist import normalize_haplotypes
from scipy.special import gammaln
import matplotlib.pyplot as plt
def log_factorial(n):
return gammaln(n+1)
def log_multinomial(xs, ps):
n = np.sum(xs)
log_prob = log_fa... | <commit_before><commit_msg>Add prediction script for cross multinomial<commit_after> | import sys
import numpy as np
import numpy.random as npr
from calculate_phist import read_counts
from calculate_phist import normalize_haplotypes
from scipy.special import gammaln
import matplotlib.pyplot as plt
def log_factorial(n):
return gammaln(n+1)
def log_multinomial(xs, ps):
n = np.sum(xs)
log_prob = log_fa... | Add prediction script for cross multinomialimport sys
import numpy as np
import numpy.random as npr
from calculate_phist import read_counts
from calculate_phist import normalize_haplotypes
from scipy.special import gammaln
import matplotlib.pyplot as plt
def log_factorial(n):
return gammaln(n+1)
def log_multinomial(... | <commit_before><commit_msg>Add prediction script for cross multinomial<commit_after>import sys
import numpy as np
import numpy.random as npr
from calculate_phist import read_counts
from calculate_phist import normalize_haplotypes
from scipy.special import gammaln
import matplotlib.pyplot as plt
def log_factorial(n):
... | |
0ffeb772e52c2af555c5fe65f4763059edb31148 | tensor2tensor/models/research/transformer_vae_test.py | tensor2tensor/models/research/transformer_vae_test.py | # coding=utf-8
# Copyright 2018 The Tensor2Tensor Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | Add a test for transformer_vae | Add a test for transformer_vae
PiperOrigin-RevId: 190462534
| Python | apache-2.0 | tensorflow/tensor2tensor,tensorflow/tensor2tensor,vthorsteinsson/tensor2tensor,tensorflow/tensor2tensor,vthorsteinsson/tensor2tensor,tensorflow/tensor2tensor,tensorflow/tensor2tensor,vthorsteinsson/tensor2tensor,vthorsteinsson/tensor2tensor | Add a test for transformer_vae
PiperOrigin-RevId: 190462534 | # coding=utf-8
# Copyright 2018 The Tensor2Tensor Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | <commit_before><commit_msg>Add a test for transformer_vae
PiperOrigin-RevId: 190462534<commit_after> | # coding=utf-8
# Copyright 2018 The Tensor2Tensor Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | Add a test for transformer_vae
PiperOrigin-RevId: 190462534# coding=utf-8
# Copyright 2018 The Tensor2Tensor Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apach... | <commit_before><commit_msg>Add a test for transformer_vae
PiperOrigin-RevId: 190462534<commit_after># coding=utf-8
# Copyright 2018 The Tensor2Tensor Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy o... | |
3ab745087ef563e30f86fdeed09b12868095f85d | tests/test_classifierbase.py | tests/test_classifierbase.py | from itertools import izip, cycle
from unittest import TestCase
class TestClassifierBase(TestCase):
def _training(self, classification_type, classifier, alphabet,
reversed_alphabet):
reversed_alphabet = reversed(alphabet)
messagePairs = izip(alphabet, reversed_alphabet)
... | Add a base class for certain types of tests | Add a base class for certain types of tests
This is so training data can be reused across classes.
In fact nonsensefilter's test have been using this for quite some time
but this was not committed. Oops!
| Python | mpl-2.0 | mozilla/spicedham,mozilla/spicedham | Add a base class for certain types of tests
This is so training data can be reused across classes.
In fact nonsensefilter's test have been using this for quite some time
but this was not committed. Oops! | from itertools import izip, cycle
from unittest import TestCase
class TestClassifierBase(TestCase):
def _training(self, classification_type, classifier, alphabet,
reversed_alphabet):
reversed_alphabet = reversed(alphabet)
messagePairs = izip(alphabet, reversed_alphabet)
... | <commit_before><commit_msg>Add a base class for certain types of tests
This is so training data can be reused across classes.
In fact nonsensefilter's test have been using this for quite some time
but this was not committed. Oops!<commit_after> | from itertools import izip, cycle
from unittest import TestCase
class TestClassifierBase(TestCase):
def _training(self, classification_type, classifier, alphabet,
reversed_alphabet):
reversed_alphabet = reversed(alphabet)
messagePairs = izip(alphabet, reversed_alphabet)
... | Add a base class for certain types of tests
This is so training data can be reused across classes.
In fact nonsensefilter's test have been using this for quite some time
but this was not committed. Oops!from itertools import izip, cycle
from unittest import TestCase
class TestClassifierBase(TestCase):
def _tra... | <commit_before><commit_msg>Add a base class for certain types of tests
This is so training data can be reused across classes.
In fact nonsensefilter's test have been using this for quite some time
but this was not committed. Oops!<commit_after>from itertools import izip, cycle
from unittest import TestCase
class Te... | |
1e3d9f01d25d89ac6167226a24e03b79d0c34a3c | test/trainer_test.py | test/trainer_test.py | import theanets
import util
class TestTrainer(util.MNIST):
def setUp(self):
super(TestTrainer, self).setUp()
self.exp = theanets.Experiment(
theanets.Autoencoder,
layers=(self.DIGIT_SIZE, 10, self.DIGIT_SIZE))
def assert_progress(self, *args, **kwargs):
self.e... | Add basic "does this even run" test for trainers. | Add basic "does this even run" test for trainers.
| Python | mit | lmjohns3/theanets,chrinide/theanets,devdoer/theanets | Add basic "does this even run" test for trainers. | import theanets
import util
class TestTrainer(util.MNIST):
def setUp(self):
super(TestTrainer, self).setUp()
self.exp = theanets.Experiment(
theanets.Autoencoder,
layers=(self.DIGIT_SIZE, 10, self.DIGIT_SIZE))
def assert_progress(self, *args, **kwargs):
self.e... | <commit_before><commit_msg>Add basic "does this even run" test for trainers.<commit_after> | import theanets
import util
class TestTrainer(util.MNIST):
def setUp(self):
super(TestTrainer, self).setUp()
self.exp = theanets.Experiment(
theanets.Autoencoder,
layers=(self.DIGIT_SIZE, 10, self.DIGIT_SIZE))
def assert_progress(self, *args, **kwargs):
self.e... | Add basic "does this even run" test for trainers.import theanets
import util
class TestTrainer(util.MNIST):
def setUp(self):
super(TestTrainer, self).setUp()
self.exp = theanets.Experiment(
theanets.Autoencoder,
layers=(self.DIGIT_SIZE, 10, self.DIGIT_SIZE))
def asser... | <commit_before><commit_msg>Add basic "does this even run" test for trainers.<commit_after>import theanets
import util
class TestTrainer(util.MNIST):
def setUp(self):
super(TestTrainer, self).setUp()
self.exp = theanets.Experiment(
theanets.Autoencoder,
layers=(self.DIGIT_S... | |
916053366034f857ad431bff069121e56c4e027b | pylearn2/scripts/tests/test_show_examples.py | pylearn2/scripts/tests/test_show_examples.py | """
Tests for the show_examples.py script
"""
import os
from pylearn2.testing.skip import skip_if_no_matplotlib, skip_if_no_data
from pylearn2.scripts.show_examples import show_examples
def test_show_examples():
"""
Create a YAML file of the MNIST dataset and show examples
"""
skip_if_no_matplotlib()... | Add unit test for show_examples.py | Add unit test for show_examples.py
| Python | bsd-3-clause | cosmoharrigan/pylearn2,skearnes/pylearn2,matrogers/pylearn2,ashhher3/pylearn2,w1kke/pylearn2,JesseLivezey/plankton,CIFASIS/pylearn2,lancezlin/pylearn2,alexjc/pylearn2,jamessergeant/pylearn2,theoryno3/pylearn2,ashhher3/pylearn2,Refefer/pylearn2,lamblin/pylearn2,alexjc/pylearn2,nouiz/pylearn2,fishcorn/pylearn2,msingh172/... | Add unit test for show_examples.py | """
Tests for the show_examples.py script
"""
import os
from pylearn2.testing.skip import skip_if_no_matplotlib, skip_if_no_data
from pylearn2.scripts.show_examples import show_examples
def test_show_examples():
"""
Create a YAML file of the MNIST dataset and show examples
"""
skip_if_no_matplotlib()... | <commit_before><commit_msg>Add unit test for show_examples.py<commit_after> | """
Tests for the show_examples.py script
"""
import os
from pylearn2.testing.skip import skip_if_no_matplotlib, skip_if_no_data
from pylearn2.scripts.show_examples import show_examples
def test_show_examples():
"""
Create a YAML file of the MNIST dataset and show examples
"""
skip_if_no_matplotlib()... | Add unit test for show_examples.py"""
Tests for the show_examples.py script
"""
import os
from pylearn2.testing.skip import skip_if_no_matplotlib, skip_if_no_data
from pylearn2.scripts.show_examples import show_examples
def test_show_examples():
"""
Create a YAML file of the MNIST dataset and show examples
... | <commit_before><commit_msg>Add unit test for show_examples.py<commit_after>"""
Tests for the show_examples.py script
"""
import os
from pylearn2.testing.skip import skip_if_no_matplotlib, skip_if_no_data
from pylearn2.scripts.show_examples import show_examples
def test_show_examples():
"""
Create a YAML file... | |
9dcb5ee0738e339f3d17876d5470b6627393862c | tensorflow_cloud/python/tests/integration/on_script_call_test.py | tensorflow_cloud/python/tests/integration/on_script_call_test.py | # Copyright 2020 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | Add details to integration test as a template | Add details to integration test as a template
| Python | apache-2.0 | tensorflow/cloud,tensorflow/cloud | Add details to integration test as a template | # Copyright 2020 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | <commit_before><commit_msg>Add details to integration test as a template<commit_after> | # Copyright 2020 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | Add details to integration test as a template# Copyright 2020 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2... | <commit_before><commit_msg>Add details to integration test as a template<commit_after># Copyright 2020 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# ... | |
e09dc29716f65b5e36436538df211363c9d874ab | vm_server/send/compile_proto.py | vm_server/send/compile_proto.py | #!/usr/bin/python
"""Script to create the protofile
It compiles the proto definition and
then creates the proto file from the text specified
in query1.txt, query2.txt etc
"""
import logging
import os
import sys
def compile_proto():
"""Executes commands to compile a proto file"""
logging.debug("Compile... | Add code to compile proto | Add code to compile proto
| Python | apache-2.0 | googleinterns/automated-windows-vms,googleinterns/automated-windows-vms | Add code to compile proto | #!/usr/bin/python
"""Script to create the protofile
It compiles the proto definition and
then creates the proto file from the text specified
in query1.txt, query2.txt etc
"""
import logging
import os
import sys
def compile_proto():
"""Executes commands to compile a proto file"""
logging.debug("Compile... | <commit_before><commit_msg>Add code to compile proto<commit_after> | #!/usr/bin/python
"""Script to create the protofile
It compiles the proto definition and
then creates the proto file from the text specified
in query1.txt, query2.txt etc
"""
import logging
import os
import sys
def compile_proto():
"""Executes commands to compile a proto file"""
logging.debug("Compile... | Add code to compile proto#!/usr/bin/python
"""Script to create the protofile
It compiles the proto definition and
then creates the proto file from the text specified
in query1.txt, query2.txt etc
"""
import logging
import os
import sys
def compile_proto():
"""Executes commands to compile a proto file"""... | <commit_before><commit_msg>Add code to compile proto<commit_after>#!/usr/bin/python
"""Script to create the protofile
It compiles the proto definition and
then creates the proto file from the text specified
in query1.txt, query2.txt etc
"""
import logging
import os
import sys
def compile_proto():
"""Exe... | |
3cc76b052ea7a49ad50247bff218f34056470f31 | accelerator/migrations/0033_migrate_gender_data.py | accelerator/migrations/0033_migrate_gender_data.py | # Generated by Django 2.2.10 on 2021-01-22 12:13
from django.db import migrations
# gender identity
GENDER_MALE = "Male"
GENDER_FEMALE = "Female"
GENDER_PREFER_TO_SELF_DESCRIBE = "I Prefer To Self-describe"
GENDER_PREFER_NOT_TO_SAY = "I Prefer Not To Say"
# gender
MALE_CHOICE = "m"
FEMALE_CHOICE = "f"
OTHER_CHOICE = ... | Merge remote-tracking branch 'origin' into AC-8354 | [AC-8354] Merge remote-tracking branch 'origin' into AC-8354
| Python | mit | masschallenge/django-accelerator,masschallenge/django-accelerator | [AC-8354] Merge remote-tracking branch 'origin' into AC-8354 | # Generated by Django 2.2.10 on 2021-01-22 12:13
from django.db import migrations
# gender identity
GENDER_MALE = "Male"
GENDER_FEMALE = "Female"
GENDER_PREFER_TO_SELF_DESCRIBE = "I Prefer To Self-describe"
GENDER_PREFER_NOT_TO_SAY = "I Prefer Not To Say"
# gender
MALE_CHOICE = "m"
FEMALE_CHOICE = "f"
OTHER_CHOICE = ... | <commit_before><commit_msg>[AC-8354] Merge remote-tracking branch 'origin' into AC-8354<commit_after> | # Generated by Django 2.2.10 on 2021-01-22 12:13
from django.db import migrations
# gender identity
GENDER_MALE = "Male"
GENDER_FEMALE = "Female"
GENDER_PREFER_TO_SELF_DESCRIBE = "I Prefer To Self-describe"
GENDER_PREFER_NOT_TO_SAY = "I Prefer Not To Say"
# gender
MALE_CHOICE = "m"
FEMALE_CHOICE = "f"
OTHER_CHOICE = ... | [AC-8354] Merge remote-tracking branch 'origin' into AC-8354# Generated by Django 2.2.10 on 2021-01-22 12:13
from django.db import migrations
# gender identity
GENDER_MALE = "Male"
GENDER_FEMALE = "Female"
GENDER_PREFER_TO_SELF_DESCRIBE = "I Prefer To Self-describe"
GENDER_PREFER_NOT_TO_SAY = "I Prefer Not To Say"
# ... | <commit_before><commit_msg>[AC-8354] Merge remote-tracking branch 'origin' into AC-8354<commit_after># Generated by Django 2.2.10 on 2021-01-22 12:13
from django.db import migrations
# gender identity
GENDER_MALE = "Male"
GENDER_FEMALE = "Female"
GENDER_PREFER_TO_SELF_DESCRIBE = "I Prefer To Self-describe"
GENDER_PREF... | |
8c47266c9bf8f91e03a0770befced1eef443b055 | problem4.py | problem4.py | """
Given: Positive integers n>=40 and k<=5.
Return: The total number of rabbit pairs that will be present after n months if we begin with 1 pair and in
each generation, every pair of reproduction-age rabbits produces a litter of k rabbit pairs
(instead of only 1 pair).
"""
# Based off a fibonnaci sequence with rabb... | Add solution to Rabbits and Recurrence Relations | Add solution to Rabbits and Recurrence Relations
| Python | mit | MichaelAquilina/rosalind-solutions | Add solution to Rabbits and Recurrence Relations | """
Given: Positive integers n>=40 and k<=5.
Return: The total number of rabbit pairs that will be present after n months if we begin with 1 pair and in
each generation, every pair of reproduction-age rabbits produces a litter of k rabbit pairs
(instead of only 1 pair).
"""
# Based off a fibonnaci sequence with rabb... | <commit_before><commit_msg>Add solution to Rabbits and Recurrence Relations<commit_after> | """
Given: Positive integers n>=40 and k<=5.
Return: The total number of rabbit pairs that will be present after n months if we begin with 1 pair and in
each generation, every pair of reproduction-age rabbits produces a litter of k rabbit pairs
(instead of only 1 pair).
"""
# Based off a fibonnaci sequence with rabb... | Add solution to Rabbits and Recurrence Relations"""
Given: Positive integers n>=40 and k<=5.
Return: The total number of rabbit pairs that will be present after n months if we begin with 1 pair and in
each generation, every pair of reproduction-age rabbits produces a litter of k rabbit pairs
(instead of only 1 pair).
... | <commit_before><commit_msg>Add solution to Rabbits and Recurrence Relations<commit_after>"""
Given: Positive integers n>=40 and k<=5.
Return: The total number of rabbit pairs that will be present after n months if we begin with 1 pair and in
each generation, every pair of reproduction-age rabbits produces a litter of ... | |
32a1ccc5856fd58e841a3960dbe17ceb029eb8d8 | examples/test_parse_bool.py | examples/test_parse_bool.py | def test_parse_bool(wish):
parse_bool = wish
assert not parse_bool('false')
assert not parse_bool('FALSE')
assert not parse_bool('0')
assert parse_bool('true')
assert parse_bool('TRUE')
assert parse_bool('1')
| Add test file use in the docs. | Add test file use in the docs.
| Python | mit | alexamici/pytest-wish,alexamici/pytest-nodev,nodev-io/pytest-nodev | Add test file use in the docs. | def test_parse_bool(wish):
parse_bool = wish
assert not parse_bool('false')
assert not parse_bool('FALSE')
assert not parse_bool('0')
assert parse_bool('true')
assert parse_bool('TRUE')
assert parse_bool('1')
| <commit_before><commit_msg>Add test file use in the docs.<commit_after> | def test_parse_bool(wish):
parse_bool = wish
assert not parse_bool('false')
assert not parse_bool('FALSE')
assert not parse_bool('0')
assert parse_bool('true')
assert parse_bool('TRUE')
assert parse_bool('1')
| Add test file use in the docs.def test_parse_bool(wish):
parse_bool = wish
assert not parse_bool('false')
assert not parse_bool('FALSE')
assert not parse_bool('0')
assert parse_bool('true')
assert parse_bool('TRUE')
assert parse_bool('1')
| <commit_before><commit_msg>Add test file use in the docs.<commit_after>def test_parse_bool(wish):
parse_bool = wish
assert not parse_bool('false')
assert not parse_bool('FALSE')
assert not parse_bool('0')
assert parse_bool('true')
assert parse_bool('TRUE')
assert parse_bool('1')
| |
bc551c0d46c63ed688184e45e3211e29a30214d7 | dbaas/workflow/steps/util/resize/check_database_status.py | dbaas/workflow/steps/util/resize/check_database_status.py | # -*- coding: utf-8 -*-
import logging
from ...util.base import BaseStep
LOG = logging.getLogger(__name__)
class CheckDatabaseStatus(BaseStep):
def __unicode__(self):
return "Checking database status..."
def do(self, workflow_dict):
try:
if not 'database' in workflow_dict:
... | Add check db status step | Add check db status step
| Python | bsd-3-clause | globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service | Add check db status step | # -*- coding: utf-8 -*-
import logging
from ...util.base import BaseStep
LOG = logging.getLogger(__name__)
class CheckDatabaseStatus(BaseStep):
def __unicode__(self):
return "Checking database status..."
def do(self, workflow_dict):
try:
if not 'database' in workflow_dict:
... | <commit_before><commit_msg>Add check db status step<commit_after> | # -*- coding: utf-8 -*-
import logging
from ...util.base import BaseStep
LOG = logging.getLogger(__name__)
class CheckDatabaseStatus(BaseStep):
def __unicode__(self):
return "Checking database status..."
def do(self, workflow_dict):
try:
if not 'database' in workflow_dict:
... | Add check db status step# -*- coding: utf-8 -*-
import logging
from ...util.base import BaseStep
LOG = logging.getLogger(__name__)
class CheckDatabaseStatus(BaseStep):
def __unicode__(self):
return "Checking database status..."
def do(self, workflow_dict):
try:
if not 'database'... | <commit_before><commit_msg>Add check db status step<commit_after># -*- coding: utf-8 -*-
import logging
from ...util.base import BaseStep
LOG = logging.getLogger(__name__)
class CheckDatabaseStatus(BaseStep):
def __unicode__(self):
return "Checking database status..."
def do(self, workflow_dict):
... | |
2399f02a08f06aa120aae351aad7a4402aca8eb4 | climate_data/migrations/0021_auto_20170619_2053.py | climate_data/migrations/0021_auto_20170619_2053.py | # -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-06-19 20:53
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('climate_data', '0020_annotation'),
]
operations = ... | Move data type field from sensor model to station-sensor link model to make sensors more 'global'. | Move data type field from sensor model to station-sensor link model to make sensors more 'global'.
| Python | apache-2.0 | qubs/climate-data-api,qubs/data-centre,qubs/climate-data-api,qubs/data-centre | Move data type field from sensor model to station-sensor link model to make sensors more 'global'. | # -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-06-19 20:53
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('climate_data', '0020_annotation'),
]
operations = ... | <commit_before><commit_msg>Move data type field from sensor model to station-sensor link model to make sensors more 'global'.<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-06-19 20:53
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('climate_data', '0020_annotation'),
]
operations = ... | Move data type field from sensor model to station-sensor link model to make sensors more 'global'.# -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-06-19 20:53
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migra... | <commit_before><commit_msg>Move data type field from sensor model to station-sensor link model to make sensors more 'global'.<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-06-19 20:53
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.de... | |
9ee332f6f0af3d632860581971446f9edf4f74be | changetext/WIKIXML2TW.py | changetext/WIKIXML2TW.py |
def WIKIXML2TW(inputfilename, outputfilename):
"Convert Wikimedia XML dump to TiddlyWiki import file"
inputfile = open(inputfilename, "r")
xmlinput = unicode(inputfile.read(), errors='ignore')
outputfilemenu = open(outputfilename + '.menu', "w")
outputfile = open(outputfilename, "w")
outputfile.w... | Convert wiki XML to Tiddlywiki import format | Convert wiki XML to Tiddlywiki import format
| Python | mit | cottley/moruga | Convert wiki XML to Tiddlywiki import format |
def WIKIXML2TW(inputfilename, outputfilename):
"Convert Wikimedia XML dump to TiddlyWiki import file"
inputfile = open(inputfilename, "r")
xmlinput = unicode(inputfile.read(), errors='ignore')
outputfilemenu = open(outputfilename + '.menu', "w")
outputfile = open(outputfilename, "w")
outputfile.w... | <commit_before><commit_msg>Convert wiki XML to Tiddlywiki import format<commit_after> |
def WIKIXML2TW(inputfilename, outputfilename):
"Convert Wikimedia XML dump to TiddlyWiki import file"
inputfile = open(inputfilename, "r")
xmlinput = unicode(inputfile.read(), errors='ignore')
outputfilemenu = open(outputfilename + '.menu', "w")
outputfile = open(outputfilename, "w")
outputfile.w... | Convert wiki XML to Tiddlywiki import format
def WIKIXML2TW(inputfilename, outputfilename):
"Convert Wikimedia XML dump to TiddlyWiki import file"
inputfile = open(inputfilename, "r")
xmlinput = unicode(inputfile.read(), errors='ignore')
outputfilemenu = open(outputfilename + '.menu', "w")
outputfile... | <commit_before><commit_msg>Convert wiki XML to Tiddlywiki import format<commit_after>
def WIKIXML2TW(inputfilename, outputfilename):
"Convert Wikimedia XML dump to TiddlyWiki import file"
inputfile = open(inputfilename, "r")
xmlinput = unicode(inputfile.read(), errors='ignore')
outputfilemenu = open(out... | |
2f508798b0a5f78ef6c04302e6983e4b824a2df0 | exercises/chapter_02/exercise_02_10/exercise_02_10.py | exercises/chapter_02/exercise_02_10/exercise_02_10.py | # 2-10 Adding Comments
# Below follows the code from exercise 2-3 and 2-4 again but now with (more) comments
# 2-3 Personal Message
name = "Henrik" # variable for storing user name
# concatenate the user name with some strings to form a personal message
print("Hello " + name + ", lets learn some Python today!") #
... | Add solution to exercise 2-10. | Add solution to exercise 2-10.
| Python | mit | HenrikSamuelsson/python-crash-course | Add solution to exercise 2-10. | # 2-10 Adding Comments
# Below follows the code from exercise 2-3 and 2-4 again but now with (more) comments
# 2-3 Personal Message
name = "Henrik" # variable for storing user name
# concatenate the user name with some strings to form a personal message
print("Hello " + name + ", lets learn some Python today!") #
... | <commit_before><commit_msg>Add solution to exercise 2-10.<commit_after> | # 2-10 Adding Comments
# Below follows the code from exercise 2-3 and 2-4 again but now with (more) comments
# 2-3 Personal Message
name = "Henrik" # variable for storing user name
# concatenate the user name with some strings to form a personal message
print("Hello " + name + ", lets learn some Python today!") #
... | Add solution to exercise 2-10.# 2-10 Adding Comments
# Below follows the code from exercise 2-3 and 2-4 again but now with (more) comments
# 2-3 Personal Message
name = "Henrik" # variable for storing user name
# concatenate the user name with some strings to form a personal message
print("Hello " + name + ", lets ... | <commit_before><commit_msg>Add solution to exercise 2-10.<commit_after># 2-10 Adding Comments
# Below follows the code from exercise 2-3 and 2-4 again but now with (more) comments
# 2-3 Personal Message
name = "Henrik" # variable for storing user name
# concatenate the user name with some strings to form a personal... | |
ac8c5a878c8ef05c139ce3d86378b3c9ca4845d3 | tools/bbox_post_process.py | tools/bbox_post_process.py | #!/usr/bin/env python
import init
import caffe
import numpy as np
import argparse
import cPickle
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('net_def')
parser.add_argument('net_param')
parser.add_argument('save_file')
parser.add_argument('--bbox_means', default... | Add a script for process trained models using bbox means and stds. | Add a script for process trained models using bbox means and stds.
| Python | mit | myfavouritekk/TPN | Add a script for process trained models using bbox means and stds. | #!/usr/bin/env python
import init
import caffe
import numpy as np
import argparse
import cPickle
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('net_def')
parser.add_argument('net_param')
parser.add_argument('save_file')
parser.add_argument('--bbox_means', default... | <commit_before><commit_msg>Add a script for process trained models using bbox means and stds.<commit_after> | #!/usr/bin/env python
import init
import caffe
import numpy as np
import argparse
import cPickle
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('net_def')
parser.add_argument('net_param')
parser.add_argument('save_file')
parser.add_argument('--bbox_means', default... | Add a script for process trained models using bbox means and stds.#!/usr/bin/env python
import init
import caffe
import numpy as np
import argparse
import cPickle
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('net_def')
parser.add_argument('net_param')
parser.add_arg... | <commit_before><commit_msg>Add a script for process trained models using bbox means and stds.<commit_after>#!/usr/bin/env python
import init
import caffe
import numpy as np
import argparse
import cPickle
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('net_def')
parser.add... | |
e00936dda07b53c22f66897d7e487d12374172e7 | corehq/apps/cleanup/management/commands/get_doc_domains.py | corehq/apps/cleanup/management/commands/get_doc_domains.py | import os
from optparse import make_option
from django.core.management.base import BaseCommand
from couchdbkit.exceptions import ResourceNotFound
from dimagi.utils.couch.database import get_db
# possible expansion: allow this to accept doc ids on the command line
# for use like `cat doc_ids.txt | ./manage.py get_do... | Add management command to get domains from doc ids | Add management command to get domains from doc ids
| Python | bsd-3-clause | gmimano/commcaretest,qedsoftware/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,gmimano/commcaretest,dimagi/commcare-hq,SEL-Columbia/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,SEL-Columbia/commcare-hq,puttarajubr/commcare-hq,gmimano/... | Add management command to get domains from doc ids | import os
from optparse import make_option
from django.core.management.base import BaseCommand
from couchdbkit.exceptions import ResourceNotFound
from dimagi.utils.couch.database import get_db
# possible expansion: allow this to accept doc ids on the command line
# for use like `cat doc_ids.txt | ./manage.py get_do... | <commit_before><commit_msg>Add management command to get domains from doc ids<commit_after> | import os
from optparse import make_option
from django.core.management.base import BaseCommand
from couchdbkit.exceptions import ResourceNotFound
from dimagi.utils.couch.database import get_db
# possible expansion: allow this to accept doc ids on the command line
# for use like `cat doc_ids.txt | ./manage.py get_do... | Add management command to get domains from doc idsimport os
from optparse import make_option
from django.core.management.base import BaseCommand
from couchdbkit.exceptions import ResourceNotFound
from dimagi.utils.couch.database import get_db
# possible expansion: allow this to accept doc ids on the command line
# ... | <commit_before><commit_msg>Add management command to get domains from doc ids<commit_after>import os
from optparse import make_option
from django.core.management.base import BaseCommand
from couchdbkit.exceptions import ResourceNotFound
from dimagi.utils.couch.database import get_db
# possible expansion: allow this... | |
24f6ebb4218f3fc1858c6100c78c41c9744b55ec | migrations/versions/0186_rename_is_active_columns.py | migrations/versions/0186_rename_is_active_columns.py | """
Revision ID: 0186_rename_is_active_columns
Revises: 0185_add_is_active_to_reply_tos
Create Date: 2018-04-27 16:35:41.824981
"""
from alembic import op
import sqlalchemy as sa
revision = '0186_rename_is_active_columns'
down_revision = '0185_add_is_active_to_reply_tos'
def upgrade():
op.alter_column('servic... | Change 'is_active' column names to 'archived' | Change 'is_active' column names to 'archived'
Changed the name of the `is_active` columns to `archived` in these
tables:
* `service_email_reply_to`
* `service_sms_senders`
* `service_letter_contacts`
These columns were not being used yet, but because we describe Templates as
'archived', this keeps the names consisten... | Python | mit | alphagov/notifications-api,alphagov/notifications-api | Change 'is_active' column names to 'archived'
Changed the name of the `is_active` columns to `archived` in these
tables:
* `service_email_reply_to`
* `service_sms_senders`
* `service_letter_contacts`
These columns were not being used yet, but because we describe Templates as
'archived', this keeps the names consisten... | """
Revision ID: 0186_rename_is_active_columns
Revises: 0185_add_is_active_to_reply_tos
Create Date: 2018-04-27 16:35:41.824981
"""
from alembic import op
import sqlalchemy as sa
revision = '0186_rename_is_active_columns'
down_revision = '0185_add_is_active_to_reply_tos'
def upgrade():
op.alter_column('servic... | <commit_before><commit_msg>Change 'is_active' column names to 'archived'
Changed the name of the `is_active` columns to `archived` in these
tables:
* `service_email_reply_to`
* `service_sms_senders`
* `service_letter_contacts`
These columns were not being used yet, but because we describe Templates as
'archived', thi... | """
Revision ID: 0186_rename_is_active_columns
Revises: 0185_add_is_active_to_reply_tos
Create Date: 2018-04-27 16:35:41.824981
"""
from alembic import op
import sqlalchemy as sa
revision = '0186_rename_is_active_columns'
down_revision = '0185_add_is_active_to_reply_tos'
def upgrade():
op.alter_column('servic... | Change 'is_active' column names to 'archived'
Changed the name of the `is_active` columns to `archived` in these
tables:
* `service_email_reply_to`
* `service_sms_senders`
* `service_letter_contacts`
These columns were not being used yet, but because we describe Templates as
'archived', this keeps the names consisten... | <commit_before><commit_msg>Change 'is_active' column names to 'archived'
Changed the name of the `is_active` columns to `archived` in these
tables:
* `service_email_reply_to`
* `service_sms_senders`
* `service_letter_contacts`
These columns were not being used yet, but because we describe Templates as
'archived', thi... | |
2774139f64f83e5b173a5acb67f986648b52dd6f | error-propagation/error-propagation.py | error-propagation/error-propagation.py | #!/usr/bin/env python3
import numpy
import random
random.seed(54864218)
def f(a, b):
coefficients = numpy.array([[2*a + b, a + b], [a - b, a - 2*b]])
inv_coefficients = numpy.linalg.inv(coefficients)
vars = numpy.array([2.5306, 10.1])
elements = numpy.matmul(inv_coefficients, vars)
return elem... | Add a basic error propagation example. | Add a basic error propagation example.
| Python | mpl-2.0 | DanielBrookRoberge/MonteCarloExamples | Add a basic error propagation example. | #!/usr/bin/env python3
import numpy
import random
random.seed(54864218)
def f(a, b):
coefficients = numpy.array([[2*a + b, a + b], [a - b, a - 2*b]])
inv_coefficients = numpy.linalg.inv(coefficients)
vars = numpy.array([2.5306, 10.1])
elements = numpy.matmul(inv_coefficients, vars)
return elem... | <commit_before><commit_msg>Add a basic error propagation example.<commit_after> | #!/usr/bin/env python3
import numpy
import random
random.seed(54864218)
def f(a, b):
coefficients = numpy.array([[2*a + b, a + b], [a - b, a - 2*b]])
inv_coefficients = numpy.linalg.inv(coefficients)
vars = numpy.array([2.5306, 10.1])
elements = numpy.matmul(inv_coefficients, vars)
return elem... | Add a basic error propagation example.#!/usr/bin/env python3
import numpy
import random
random.seed(54864218)
def f(a, b):
coefficients = numpy.array([[2*a + b, a + b], [a - b, a - 2*b]])
inv_coefficients = numpy.linalg.inv(coefficients)
vars = numpy.array([2.5306, 10.1])
elements = numpy.matmul(in... | <commit_before><commit_msg>Add a basic error propagation example.<commit_after>#!/usr/bin/env python3
import numpy
import random
random.seed(54864218)
def f(a, b):
coefficients = numpy.array([[2*a + b, a + b], [a - b, a - 2*b]])
inv_coefficients = numpy.linalg.inv(coefficients)
vars = numpy.array([2.530... | |
22c668a74682769fd6c41709630047e07f6d8915 | example/NaCl/NaCl-gv.py | example/NaCl/NaCl-gv.py | import numpy as np
from phonopy import Phonopy
from phonopy.interface.vasp import read_vasp
from phonopy.file_IO import parse_FORCE_SETS, parse_BORN
import matplotlib.pyplot as plt
unitcell = read_vasp("POSCAR")
phonon = Phonopy(unitcell,
[[2, 0, 0],
[0, 2, 0],
[0, ... | Add an API example for NaCl | Add an API example for NaCl
| Python | bsd-3-clause | atztogo/phonopy,atztogo/phonopy,atztogo/phonopy,atztogo/phonopy | Add an API example for NaCl | import numpy as np
from phonopy import Phonopy
from phonopy.interface.vasp import read_vasp
from phonopy.file_IO import parse_FORCE_SETS, parse_BORN
import matplotlib.pyplot as plt
unitcell = read_vasp("POSCAR")
phonon = Phonopy(unitcell,
[[2, 0, 0],
[0, 2, 0],
[0, ... | <commit_before><commit_msg>Add an API example for NaCl<commit_after> | import numpy as np
from phonopy import Phonopy
from phonopy.interface.vasp import read_vasp
from phonopy.file_IO import parse_FORCE_SETS, parse_BORN
import matplotlib.pyplot as plt
unitcell = read_vasp("POSCAR")
phonon = Phonopy(unitcell,
[[2, 0, 0],
[0, 2, 0],
[0, ... | Add an API example for NaClimport numpy as np
from phonopy import Phonopy
from phonopy.interface.vasp import read_vasp
from phonopy.file_IO import parse_FORCE_SETS, parse_BORN
import matplotlib.pyplot as plt
unitcell = read_vasp("POSCAR")
phonon = Phonopy(unitcell,
[[2, 0, 0],
[0, 2,... | <commit_before><commit_msg>Add an API example for NaCl<commit_after>import numpy as np
from phonopy import Phonopy
from phonopy.interface.vasp import read_vasp
from phonopy.file_IO import parse_FORCE_SETS, parse_BORN
import matplotlib.pyplot as plt
unitcell = read_vasp("POSCAR")
phonon = Phonopy(unitcell,
... | |
1c951a5aa3ff604177087e847faaeca1570bd898 | experiments/keras_tensorflow/kuza55/original_examples/kuza55_blog_example.py | experiments/keras_tensorflow/kuza55/original_examples/kuza55_blog_example.py | # Multi-GPU Keras training.
# Captured from https://github.com/kuza55/keras-extras and
# https://medium.com/@kuza55/transparent-multi-gpu-training-on-tensorflow-with-keras-8b0016fd9012.
from keras.layers import merge
from keras.layers.core import Lambda
from keras.models import Model
import tensorflow as tf
def make... | Add an example from kuza55. | Add an example from kuza55.
| Python | mit | rossumai/keras-multi-gpu,rossumai/keras-multi-gpu | Add an example from kuza55. | # Multi-GPU Keras training.
# Captured from https://github.com/kuza55/keras-extras and
# https://medium.com/@kuza55/transparent-multi-gpu-training-on-tensorflow-with-keras-8b0016fd9012.
from keras.layers import merge
from keras.layers.core import Lambda
from keras.models import Model
import tensorflow as tf
def make... | <commit_before><commit_msg>Add an example from kuza55.<commit_after> | # Multi-GPU Keras training.
# Captured from https://github.com/kuza55/keras-extras and
# https://medium.com/@kuza55/transparent-multi-gpu-training-on-tensorflow-with-keras-8b0016fd9012.
from keras.layers import merge
from keras.layers.core import Lambda
from keras.models import Model
import tensorflow as tf
def make... | Add an example from kuza55.# Multi-GPU Keras training.
# Captured from https://github.com/kuza55/keras-extras and
# https://medium.com/@kuza55/transparent-multi-gpu-training-on-tensorflow-with-keras-8b0016fd9012.
from keras.layers import merge
from keras.layers.core import Lambda
from keras.models import Model
import... | <commit_before><commit_msg>Add an example from kuza55.<commit_after># Multi-GPU Keras training.
# Captured from https://github.com/kuza55/keras-extras and
# https://medium.com/@kuza55/transparent-multi-gpu-training-on-tensorflow-with-keras-8b0016fd9012.
from keras.layers import merge
from keras.layers.core import Lamb... | |
e6cf9cb9d27523fd72242f6ea137d14bff5f2039 | interface/plugin/farmanager/02title/__init__.py | interface/plugin/farmanager/02title/__init__.py | """
Gets plugin info from global fields
Low-level Far Manager API is here:
* https://api.farmanager.com/en/exported_functions/getglobalinfow.html
"""
__title__ = "02fields"
__author__ = "anatoly techtonik <[email protected]>"
__license__ = "Public Domain"
# --- utility functions ---
import hashlib
def getu... | Add 02title/ plugin that gets own info from global fields | Add 02title/ plugin that gets own info from global fields
| Python | unlicense | techtonik/discovery,techtonik/discovery,techtonik/discovery | Add 02title/ plugin that gets own info from global fields | """
Gets plugin info from global fields
Low-level Far Manager API is here:
* https://api.farmanager.com/en/exported_functions/getglobalinfow.html
"""
__title__ = "02fields"
__author__ = "anatoly techtonik <[email protected]>"
__license__ = "Public Domain"
# --- utility functions ---
import hashlib
def getu... | <commit_before><commit_msg>Add 02title/ plugin that gets own info from global fields<commit_after> | """
Gets plugin info from global fields
Low-level Far Manager API is here:
* https://api.farmanager.com/en/exported_functions/getglobalinfow.html
"""
__title__ = "02fields"
__author__ = "anatoly techtonik <[email protected]>"
__license__ = "Public Domain"
# --- utility functions ---
import hashlib
def getu... | Add 02title/ plugin that gets own info from global fields"""
Gets plugin info from global fields
Low-level Far Manager API is here:
* https://api.farmanager.com/en/exported_functions/getglobalinfow.html
"""
__title__ = "02fields"
__author__ = "anatoly techtonik <[email protected]>"
__license__ = "Public Domain... | <commit_before><commit_msg>Add 02title/ plugin that gets own info from global fields<commit_after>"""
Gets plugin info from global fields
Low-level Far Manager API is here:
* https://api.farmanager.com/en/exported_functions/getglobalinfow.html
"""
__title__ = "02fields"
__author__ = "anatoly techtonik <techtonik... | |
2bfe7b5324ef79cc60b81ffc392bd2dd1b7b2bc5 | src/tests/templateedit.py | src/tests/templateedit.py | # -*- encoding: utf-8 -*-
from __future__ import unicode_literals
import unittest
from oabot.main import *
class TemplateEditTests(unittest.TestCase):
def propose_change(self, text, page_name='Test page'):
wikicode = mwparserfromhell.parse(text)
for template in wikicode.filter_templates():
... | Add test utilities for proposed changes - feel free to add other cases | Add test utilities for proposed changes - feel free to add other cases
| Python | mit | dissemin/oabot,dissemin/oabot,dissemin/oabot | Add test utilities for proposed changes - feel free to add other cases | # -*- encoding: utf-8 -*-
from __future__ import unicode_literals
import unittest
from oabot.main import *
class TemplateEditTests(unittest.TestCase):
def propose_change(self, text, page_name='Test page'):
wikicode = mwparserfromhell.parse(text)
for template in wikicode.filter_templates():
... | <commit_before><commit_msg>Add test utilities for proposed changes - feel free to add other cases<commit_after> | # -*- encoding: utf-8 -*-
from __future__ import unicode_literals
import unittest
from oabot.main import *
class TemplateEditTests(unittest.TestCase):
def propose_change(self, text, page_name='Test page'):
wikicode = mwparserfromhell.parse(text)
for template in wikicode.filter_templates():
... | Add test utilities for proposed changes - feel free to add other cases# -*- encoding: utf-8 -*-
from __future__ import unicode_literals
import unittest
from oabot.main import *
class TemplateEditTests(unittest.TestCase):
def propose_change(self, text, page_name='Test page'):
wikicode = mwparserfromhell.... | <commit_before><commit_msg>Add test utilities for proposed changes - feel free to add other cases<commit_after># -*- encoding: utf-8 -*-
from __future__ import unicode_literals
import unittest
from oabot.main import *
class TemplateEditTests(unittest.TestCase):
def propose_change(self, text, page_name='Test pag... | |
52accd6f82893fe167a78eabe8cdf30dfa0bc841 | glance/tests/unit/test_data_migration_version.py | glance/tests/unit/test_data_migration_version.py | # Copyright 2019 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by... | Add test for data migration version | Add test for data migration version
Detect when the CURRENT_VERSION constant used by the database
migration code is out of sync with the version number of the
current release.
Change-Id: I9e8b5ee98c6c111aaf065cb310906d6442f5f79a
| Python | apache-2.0 | openstack/glance,openstack/glance,openstack/glance | Add test for data migration version
Detect when the CURRENT_VERSION constant used by the database
migration code is out of sync with the version number of the
current release.
Change-Id: I9e8b5ee98c6c111aaf065cb310906d6442f5f79a | # Copyright 2019 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by... | <commit_before><commit_msg>Add test for data migration version
Detect when the CURRENT_VERSION constant used by the database
migration code is out of sync with the version number of the
current release.
Change-Id: I9e8b5ee98c6c111aaf065cb310906d6442f5f79a<commit_after> | # Copyright 2019 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by... | Add test for data migration version
Detect when the CURRENT_VERSION constant used by the database
migration code is out of sync with the version number of the
current release.
Change-Id: I9e8b5ee98c6c111aaf065cb310906d6442f5f79a# Copyright 2019 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache Lic... | <commit_before><commit_msg>Add test for data migration version
Detect when the CURRENT_VERSION constant used by the database
migration code is out of sync with the version number of the
current release.
Change-Id: I9e8b5ee98c6c111aaf065cb310906d6442f5f79a<commit_after># Copyright 2019 Red Hat, Inc.
# All Rights Reser... | |
11d39551f85a1490ebe370b97ed729d85df06b0b | shuup/xtheme/migrations/0004_convert_shop_themes.py | shuup/xtheme/migrations/0004_convert_shop_themes.py | # -*- coding: utf-8 -*-
# Generated by Django 1.9.11 on 2016-12-07 23:22
from __future__ import unicode_literals
from django.db.transaction import atomic
from django.db import migrations
from shuup.core.models import Shop
from shuup.xtheme.models import SavedViewConfig, ThemeSettings
@atomic
def convert_shop_themes... | Add migration to add shop information | Xtheme: Add migration to add shop information
Add shop information in ThemeSettings and SavedViewConfig
The process will clone every settings for each existent shop
| Python | agpl-3.0 | shoopio/shoop,shoopio/shoop,shoopio/shoop | Xtheme: Add migration to add shop information
Add shop information in ThemeSettings and SavedViewConfig
The process will clone every settings for each existent shop | # -*- coding: utf-8 -*-
# Generated by Django 1.9.11 on 2016-12-07 23:22
from __future__ import unicode_literals
from django.db.transaction import atomic
from django.db import migrations
from shuup.core.models import Shop
from shuup.xtheme.models import SavedViewConfig, ThemeSettings
@atomic
def convert_shop_themes... | <commit_before><commit_msg>Xtheme: Add migration to add shop information
Add shop information in ThemeSettings and SavedViewConfig
The process will clone every settings for each existent shop<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.9.11 on 2016-12-07 23:22
from __future__ import unicode_literals
from django.db.transaction import atomic
from django.db import migrations
from shuup.core.models import Shop
from shuup.xtheme.models import SavedViewConfig, ThemeSettings
@atomic
def convert_shop_themes... | Xtheme: Add migration to add shop information
Add shop information in ThemeSettings and SavedViewConfig
The process will clone every settings for each existent shop# -*- coding: utf-8 -*-
# Generated by Django 1.9.11 on 2016-12-07 23:22
from __future__ import unicode_literals
from django.db.transaction import atomic... | <commit_before><commit_msg>Xtheme: Add migration to add shop information
Add shop information in ThemeSettings and SavedViewConfig
The process will clone every settings for each existent shop<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.9.11 on 2016-12-07 23:22
from __future__ import unicode_literals
... | |
b055b5094eaca3191d066749ea6aff16dd6b9867 | nova/tests/unit/policies/test_console_auth_tokens.py | nova/tests/unit/policies/test_console_auth_tokens.py | # Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# d... | Add test coverage of existing os-console-auth-tokens policies | Add test coverage of existing os-console-auth-tokens policies
Current tests do not have good test coverage of existing policies.
Either tests for policies do not exist or if they exist then they
do not cover the actual negative and positive testing.
For Example, if any policy with default rule as admin only then
test... | Python | apache-2.0 | mahak/nova,openstack/nova,mahak/nova,openstack/nova,klmitch/nova,openstack/nova,klmitch/nova,klmitch/nova,mahak/nova,klmitch/nova | Add test coverage of existing os-console-auth-tokens policies
Current tests do not have good test coverage of existing policies.
Either tests for policies do not exist or if they exist then they
do not cover the actual negative and positive testing.
For Example, if any policy with default rule as admin only then
test... | # Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# d... | <commit_before><commit_msg>Add test coverage of existing os-console-auth-tokens policies
Current tests do not have good test coverage of existing policies.
Either tests for policies do not exist or if they exist then they
do not cover the actual negative and positive testing.
For Example, if any policy with default r... | # Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# d... | Add test coverage of existing os-console-auth-tokens policies
Current tests do not have good test coverage of existing policies.
Either tests for policies do not exist or if they exist then they
do not cover the actual negative and positive testing.
For Example, if any policy with default rule as admin only then
test... | <commit_before><commit_msg>Add test coverage of existing os-console-auth-tokens policies
Current tests do not have good test coverage of existing policies.
Either tests for policies do not exist or if they exist then they
do not cover the actual negative and positive testing.
For Example, if any policy with default r... | |
babbc80437df31b61973576065296b4eb749e243 | src/analyses/migrations/0002_auto_20160616_1724.py | src/analyses/migrations/0002_auto_20160616_1724.py | # -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-06-16 17:24
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('analyses', '0001_initial'),
]
operations = [
... | Update genome reference model migration | Update genome reference model migration
| Python | mit | ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai | Update genome reference model migration | # -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-06-16 17:24
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('analyses', '0001_initial'),
]
operations = [
... | <commit_before><commit_msg>Update genome reference model migration<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-06-16 17:24
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('analyses', '0001_initial'),
]
operations = [
... | Update genome reference model migration# -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-06-16 17:24
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('analyses', '0001_init... | <commit_before><commit_msg>Update genome reference model migration<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.9.7 on 2016-06-16 17:24
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
depende... | |
39327cb18d3551618e84aaa2f70f5e3baaf734de | tests/terminal_tests/CorrectGrammarHandlingTest.py | tests/terminal_tests/CorrectGrammarHandlingTest.py | #!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase
from grammpy import Terminal
class TempClass:
pass
class CorrectGrammarHandlingTest(TestCase):
def test_sameNumber(self):
ter1 = Terminal(0, 1)
ter2 =... | Add test for correct handling of grammar in terminal | Add test for correct handling of grammar in terminal
| Python | mit | PatrikValkovic/grammpy | Add test for correct handling of grammar in terminal | #!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase
from grammpy import Terminal
class TempClass:
pass
class CorrectGrammarHandlingTest(TestCase):
def test_sameNumber(self):
ter1 = Terminal(0, 1)
ter2 =... | <commit_before><commit_msg>Add test for correct handling of grammar in terminal<commit_after> | #!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase
from grammpy import Terminal
class TempClass:
pass
class CorrectGrammarHandlingTest(TestCase):
def test_sameNumber(self):
ter1 = Terminal(0, 1)
ter2 =... | Add test for correct handling of grammar in terminal#!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase
from grammpy import Terminal
class TempClass:
pass
class CorrectGrammarHandlingTest(TestCase):
def test_sameNumber... | <commit_before><commit_msg>Add test for correct handling of grammar in terminal<commit_after>#!/usr/bin/env python
"""
:Author Patrik Valkovic
:Created 23.06.2017 16:39
:Licence GNUv3
Part of grammpy
"""
from unittest import TestCase
from grammpy import Terminal
class TempClass:
pass
class CorrectGrammarHandli... | |
42851304fecec95c32eebd618aae192743aab5d4 | src/ggrc_risk_assessment_v2/migrations/versions/20141121231826_1347acbb4dc2_add_unique_constraint_to_threat_actors.py | src/ggrc_risk_assessment_v2/migrations/versions/20141121231826_1347acbb4dc2_add_unique_constraint_to_threat_actors.py |
"""Add unique constraint to threat actors
Revision ID: 1347acbb4dc2
Revises: 5ada65dc60b3
Create Date: 2014-11-21 23:18:26.689048
"""
# revision identifiers, used by Alembic.
revision = '1347acbb4dc2'
down_revision = '5ada65dc60b3'
from alembic import op
def upgrade():
op.create_unique_constraint('uq_t_actors'... | Add unique constraint on Threat Actor title | Add unique constraint on Threat Actor title
| Python | apache-2.0 | prasannav7/ggrc-core,kr41/ggrc-core,andrei-karalionak/ggrc-core,NejcZupec/ggrc-core,plamut/ggrc-core,VinnieJohns/ggrc-core,j0gurt/ggrc-core,prasannav7/ggrc-core,selahssea/ggrc-core,hyperNURb/ggrc-core,VinnieJohns/ggrc-core,VinnieJohns/ggrc-core,plamut/ggrc-core,kr41/ggrc-core,AleksNeStu/ggrc-core,hasanalom/ggrc-core,j0... | Add unique constraint on Threat Actor title |
"""Add unique constraint to threat actors
Revision ID: 1347acbb4dc2
Revises: 5ada65dc60b3
Create Date: 2014-11-21 23:18:26.689048
"""
# revision identifiers, used by Alembic.
revision = '1347acbb4dc2'
down_revision = '5ada65dc60b3'
from alembic import op
def upgrade():
op.create_unique_constraint('uq_t_actors'... | <commit_before><commit_msg>Add unique constraint on Threat Actor title<commit_after> |
"""Add unique constraint to threat actors
Revision ID: 1347acbb4dc2
Revises: 5ada65dc60b3
Create Date: 2014-11-21 23:18:26.689048
"""
# revision identifiers, used by Alembic.
revision = '1347acbb4dc2'
down_revision = '5ada65dc60b3'
from alembic import op
def upgrade():
op.create_unique_constraint('uq_t_actors'... | Add unique constraint on Threat Actor title
"""Add unique constraint to threat actors
Revision ID: 1347acbb4dc2
Revises: 5ada65dc60b3
Create Date: 2014-11-21 23:18:26.689048
"""
# revision identifiers, used by Alembic.
revision = '1347acbb4dc2'
down_revision = '5ada65dc60b3'
from alembic import op
def upgrade():
... | <commit_before><commit_msg>Add unique constraint on Threat Actor title<commit_after>
"""Add unique constraint to threat actors
Revision ID: 1347acbb4dc2
Revises: 5ada65dc60b3
Create Date: 2014-11-21 23:18:26.689048
"""
# revision identifiers, used by Alembic.
revision = '1347acbb4dc2'
down_revision = '5ada65dc60b3'
... | |
66e6ba2988aee0cf9301f735e170b7f74ca310ac | qcfractal/alembic/versions/4bb79efa9855_add_queue_manager_id_to_base_results.py | qcfractal/alembic/versions/4bb79efa9855_add_queue_manager_id_to_base_results.py | """Add queue_manager id to Base_results
Revision ID: 4bb79efa9855
Revises: 26cfd7b0439e
Create Date: 2019-07-26 15:31:14.811337
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '4bb79efa9855'
down_revision = '26cfd7b0439e'
branch_labels = None
depends_on = None
... | Add db migration to add manager_id to base_results | Add db migration to add manager_id to base_results
| Python | bsd-3-clause | psi4/mongo_qcdb,psi4/DatenQM,psi4/mongo_qcdb,psi4/DatenQM | Add db migration to add manager_id to base_results | """Add queue_manager id to Base_results
Revision ID: 4bb79efa9855
Revises: 26cfd7b0439e
Create Date: 2019-07-26 15:31:14.811337
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '4bb79efa9855'
down_revision = '26cfd7b0439e'
branch_labels = None
depends_on = None
... | <commit_before><commit_msg>Add db migration to add manager_id to base_results<commit_after> | """Add queue_manager id to Base_results
Revision ID: 4bb79efa9855
Revises: 26cfd7b0439e
Create Date: 2019-07-26 15:31:14.811337
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '4bb79efa9855'
down_revision = '26cfd7b0439e'
branch_labels = None
depends_on = None
... | Add db migration to add manager_id to base_results"""Add queue_manager id to Base_results
Revision ID: 4bb79efa9855
Revises: 26cfd7b0439e
Create Date: 2019-07-26 15:31:14.811337
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = '4bb79efa9855'
down_revision = '26c... | <commit_before><commit_msg>Add db migration to add manager_id to base_results<commit_after>"""Add queue_manager id to Base_results
Revision ID: 4bb79efa9855
Revises: 26cfd7b0439e
Create Date: 2019-07-26 15:31:14.811337
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revis... | |
807bb8425665e1cdd9040dfdd830603b95e9ab0a | tests/test_inheritance.py | tests/test_inheritance.py | from dtest import *
from dtest.util import *
# Define setUpClass/tearDownClass/setUp/tearDown for inheritance
class TestInheritanceBase(DTestCase):
class_setup = None
instance_setup = None
@classmethod
def setUpClass(cls):
assert_is_none(cls.class_setup)
cls.class_setup = True
@c... | Test that inheritance of setUp/tearDown/setUpClass/tearDownClass works as expected | Test that inheritance of setUp/tearDown/setUpClass/tearDownClass works
as expected
| Python | apache-2.0 | klmitch/dtest,klmitch/dtest | Test that inheritance of setUp/tearDown/setUpClass/tearDownClass works
as expected | from dtest import *
from dtest.util import *
# Define setUpClass/tearDownClass/setUp/tearDown for inheritance
class TestInheritanceBase(DTestCase):
class_setup = None
instance_setup = None
@classmethod
def setUpClass(cls):
assert_is_none(cls.class_setup)
cls.class_setup = True
@c... | <commit_before><commit_msg>Test that inheritance of setUp/tearDown/setUpClass/tearDownClass works
as expected<commit_after> | from dtest import *
from dtest.util import *
# Define setUpClass/tearDownClass/setUp/tearDown for inheritance
class TestInheritanceBase(DTestCase):
class_setup = None
instance_setup = None
@classmethod
def setUpClass(cls):
assert_is_none(cls.class_setup)
cls.class_setup = True
@c... | Test that inheritance of setUp/tearDown/setUpClass/tearDownClass works
as expectedfrom dtest import *
from dtest.util import *
# Define setUpClass/tearDownClass/setUp/tearDown for inheritance
class TestInheritanceBase(DTestCase):
class_setup = None
instance_setup = None
@classmethod
def setUpClass(cl... | <commit_before><commit_msg>Test that inheritance of setUp/tearDown/setUpClass/tearDownClass works
as expected<commit_after>from dtest import *
from dtest.util import *
# Define setUpClass/tearDownClass/setUp/tearDown for inheritance
class TestInheritanceBase(DTestCase):
class_setup = None
instance_setup = Non... | |
eda80dd9a903a7baaddad123978981352de6d337 | project/app/migrations/0003_auto_20170311_0837.py | project/app/migrations/0003_auto_20170311_0837.py | # -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-03-11 16:37
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('app', '0002_auto_20170311_0811'),
]
operations = [
migrations.AlterUniqueTogether(
... | Remove constraint on Session model | Remove constraint on Session model
| Python | bsd-2-clause | barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,barberscore/barberscore-api,dbinetti/barberscore,dbinetti/barberscore,dbinetti/barberscore-django,barberscore/barberscore-api | Remove constraint on Session model | # -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-03-11 16:37
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('app', '0002_auto_20170311_0811'),
]
operations = [
migrations.AlterUniqueTogether(
... | <commit_before><commit_msg>Remove constraint on Session model<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-03-11 16:37
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('app', '0002_auto_20170311_0811'),
]
operations = [
migrations.AlterUniqueTogether(
... | Remove constraint on Session model# -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-03-11 16:37
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('app', '0002_auto_20170311_0811'),
]
operations = [
... | <commit_before><commit_msg>Remove constraint on Session model<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.10.6 on 2017-03-11 16:37
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('app', '0002_auto_20170311... | |
03f9355de0c25f41fa5ed78b81b0fd0ee988a117 | py/kth-largest-element-in-an-array.py | py/kth-largest-element-in-an-array.py | import random
class Solution(object):
def findKthLargest(self, nums, k, start=0, end=None):
"""
:type nums: List[int]
:type k: int
:rtype: int
"""
l = len(nums)
if end is None:
end = l
if end == start + 1:
return nums[start]
... | Add py solution for 215. Kth Largest Element in an Array | Add py solution for 215. Kth Largest Element in an Array
215. Kth Largest Element in an Array: https://leetcode.com/problems/kth-largest-element-in-an-array/
| Python | apache-2.0 | ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode | Add py solution for 215. Kth Largest Element in an Array
215. Kth Largest Element in an Array: https://leetcode.com/problems/kth-largest-element-in-an-array/ | import random
class Solution(object):
def findKthLargest(self, nums, k, start=0, end=None):
"""
:type nums: List[int]
:type k: int
:rtype: int
"""
l = len(nums)
if end is None:
end = l
if end == start + 1:
return nums[start]
... | <commit_before><commit_msg>Add py solution for 215. Kth Largest Element in an Array
215. Kth Largest Element in an Array: https://leetcode.com/problems/kth-largest-element-in-an-array/<commit_after> | import random
class Solution(object):
def findKthLargest(self, nums, k, start=0, end=None):
"""
:type nums: List[int]
:type k: int
:rtype: int
"""
l = len(nums)
if end is None:
end = l
if end == start + 1:
return nums[start]
... | Add py solution for 215. Kth Largest Element in an Array
215. Kth Largest Element in an Array: https://leetcode.com/problems/kth-largest-element-in-an-array/import random
class Solution(object):
def findKthLargest(self, nums, k, start=0, end=None):
"""
:type nums: List[int]
:type k: int
... | <commit_before><commit_msg>Add py solution for 215. Kth Largest Element in an Array
215. Kth Largest Element in an Array: https://leetcode.com/problems/kth-largest-element-in-an-array/<commit_after>import random
class Solution(object):
def findKthLargest(self, nums, k, start=0, end=None):
"""
:type... | |
28208126b532edb64c97c63a8cbff841c8da3c4f | scripts/officediff/xlsx-dump.py | scripts/officediff/xlsx-dump.py | from io import StringIO
import sys
import pandas as pd
for sheet_name in pd.ExcelFile(sys.argv[1]).sheet_names:
output = StringIO()
print('Sheet: %s' % sheet_name)
pd.read_excel(sys.argv[1], sheet_name=sheet_name)\
.to_csv(output,
header=True,
index=False)
print(outp... | Add Python script for powerpoint file diffs | Add Python script for powerpoint file diffs
| Python | mit | Stratus3D/dotfiles,Stratus3D/dotfiles,Stratus3D/dotfiles | Add Python script for powerpoint file diffs | from io import StringIO
import sys
import pandas as pd
for sheet_name in pd.ExcelFile(sys.argv[1]).sheet_names:
output = StringIO()
print('Sheet: %s' % sheet_name)
pd.read_excel(sys.argv[1], sheet_name=sheet_name)\
.to_csv(output,
header=True,
index=False)
print(outp... | <commit_before><commit_msg>Add Python script for powerpoint file diffs<commit_after> | from io import StringIO
import sys
import pandas as pd
for sheet_name in pd.ExcelFile(sys.argv[1]).sheet_names:
output = StringIO()
print('Sheet: %s' % sheet_name)
pd.read_excel(sys.argv[1], sheet_name=sheet_name)\
.to_csv(output,
header=True,
index=False)
print(outp... | Add Python script for powerpoint file diffsfrom io import StringIO
import sys
import pandas as pd
for sheet_name in pd.ExcelFile(sys.argv[1]).sheet_names:
output = StringIO()
print('Sheet: %s' % sheet_name)
pd.read_excel(sys.argv[1], sheet_name=sheet_name)\
.to_csv(output,
header=True... | <commit_before><commit_msg>Add Python script for powerpoint file diffs<commit_after>from io import StringIO
import sys
import pandas as pd
for sheet_name in pd.ExcelFile(sys.argv[1]).sheet_names:
output = StringIO()
print('Sheet: %s' % sheet_name)
pd.read_excel(sys.argv[1], sheet_name=sheet_name)\
... | |
8ea2670402812738e41400e5b513bd902757635e | ideascube/conf/idb_sen_fap.py | ideascube/conf/idb_sen_fap.py | # -*- coding: utf-8 -*-
"""Ideaxbox for FAP in Sénégal"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Futur au Présent"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['SN']
TIME_ZONE = None
LANGUAGE_CODE = 'fr'
LOAN_DURATION = 14
MONITORING_ENTRY_EXPORT_FI... | Add conf file for Ideasbox in Sénégal | Add conf file for Ideasbox in Sénégal
| Python | agpl-3.0 | ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube | Add conf file for Ideasbox in Sénégal | # -*- coding: utf-8 -*-
"""Ideaxbox for FAP in Sénégal"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Futur au Présent"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['SN']
TIME_ZONE = None
LANGUAGE_CODE = 'fr'
LOAN_DURATION = 14
MONITORING_ENTRY_EXPORT_FI... | <commit_before><commit_msg>Add conf file for Ideasbox in Sénégal<commit_after> | # -*- coding: utf-8 -*-
"""Ideaxbox for FAP in Sénégal"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Futur au Présent"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['SN']
TIME_ZONE = None
LANGUAGE_CODE = 'fr'
LOAN_DURATION = 14
MONITORING_ENTRY_EXPORT_FI... | Add conf file for Ideasbox in Sénégal# -*- coding: utf-8 -*-
"""Ideaxbox for FAP in Sénégal"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Futur au Présent"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['SN']
TIME_ZONE = None
LANGUAGE_CODE = 'fr'
LOAN_DUR... | <commit_before><commit_msg>Add conf file for Ideasbox in Sénégal<commit_after># -*- coding: utf-8 -*-
"""Ideaxbox for FAP in Sénégal"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Futur au Présent"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['SN']
TIME_... | |
0f16c249f048e27a67f14ff834c690fc2434b55f | admin/common_auth/migrations/0006_auto_20170130_1611.py | admin/common_auth/migrations/0006_auto_20170130_1611.py | # -*- coding: utf-8 -*-
# Generated by Django 1.9 on 2017-01-30 22:11
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('common_auth', '0005_auto_20170111_1513'),
]
operations = [
migrations.AlterModelOption... | Add migration that adds permission for spam, metrics and prereg | Add migration that adds permission for spam, metrics and prereg
| Python | apache-2.0 | CenterForOpenScience/osf.io,monikagrabowska/osf.io,Nesiehr/osf.io,adlius/osf.io,caseyrollins/osf.io,brianjgeiger/osf.io,Nesiehr/osf.io,acshi/osf.io,laurenrevere/osf.io,pattisdr/osf.io,binoculars/osf.io,CenterForOpenScience/osf.io,mfraezz/osf.io,felliott/osf.io,mattclark/osf.io,cslzchen/osf.io,brianjgeiger/osf.io,caseyr... | Add migration that adds permission for spam, metrics and prereg | # -*- coding: utf-8 -*-
# Generated by Django 1.9 on 2017-01-30 22:11
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('common_auth', '0005_auto_20170111_1513'),
]
operations = [
migrations.AlterModelOption... | <commit_before><commit_msg>Add migration that adds permission for spam, metrics and prereg<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.9 on 2017-01-30 22:11
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('common_auth', '0005_auto_20170111_1513'),
]
operations = [
migrations.AlterModelOption... | Add migration that adds permission for spam, metrics and prereg# -*- coding: utf-8 -*-
# Generated by Django 1.9 on 2017-01-30 22:11
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('common_auth', '0005_auto_20170111_1513')... | <commit_before><commit_msg>Add migration that adds permission for spam, metrics and prereg<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.9 on 2017-01-30 22:11
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
(... | |
7e04bc41e977ef7304972cfc630cd9bf9d2c0aa2 | examples/anonymized-real-case.py | examples/anonymized-real-case.py | #!/usr/bin/env python2
# coding: utf-8
# this example is used in production, it is depending on private libraries
# to communicate with internal APIs, but it can help you build your own
# production company-specific hook.
import sys
import json
import pprint
pp = pprint.PrettyPrinter(indent=4)
sys.path.insert(0, '/o... | Add anonymzed real case example | Add anonymzed real case example
| Python | mit | moul/ssh2docker,moul/ssh2docker | Add anonymzed real case example | #!/usr/bin/env python2
# coding: utf-8
# this example is used in production, it is depending on private libraries
# to communicate with internal APIs, but it can help you build your own
# production company-specific hook.
import sys
import json
import pprint
pp = pprint.PrettyPrinter(indent=4)
sys.path.insert(0, '/o... | <commit_before><commit_msg>Add anonymzed real case example<commit_after> | #!/usr/bin/env python2
# coding: utf-8
# this example is used in production, it is depending on private libraries
# to communicate with internal APIs, but it can help you build your own
# production company-specific hook.
import sys
import json
import pprint
pp = pprint.PrettyPrinter(indent=4)
sys.path.insert(0, '/o... | Add anonymzed real case example#!/usr/bin/env python2
# coding: utf-8
# this example is used in production, it is depending on private libraries
# to communicate with internal APIs, but it can help you build your own
# production company-specific hook.
import sys
import json
import pprint
pp = pprint.PrettyPrinter(in... | <commit_before><commit_msg>Add anonymzed real case example<commit_after>#!/usr/bin/env python2
# coding: utf-8
# this example is used in production, it is depending on private libraries
# to communicate with internal APIs, but it can help you build your own
# production company-specific hook.
import sys
import json
i... | |
cee56f48a8ad726498c75b38cf39e4c83ceeb359 | analyze.py | analyze.py | import os
import pickle
import numpy as np
import matplotlib.pyplot as plt
from datetime import datetime
def load_data(data_path):
'''Return dictionary `data` from string `data_path`
'''
os.path.join(data_path, '1.dat')
data = pickle.load(open(data_path, 'rb'))
return data
def get_baseline(data... | Add helper functions for analyzing data | Add helper functions for analyzing data
| Python | mit | JustinShenk/sensei | Add helper functions for analyzing data | import os
import pickle
import numpy as np
import matplotlib.pyplot as plt
from datetime import datetime
def load_data(data_path):
'''Return dictionary `data` from string `data_path`
'''
os.path.join(data_path, '1.dat')
data = pickle.load(open(data_path, 'rb'))
return data
def get_baseline(data... | <commit_before><commit_msg>Add helper functions for analyzing data<commit_after> | import os
import pickle
import numpy as np
import matplotlib.pyplot as plt
from datetime import datetime
def load_data(data_path):
'''Return dictionary `data` from string `data_path`
'''
os.path.join(data_path, '1.dat')
data = pickle.load(open(data_path, 'rb'))
return data
def get_baseline(data... | Add helper functions for analyzing dataimport os
import pickle
import numpy as np
import matplotlib.pyplot as plt
from datetime import datetime
def load_data(data_path):
'''Return dictionary `data` from string `data_path`
'''
os.path.join(data_path, '1.dat')
data = pickle.load(open(data_path, 'rb'))
... | <commit_before><commit_msg>Add helper functions for analyzing data<commit_after>import os
import pickle
import numpy as np
import matplotlib.pyplot as plt
from datetime import datetime
def load_data(data_path):
'''Return dictionary `data` from string `data_path`
'''
os.path.join(data_path, '1.dat')
d... | |
5811966fa895669aa13395a8247e966f4b957ab2 | CodeFights/isDigit.py | CodeFights/isDigit.py | #!/usr/local/bin/python
# Code Fights Is Digit Problem
import re
def isDigit(symbol):
return bool(re.search(r'\d', symbol))
def main():
tests = [
["0", True],
["-", False],
["o", False],
["1", True],
["2", True],
["!", False],
["@", False],
["... | Solve Code Fights is digit problem | Solve Code Fights is digit problem
| Python | mit | HKuz/Test_Code | Solve Code Fights is digit problem | #!/usr/local/bin/python
# Code Fights Is Digit Problem
import re
def isDigit(symbol):
return bool(re.search(r'\d', symbol))
def main():
tests = [
["0", True],
["-", False],
["o", False],
["1", True],
["2", True],
["!", False],
["@", False],
["... | <commit_before><commit_msg>Solve Code Fights is digit problem<commit_after> | #!/usr/local/bin/python
# Code Fights Is Digit Problem
import re
def isDigit(symbol):
return bool(re.search(r'\d', symbol))
def main():
tests = [
["0", True],
["-", False],
["o", False],
["1", True],
["2", True],
["!", False],
["@", False],
["... | Solve Code Fights is digit problem#!/usr/local/bin/python
# Code Fights Is Digit Problem
import re
def isDigit(symbol):
return bool(re.search(r'\d', symbol))
def main():
tests = [
["0", True],
["-", False],
["o", False],
["1", True],
["2", True],
["!", False]... | <commit_before><commit_msg>Solve Code Fights is digit problem<commit_after>#!/usr/local/bin/python
# Code Fights Is Digit Problem
import re
def isDigit(symbol):
return bool(re.search(r'\d', symbol))
def main():
tests = [
["0", True],
["-", False],
["o", False],
["1", True],
... | |
71b9ae31e34504b5805d37e72f7161aaa56ad5f8 | pyconcz_2016/cfp/migrations/0002_auto_20160716_2222.py | pyconcz_2016/cfp/migrations/0002_auto_20160716_2222.py | # -*- coding: utf-8 -*-
# Generated by Django 1.9.3 on 2016-07-16 20:22
from __future__ import unicode_literals
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
('cfp', '0001_initial'),
]
operations = [
migrati... | Add migrations for previous commit | Add migrations for previous commit
| Python | mit | benabraham/cz.pycon.org-2017,pyvec/cz.pycon.org-2017,pyvec/cz.pycon.org-2017,pyvec/cz.pycon.org-2016,benabraham/cz.pycon.org-2017,pyvec/cz.pycon.org-2016,benabraham/cz.pycon.org-2017,pyvec/cz.pycon.org-2017,pyvec/cz.pycon.org-2016 | Add migrations for previous commit | # -*- coding: utf-8 -*-
# Generated by Django 1.9.3 on 2016-07-16 20:22
from __future__ import unicode_literals
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
('cfp', '0001_initial'),
]
operations = [
migrati... | <commit_before><commit_msg>Add migrations for previous commit<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.9.3 on 2016-07-16 20:22
from __future__ import unicode_literals
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
('cfp', '0001_initial'),
]
operations = [
migrati... | Add migrations for previous commit# -*- coding: utf-8 -*-
# Generated by Django 1.9.3 on 2016-07-16 20:22
from __future__ import unicode_literals
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
('cfp', '0001_initial'),
]
... | <commit_before><commit_msg>Add migrations for previous commit<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.9.3 on 2016-07-16 20:22
from __future__ import unicode_literals
from django.db import migrations, models
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [... | |
d94ab69dff6fc950b27105bf334454fcee83a1b8 | create_properties.py | create_properties.py | import enki
e = enki.Enki('a', 'http://localhost:5001', 'translationsvoting')
e.get_all()
lines = []
for t in e.tasks:
msgstr = e.task_runs_df[t.id]['msgstr'].describe()['top']
var_id = t.info['var_id']
line = "%s= %s\n" % (var_id, msgstr)
lines.append(line)
file = open('/tmp/file.properties', 'w')
... | Create properties files from votes. | Create properties files from votes.
| Python | agpl-3.0 | PyBossa/app-translations | Create properties files from votes. | import enki
e = enki.Enki('a', 'http://localhost:5001', 'translationsvoting')
e.get_all()
lines = []
for t in e.tasks:
msgstr = e.task_runs_df[t.id]['msgstr'].describe()['top']
var_id = t.info['var_id']
line = "%s= %s\n" % (var_id, msgstr)
lines.append(line)
file = open('/tmp/file.properties', 'w')
... | <commit_before><commit_msg>Create properties files from votes.<commit_after> | import enki
e = enki.Enki('a', 'http://localhost:5001', 'translationsvoting')
e.get_all()
lines = []
for t in e.tasks:
msgstr = e.task_runs_df[t.id]['msgstr'].describe()['top']
var_id = t.info['var_id']
line = "%s= %s\n" % (var_id, msgstr)
lines.append(line)
file = open('/tmp/file.properties', 'w')
... | Create properties files from votes.import enki
e = enki.Enki('a', 'http://localhost:5001', 'translationsvoting')
e.get_all()
lines = []
for t in e.tasks:
msgstr = e.task_runs_df[t.id]['msgstr'].describe()['top']
var_id = t.info['var_id']
line = "%s= %s\n" % (var_id, msgstr)
lines.append(line)
file =... | <commit_before><commit_msg>Create properties files from votes.<commit_after>import enki
e = enki.Enki('a', 'http://localhost:5001', 'translationsvoting')
e.get_all()
lines = []
for t in e.tasks:
msgstr = e.task_runs_df[t.id]['msgstr'].describe()['top']
var_id = t.info['var_id']
line = "%s= %s\n" % (var_i... | |
e7caa52ffdc547c66454a200e5e9901d6015bb3e | hunittest/test/test_stopwatch.py | hunittest/test/test_stopwatch.py | # -*- encoding: utf-8 -*-
"""Test 'stopwath' module.
"""
import unittest
import time
from datetime import timedelta
from hunittest.stopwatch import StopWatch
class TestStopWatch(unittest.TestCase):
def assertTimedeltaAlmostEqual(self, td1, td2, prec=1e-3):
return abs(td1 - td2).total_seconds <= prec
... | Add unit test for StopWatch. | Add unit test for StopWatch.
| Python | bsd-2-clause | nicolasdespres/hunittest | Add unit test for StopWatch. | # -*- encoding: utf-8 -*-
"""Test 'stopwath' module.
"""
import unittest
import time
from datetime import timedelta
from hunittest.stopwatch import StopWatch
class TestStopWatch(unittest.TestCase):
def assertTimedeltaAlmostEqual(self, td1, td2, prec=1e-3):
return abs(td1 - td2).total_seconds <= prec
... | <commit_before><commit_msg>Add unit test for StopWatch.<commit_after> | # -*- encoding: utf-8 -*-
"""Test 'stopwath' module.
"""
import unittest
import time
from datetime import timedelta
from hunittest.stopwatch import StopWatch
class TestStopWatch(unittest.TestCase):
def assertTimedeltaAlmostEqual(self, td1, td2, prec=1e-3):
return abs(td1 - td2).total_seconds <= prec
... | Add unit test for StopWatch.# -*- encoding: utf-8 -*-
"""Test 'stopwath' module.
"""
import unittest
import time
from datetime import timedelta
from hunittest.stopwatch import StopWatch
class TestStopWatch(unittest.TestCase):
def assertTimedeltaAlmostEqual(self, td1, td2, prec=1e-3):
return abs(td1 - ... | <commit_before><commit_msg>Add unit test for StopWatch.<commit_after># -*- encoding: utf-8 -*-
"""Test 'stopwath' module.
"""
import unittest
import time
from datetime import timedelta
from hunittest.stopwatch import StopWatch
class TestStopWatch(unittest.TestCase):
def assertTimedeltaAlmostEqual(self, td1, t... | |
b246723d837cddde898316b67ab5af1feba67817 | zerver/migrations/0279_message_recipient_subject_indexes.py | zerver/migrations/0279_message_recipient_subject_indexes.py | # Generated by Django 2.2.12 on 2020-04-30 00:35
from django.db import migrations
class Migration(migrations.Migration):
atomic = False
dependencies = [
('zerver', '0278_remove_userprofile_alert_words'),
]
operations = [
migrations.RunSQL("""
CREATE INDEX CONCURRENTLY IF NOT... | Add Message indexes on (recipient_id, subject). | migrations: Add Message indexes on (recipient_id, subject).
Our previous set of indexes for the Message table did not contain
anything to optimize queries for all the messages in a topic in an
organization where the same topic name might appear in 10,000s of
messages in many streams.
We add two indexes here to suppor... | Python | apache-2.0 | andersk/zulip,timabbott/zulip,rht/zulip,zulip/zulip,synicalsyntax/zulip,showell/zulip,kou/zulip,rht/zulip,eeshangarg/zulip,brainwane/zulip,andersk/zulip,zulip/zulip,kou/zulip,timabbott/zulip,zulip/zulip,eeshangarg/zulip,andersk/zulip,eeshangarg/zulip,timabbott/zulip,showell/zulip,showell/zulip,hackerkid/zulip,shubhamdh... | migrations: Add Message indexes on (recipient_id, subject).
Our previous set of indexes for the Message table did not contain
anything to optimize queries for all the messages in a topic in an
organization where the same topic name might appear in 10,000s of
messages in many streams.
We add two indexes here to suppor... | # Generated by Django 2.2.12 on 2020-04-30 00:35
from django.db import migrations
class Migration(migrations.Migration):
atomic = False
dependencies = [
('zerver', '0278_remove_userprofile_alert_words'),
]
operations = [
migrations.RunSQL("""
CREATE INDEX CONCURRENTLY IF NOT... | <commit_before><commit_msg>migrations: Add Message indexes on (recipient_id, subject).
Our previous set of indexes for the Message table did not contain
anything to optimize queries for all the messages in a topic in an
organization where the same topic name might appear in 10,000s of
messages in many streams.
We add... | # Generated by Django 2.2.12 on 2020-04-30 00:35
from django.db import migrations
class Migration(migrations.Migration):
atomic = False
dependencies = [
('zerver', '0278_remove_userprofile_alert_words'),
]
operations = [
migrations.RunSQL("""
CREATE INDEX CONCURRENTLY IF NOT... | migrations: Add Message indexes on (recipient_id, subject).
Our previous set of indexes for the Message table did not contain
anything to optimize queries for all the messages in a topic in an
organization where the same topic name might appear in 10,000s of
messages in many streams.
We add two indexes here to suppor... | <commit_before><commit_msg>migrations: Add Message indexes on (recipient_id, subject).
Our previous set of indexes for the Message table did not contain
anything to optimize queries for all the messages in a topic in an
organization where the same topic name might appear in 10,000s of
messages in many streams.
We add... | |
da596f8c04cb96355a12087eabea7151eb8771a3 | osf_tests/test_maintenance.py | osf_tests/test_maintenance.py | import unittest
from datetime import timedelta
import pytest
from django.utils import timezone
from website import maintenance
from osf.models import MaintenanceState
pytestmark = pytest.mark.django_db
class TestMaintenance(unittest.TestCase):
def tearDown(self):
MaintenanceState.objects.all().delete... | Add tests for maintenance functions | Add tests for maintenance functions
| Python | apache-2.0 | erinspace/osf.io,cslzchen/osf.io,saradbowman/osf.io,Johnetordoff/osf.io,adlius/osf.io,adlius/osf.io,erinspace/osf.io,caseyrollins/osf.io,laurenrevere/osf.io,sloria/osf.io,pattisdr/osf.io,icereval/osf.io,baylee-d/osf.io,binoculars/osf.io,crcresearch/osf.io,icereval/osf.io,caneruguz/osf.io,chrisseto/osf.io,TomBaxter/osf.... | Add tests for maintenance functions | import unittest
from datetime import timedelta
import pytest
from django.utils import timezone
from website import maintenance
from osf.models import MaintenanceState
pytestmark = pytest.mark.django_db
class TestMaintenance(unittest.TestCase):
def tearDown(self):
MaintenanceState.objects.all().delete... | <commit_before><commit_msg>Add tests for maintenance functions<commit_after> | import unittest
from datetime import timedelta
import pytest
from django.utils import timezone
from website import maintenance
from osf.models import MaintenanceState
pytestmark = pytest.mark.django_db
class TestMaintenance(unittest.TestCase):
def tearDown(self):
MaintenanceState.objects.all().delete... | Add tests for maintenance functionsimport unittest
from datetime import timedelta
import pytest
from django.utils import timezone
from website import maintenance
from osf.models import MaintenanceState
pytestmark = pytest.mark.django_db
class TestMaintenance(unittest.TestCase):
def tearDown(self):
Ma... | <commit_before><commit_msg>Add tests for maintenance functions<commit_after>import unittest
from datetime import timedelta
import pytest
from django.utils import timezone
from website import maintenance
from osf.models import MaintenanceState
pytestmark = pytest.mark.django_db
class TestMaintenance(unittest.TestC... | |
7079c8d78256bb24608f8a74c26273ab653c886a | pal/grammars/test_grammars.py | pal/grammars/test_grammars.py | import re
from collections import defaultdict
from pal.grammars.grammars import make_chomsky_normal_form
from pal.grammars.grammars import parse_grammar_from_file
from pal.grammars.parser import generate_grammar_features
from pal.grammars.parser import parse
_GRAMMARS_DIR = 'pal/grammars/services'
_EXAMPLES_FILE = '... | Add test for service grammars: hit rate against example queries | Add test for service grammars: hit rate against example queries
| Python | bsd-3-clause | Machyne/pal,Machyne/pal,Machyne/pal,Machyne/pal | Add test for service grammars: hit rate against example queries | import re
from collections import defaultdict
from pal.grammars.grammars import make_chomsky_normal_form
from pal.grammars.grammars import parse_grammar_from_file
from pal.grammars.parser import generate_grammar_features
from pal.grammars.parser import parse
_GRAMMARS_DIR = 'pal/grammars/services'
_EXAMPLES_FILE = '... | <commit_before><commit_msg>Add test for service grammars: hit rate against example queries<commit_after> | import re
from collections import defaultdict
from pal.grammars.grammars import make_chomsky_normal_form
from pal.grammars.grammars import parse_grammar_from_file
from pal.grammars.parser import generate_grammar_features
from pal.grammars.parser import parse
_GRAMMARS_DIR = 'pal/grammars/services'
_EXAMPLES_FILE = '... | Add test for service grammars: hit rate against example queriesimport re
from collections import defaultdict
from pal.grammars.grammars import make_chomsky_normal_form
from pal.grammars.grammars import parse_grammar_from_file
from pal.grammars.parser import generate_grammar_features
from pal.grammars.parser import par... | <commit_before><commit_msg>Add test for service grammars: hit rate against example queries<commit_after>import re
from collections import defaultdict
from pal.grammars.grammars import make_chomsky_normal_form
from pal.grammars.grammars import parse_grammar_from_file
from pal.grammars.parser import generate_grammar_fea... | |
470b217e8ca8687414ff4cad39ca7cfe4710d956 | tests/functional/test_pip_runner_script.py | tests/functional/test_pip_runner_script.py | import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the e... | Add a test that the runner script works in environments without pip | Add a test that the runner script works in environments without pip
This ensures that the runner script can be used in environments where
pip is not installed.
| Python | mit | sbidoul/pip,pypa/pip,pradyunsg/pip,pfmoore/pip,pradyunsg/pip,pypa/pip,sbidoul/pip,pfmoore/pip | Add a test that the runner script works in environments without pip
This ensures that the runner script can be used in environments where
pip is not installed. | import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the e... | <commit_before><commit_msg>Add a test that the runner script works in environments without pip
This ensures that the runner script can be used in environments where
pip is not installed.<commit_after> | import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
script: PipTestEnvironment, pip_src: Path
) -> None:
runner = pip_src / "src" / "pip" / "__pip-runner__.py"
# Ensure there's no pip installed in the e... | Add a test that the runner script works in environments without pip
This ensures that the runner script can be used in environments where
pip is not installed.import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_runner_work_in_environments_with_no_pip(
... | <commit_before><commit_msg>Add a test that the runner script works in environments without pip
This ensures that the runner script can be used in environments where
pip is not installed.<commit_after>import os
from pathlib import Path
from pip import __version__
from tests.lib import PipTestEnvironment
def test_run... | |
34c9dbb30a731643b6a8747995fbf760dbd5377c | tests/header_test.py | tests/header_test.py | #! /usr/bin/env python
# -*- coding: iso-8859-1 -*-
# vi:ts=4:et
import pycurl
import unittest
from . import appmanager
from . import util
setup_module, teardown_module = appmanager.setup(('app', 8380))
class HeaderTest(unittest.TestCase):
def setUp(self):
self.curl = pycurl.Curl()
def tearDown... | Add a test for sending unicode data in http headers | Add a test for sending unicode data in http headers
| Python | lgpl-2.1 | pycurl/pycurl,pycurl/pycurl,pycurl/pycurl | Add a test for sending unicode data in http headers | #! /usr/bin/env python
# -*- coding: iso-8859-1 -*-
# vi:ts=4:et
import pycurl
import unittest
from . import appmanager
from . import util
setup_module, teardown_module = appmanager.setup(('app', 8380))
class HeaderTest(unittest.TestCase):
def setUp(self):
self.curl = pycurl.Curl()
def tearDown... | <commit_before><commit_msg>Add a test for sending unicode data in http headers<commit_after> | #! /usr/bin/env python
# -*- coding: iso-8859-1 -*-
# vi:ts=4:et
import pycurl
import unittest
from . import appmanager
from . import util
setup_module, teardown_module = appmanager.setup(('app', 8380))
class HeaderTest(unittest.TestCase):
def setUp(self):
self.curl = pycurl.Curl()
def tearDown... | Add a test for sending unicode data in http headers#! /usr/bin/env python
# -*- coding: iso-8859-1 -*-
# vi:ts=4:et
import pycurl
import unittest
from . import appmanager
from . import util
setup_module, teardown_module = appmanager.setup(('app', 8380))
class HeaderTest(unittest.TestCase):
def setUp(self):
... | <commit_before><commit_msg>Add a test for sending unicode data in http headers<commit_after>#! /usr/bin/env python
# -*- coding: iso-8859-1 -*-
# vi:ts=4:et
import pycurl
import unittest
from . import appmanager
from . import util
setup_module, teardown_module = appmanager.setup(('app', 8380))
class HeaderTest(unit... | |
a205fbc29d2c6c8004b91c3fed5b2a03427e62a4 | locations/spiders/victra.py | locations/spiders/victra.py | # -*- coding: utf-8 -*-
import json
import re
import scrapy
from locations.items import GeojsonPointItem
from locations.hours import OpeningHours
class VictraSpider(scrapy.Spider):
name = "victra"
allowed_domains = []
start_urls = [
'https://victra.com/Handlers/LocationData.ashx',
]
def ... | Add spider for Victra mobile phone stores | Add spider for Victra mobile phone stores
| Python | mit | iandees/all-the-places,iandees/all-the-places,iandees/all-the-places | Add spider for Victra mobile phone stores | # -*- coding: utf-8 -*-
import json
import re
import scrapy
from locations.items import GeojsonPointItem
from locations.hours import OpeningHours
class VictraSpider(scrapy.Spider):
name = "victra"
allowed_domains = []
start_urls = [
'https://victra.com/Handlers/LocationData.ashx',
]
def ... | <commit_before><commit_msg>Add spider for Victra mobile phone stores<commit_after> | # -*- coding: utf-8 -*-
import json
import re
import scrapy
from locations.items import GeojsonPointItem
from locations.hours import OpeningHours
class VictraSpider(scrapy.Spider):
name = "victra"
allowed_domains = []
start_urls = [
'https://victra.com/Handlers/LocationData.ashx',
]
def ... | Add spider for Victra mobile phone stores# -*- coding: utf-8 -*-
import json
import re
import scrapy
from locations.items import GeojsonPointItem
from locations.hours import OpeningHours
class VictraSpider(scrapy.Spider):
name = "victra"
allowed_domains = []
start_urls = [
'https://victra.com/Han... | <commit_before><commit_msg>Add spider for Victra mobile phone stores<commit_after># -*- coding: utf-8 -*-
import json
import re
import scrapy
from locations.items import GeojsonPointItem
from locations.hours import OpeningHours
class VictraSpider(scrapy.Spider):
name = "victra"
allowed_domains = []
start... | |
6edadb2fe75345d24f5fec26bd4e519611f651f8 | integration_tests/test_s3_deletion.py | integration_tests/test_s3_deletion.py | #!/usr/bin/env python
"""
Test S3 directory deletion functionality.
Uses the lsst-the-docs-test bucket in lsst-sqre's account. Also assumes that
credentials for that account are in the ltd-dev profile of ~/.aws/credentials.
"""
import sys
import os.path
app_path = os.path.abspath(os.path.join(os.path.dirname(__file__... | Add integration test for S3 object deletion | Add integration test for S3 object deletion
By default, uses the test bucket 'lsst-the-docs-test' and assumes
credentials for the lsst-sqre account are in 'ltd-dev' profile of
~/.aws/credentials.
For DM-4951.
| Python | mit | lsst-sqre/ltd-keeper,lsst-sqre/ltd-keeper | Add integration test for S3 object deletion
By default, uses the test bucket 'lsst-the-docs-test' and assumes
credentials for the lsst-sqre account are in 'ltd-dev' profile of
~/.aws/credentials.
For DM-4951. | #!/usr/bin/env python
"""
Test S3 directory deletion functionality.
Uses the lsst-the-docs-test bucket in lsst-sqre's account. Also assumes that
credentials for that account are in the ltd-dev profile of ~/.aws/credentials.
"""
import sys
import os.path
app_path = os.path.abspath(os.path.join(os.path.dirname(__file__... | <commit_before><commit_msg>Add integration test for S3 object deletion
By default, uses the test bucket 'lsst-the-docs-test' and assumes
credentials for the lsst-sqre account are in 'ltd-dev' profile of
~/.aws/credentials.
For DM-4951.<commit_after> | #!/usr/bin/env python
"""
Test S3 directory deletion functionality.
Uses the lsst-the-docs-test bucket in lsst-sqre's account. Also assumes that
credentials for that account are in the ltd-dev profile of ~/.aws/credentials.
"""
import sys
import os.path
app_path = os.path.abspath(os.path.join(os.path.dirname(__file__... | Add integration test for S3 object deletion
By default, uses the test bucket 'lsst-the-docs-test' and assumes
credentials for the lsst-sqre account are in 'ltd-dev' profile of
~/.aws/credentials.
For DM-4951.#!/usr/bin/env python
"""
Test S3 directory deletion functionality.
Uses the lsst-the-docs-test bucket in lss... | <commit_before><commit_msg>Add integration test for S3 object deletion
By default, uses the test bucket 'lsst-the-docs-test' and assumes
credentials for the lsst-sqre account are in 'ltd-dev' profile of
~/.aws/credentials.
For DM-4951.<commit_after>#!/usr/bin/env python
"""
Test S3 directory deletion functionality.
... | |
15a9ec59e698a16cc37c3e0b430a376e3e73bd04 | tests/test_parser.py | tests/test_parser.py | import unittest
from unittest.mock import patch, call
import whitepy.lexer as lexer
from whitepy.parser import Parser
class TestParser(unittest.TestCase):
def _get_tokens(self, filename):
with open(filename, 'r') as f:
lines = f.read()
item = lexer.Lexer(line=lines)
i... | Add new test for Parser | Add new test for Parser
Tests the main Parser.parse() function with hello world and Fibonacci whitespace
samples
| Python | apache-2.0 | yasn77/whitepy | Add new test for Parser
Tests the main Parser.parse() function with hello world and Fibonacci whitespace
samples | import unittest
from unittest.mock import patch, call
import whitepy.lexer as lexer
from whitepy.parser import Parser
class TestParser(unittest.TestCase):
def _get_tokens(self, filename):
with open(filename, 'r') as f:
lines = f.read()
item = lexer.Lexer(line=lines)
i... | <commit_before><commit_msg>Add new test for Parser
Tests the main Parser.parse() function with hello world and Fibonacci whitespace
samples<commit_after> | import unittest
from unittest.mock import patch, call
import whitepy.lexer as lexer
from whitepy.parser import Parser
class TestParser(unittest.TestCase):
def _get_tokens(self, filename):
with open(filename, 'r') as f:
lines = f.read()
item = lexer.Lexer(line=lines)
i... | Add new test for Parser
Tests the main Parser.parse() function with hello world and Fibonacci whitespace
samplesimport unittest
from unittest.mock import patch, call
import whitepy.lexer as lexer
from whitepy.parser import Parser
class TestParser(unittest.TestCase):
def _get_tokens(self, filename):
wit... | <commit_before><commit_msg>Add new test for Parser
Tests the main Parser.parse() function with hello world and Fibonacci whitespace
samples<commit_after>import unittest
from unittest.mock import patch, call
import whitepy.lexer as lexer
from whitepy.parser import Parser
class TestParser(unittest.TestCase):
def... | |
dc10584666199797b77a759696c56e179ef8ca21 | billjobs/serializers.py | billjobs/serializers.py | from django.contrib.auth.models import User, Group
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('url', 'username', 'email', 'groups')
| Add serializer for User model | Add serializer for User model
| Python | mit | ioO/billjobs | Add serializer for User model | from django.contrib.auth.models import User, Group
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('url', 'username', 'email', 'groups')
| <commit_before><commit_msg>Add serializer for User model<commit_after> | from django.contrib.auth.models import User, Group
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('url', 'username', 'email', 'groups')
| Add serializer for User modelfrom django.contrib.auth.models import User, Group
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('url', 'username', 'email', 'groups')
| <commit_before><commit_msg>Add serializer for User model<commit_after>from django.contrib.auth.models import User, Group
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ('url', 'username', 'email', 'groups')
| |
1c3082e18b63ef0ba7fb73dd69544980da7ad16c | zinnia/migrations/0003_publication_date.py | zinnia/migrations/0003_publication_date.py | from django.db import models
from django.db import migrations
from django.utils import timezone
def fill_publication_date(apps, schema_editor):
Entry = apps.get_model('zinnia', 'Entry')
for entry in Entry.objects.all():
entry.publication_date = entry.creation_date
entry.save()
def unfill_pub... | Add migration for publication_date field | Add migration for publication_date field
| Python | bsd-3-clause | ghachey/django-blog-zinnia,dapeng0802/django-blog-zinnia,petecummings/django-blog-zinnia,aorzh/django-blog-zinnia,extertioner/django-blog-zinnia,Zopieux/django-blog-zinnia,bywbilly/django-blog-zinnia,extertioner/django-blog-zinnia,marctc/django-blog-zinnia,bywbilly/django-blog-zinnia,marctc/django-blog-zinnia,extertion... | Add migration for publication_date field | from django.db import models
from django.db import migrations
from django.utils import timezone
def fill_publication_date(apps, schema_editor):
Entry = apps.get_model('zinnia', 'Entry')
for entry in Entry.objects.all():
entry.publication_date = entry.creation_date
entry.save()
def unfill_pub... | <commit_before><commit_msg>Add migration for publication_date field<commit_after> | from django.db import models
from django.db import migrations
from django.utils import timezone
def fill_publication_date(apps, schema_editor):
Entry = apps.get_model('zinnia', 'Entry')
for entry in Entry.objects.all():
entry.publication_date = entry.creation_date
entry.save()
def unfill_pub... | Add migration for publication_date fieldfrom django.db import models
from django.db import migrations
from django.utils import timezone
def fill_publication_date(apps, schema_editor):
Entry = apps.get_model('zinnia', 'Entry')
for entry in Entry.objects.all():
entry.publication_date = entry.creation_da... | <commit_before><commit_msg>Add migration for publication_date field<commit_after>from django.db import models
from django.db import migrations
from django.utils import timezone
def fill_publication_date(apps, schema_editor):
Entry = apps.get_model('zinnia', 'Entry')
for entry in Entry.objects.all():
e... | |
e9f6031f9368cd036826f40c113e5ca9d420409a | lily/contacts/migrations/0013_auto_20170717_2005.py | lily/contacts/migrations/0013_auto_20170717_2005.py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('contacts', '0012_remove_contact_preposition'),
]
operations = [
migrations.AlterField(
... | Add extra migration for the manager field | Add extra migration for the manager field
| Python | agpl-3.0 | HelloLily/hellolily,HelloLily/hellolily,HelloLily/hellolily,HelloLily/hellolily | Add extra migration for the manager field | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('contacts', '0012_remove_contact_preposition'),
]
operations = [
migrations.AlterField(
... | <commit_before><commit_msg>Add extra migration for the manager field<commit_after> | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('contacts', '0012_remove_contact_preposition'),
]
operations = [
migrations.AlterField(
... | Add extra migration for the manager field# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('contacts', '0012_remove_contact_preposition'),
]
operation... | <commit_before><commit_msg>Add extra migration for the manager field<commit_after># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('contacts', '0012_remove_co... | |
b2635fd52aa474e6dd375ad804f46a16945962a4 | lib/svtplay/service/hbo.py | lib/svtplay/service/hbo.py | class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*).html", other)... | class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, options, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*).html... | Add options argument to get() | HBO: Add options argument to get()
| Python | mit | qnorsten/svtplay-dl,dalgr/svtplay-dl,olof/svtplay-dl,leakim/svtplay-dl,spaam/svtplay-dl,selepo/svtplay-dl,OakNinja/svtplay-dl,iwconfig/svtplay-dl,dalgr/svtplay-dl,spaam/svtplay-dl,OakNinja/svtplay-dl,leakim/svtplay-dl,iwconfig/svtplay-dl,qnorsten/svtplay-dl,selepo/svtplay-dl,OakNinja/svtplay-dl,leakim/svtplay-dl,olof/s... | class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*).html", other)... | class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, options, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*).html... | <commit_before>class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*... | class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, options, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*).html... | class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*).html", other)... | <commit_before>class Hbo():
def handle(self, url):
return "hbo.com" in url
def get(self, url):
parse = urlparse(url)
try:
other = parse[5]
except KeyError:
log.error("Something wrong with that url")
sys.exit(2)
match = re.search("^/(.*... |
c2817959fea63db83a328a8ce61b95031b04c4bd | mopidy/frontends/mpd/__init__.py | mopidy/frontends/mpd/__init__.py | from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.server = None
self.dispatcher = None
def start_server(self, core_queue):
"""
... | from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.process = None
self.dispatcher = None
def start_server(self, core_queue):
"""
... | Initialize the correct variable in init | Initialize the correct variable in init
| Python | apache-2.0 | dbrgn/mopidy,dbrgn/mopidy,ali/mopidy,diandiankan/mopidy,kingosticks/mopidy,SuperStarPL/mopidy,abarisain/mopidy,pacificIT/mopidy,kingosticks/mopidy,tkem/mopidy,bacontext/mopidy,pacificIT/mopidy,SuperStarPL/mopidy,ali/mopidy,vrs01/mopidy,quartz55/mopidy,quartz55/mopidy,tkem/mopidy,tkem/mopidy,mopidy/mopidy,dbrgn/mopidy,g... | from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.server = None
self.dispatcher = None
def start_server(self, core_queue):
"""
... | from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.process = None
self.dispatcher = None
def start_server(self, core_queue):
"""
... | <commit_before>from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.server = None
self.dispatcher = None
def start_server(self, core_queue):
... | from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.process = None
self.dispatcher = None
def start_server(self, core_queue):
"""
... | from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.server = None
self.dispatcher = None
def start_server(self, core_queue):
"""
... | <commit_before>from mopidy.frontends.mpd.dispatcher import MpdDispatcher
from mopidy.frontends.mpd.process import MpdProcess
class MpdFrontend(object):
"""
The MPD frontend.
"""
def __init__(self):
self.server = None
self.dispatcher = None
def start_server(self, core_queue):
... |
df810f11409212edce61263205288cfbb7221a2d | scripts/evaluation/process_api_performance_files.py | scripts/evaluation/process_api_performance_files.py | import sys
import os
output_dir = sys.argv[1]
def filename(num_duplications):
return os.path.join(output_dir, "output_%s.csv" % (str(num_duplications),))
def time_for_method(fname):
with open(fname) as f:
lines = f.readlines()[1:]
lines = [line.replace("\n", "").split(", ") for line in lines]... | Write script to parse API performance data | Write script to parse API performance data
| Python | mit | mitdbg/modeldb,mitdbg/modeldb,mitdbg/modeldb,mitdbg/modeldb,mitdbg/modeldb | Write script to parse API performance data | import sys
import os
output_dir = sys.argv[1]
def filename(num_duplications):
return os.path.join(output_dir, "output_%s.csv" % (str(num_duplications),))
def time_for_method(fname):
with open(fname) as f:
lines = f.readlines()[1:]
lines = [line.replace("\n", "").split(", ") for line in lines]... | <commit_before><commit_msg>Write script to parse API performance data<commit_after> | import sys
import os
output_dir = sys.argv[1]
def filename(num_duplications):
return os.path.join(output_dir, "output_%s.csv" % (str(num_duplications),))
def time_for_method(fname):
with open(fname) as f:
lines = f.readlines()[1:]
lines = [line.replace("\n", "").split(", ") for line in lines]... | Write script to parse API performance dataimport sys
import os
output_dir = sys.argv[1]
def filename(num_duplications):
return os.path.join(output_dir, "output_%s.csv" % (str(num_duplications),))
def time_for_method(fname):
with open(fname) as f:
lines = f.readlines()[1:]
lines = [line.replac... | <commit_before><commit_msg>Write script to parse API performance data<commit_after>import sys
import os
output_dir = sys.argv[1]
def filename(num_duplications):
return os.path.join(output_dir, "output_%s.csv" % (str(num_duplications),))
def time_for_method(fname):
with open(fname) as f:
lines = f.rea... | |
d44dd85a1020da3971a50ceb1c0c8150b05d5334 | scripts/utils/shape-to-csv.py | scripts/utils/shape-to-csv.py | #!/usr/bin/env python3
import csv
import ogr
import osr
import sys
import unicodedata
def usage():
print('shape-to-csv.py PATH_TO_INPUT_SHAPE PATH_TO_OUTPUT_CSV')
def main():
# Inspired by http://gis.stackexchange.com/a/19178
# This is design to work with the places.shp from Swisstzerland.
shpfile ... | Add script to convert ShapeFile to CSV for search | Add script to convert ShapeFile to CSV for search
Used by customer infra to have search.
| Python | agpl-3.0 | ioda-net/geo-infra,ioda-net/geo-infra,ioda-net/geo-infra,ioda-net/geo-infra | Add script to convert ShapeFile to CSV for search
Used by customer infra to have search. | #!/usr/bin/env python3
import csv
import ogr
import osr
import sys
import unicodedata
def usage():
print('shape-to-csv.py PATH_TO_INPUT_SHAPE PATH_TO_OUTPUT_CSV')
def main():
# Inspired by http://gis.stackexchange.com/a/19178
# This is design to work with the places.shp from Swisstzerland.
shpfile ... | <commit_before><commit_msg>Add script to convert ShapeFile to CSV for search
Used by customer infra to have search.<commit_after> | #!/usr/bin/env python3
import csv
import ogr
import osr
import sys
import unicodedata
def usage():
print('shape-to-csv.py PATH_TO_INPUT_SHAPE PATH_TO_OUTPUT_CSV')
def main():
# Inspired by http://gis.stackexchange.com/a/19178
# This is design to work with the places.shp from Swisstzerland.
shpfile ... | Add script to convert ShapeFile to CSV for search
Used by customer infra to have search.#!/usr/bin/env python3
import csv
import ogr
import osr
import sys
import unicodedata
def usage():
print('shape-to-csv.py PATH_TO_INPUT_SHAPE PATH_TO_OUTPUT_CSV')
def main():
# Inspired by http://gis.stackexchange.com/... | <commit_before><commit_msg>Add script to convert ShapeFile to CSV for search
Used by customer infra to have search.<commit_after>#!/usr/bin/env python3
import csv
import ogr
import osr
import sys
import unicodedata
def usage():
print('shape-to-csv.py PATH_TO_INPUT_SHAPE PATH_TO_OUTPUT_CSV')
def main():
# ... | |
ddc0b056af46bf50f3e5c4213f7598d645d378f7 | openprescribing/matrixstore/tests/test_cachelib.py | openprescribing/matrixstore/tests/test_cachelib.py | from mock import Mock
import warnings
from django.core.cache import CacheKeyWarning
from django.test import SimpleTestCase, override_settings
from matrixstore.cachelib import memoize
# The local memory cache backend we use in testing warns that our binary cache
# keys won't be compatible with memcached, but we reall... | Add tests for cachelib decorator | Add tests for cachelib decorator
| Python | mit | ebmdatalab/openprescribing,annapowellsmith/openpresc,annapowellsmith/openpresc,annapowellsmith/openpresc,annapowellsmith/openpresc,ebmdatalab/openprescribing,ebmdatalab/openprescribing,ebmdatalab/openprescribing | Add tests for cachelib decorator | from mock import Mock
import warnings
from django.core.cache import CacheKeyWarning
from django.test import SimpleTestCase, override_settings
from matrixstore.cachelib import memoize
# The local memory cache backend we use in testing warns that our binary cache
# keys won't be compatible with memcached, but we reall... | <commit_before><commit_msg>Add tests for cachelib decorator<commit_after> | from mock import Mock
import warnings
from django.core.cache import CacheKeyWarning
from django.test import SimpleTestCase, override_settings
from matrixstore.cachelib import memoize
# The local memory cache backend we use in testing warns that our binary cache
# keys won't be compatible with memcached, but we reall... | Add tests for cachelib decoratorfrom mock import Mock
import warnings
from django.core.cache import CacheKeyWarning
from django.test import SimpleTestCase, override_settings
from matrixstore.cachelib import memoize
# The local memory cache backend we use in testing warns that our binary cache
# keys won't be compati... | <commit_before><commit_msg>Add tests for cachelib decorator<commit_after>from mock import Mock
import warnings
from django.core.cache import CacheKeyWarning
from django.test import SimpleTestCase, override_settings
from matrixstore.cachelib import memoize
# The local memory cache backend we use in testing warns that... | |
e02679577219cea30896532be9ba6aea457e2454 | mixpanel_query/paginator.py | mixpanel_query/paginator.py | import math
import itertools
from multiprocessing.pool import ThreadPool
class ConcurrentPaginator(object):
"""
Concurrently fetches all pages in a paginated collection.
Currently, only the people API (`/api/2.0/engage`) supports pagination.
This class is designed to support the people API's implement... | Add ConcurrentPaginator (for the engage API) | Add ConcurrentPaginator (for the engage API)
| Python | mit | cooncesean/mixpanel-query-py | Add ConcurrentPaginator (for the engage API) | import math
import itertools
from multiprocessing.pool import ThreadPool
class ConcurrentPaginator(object):
"""
Concurrently fetches all pages in a paginated collection.
Currently, only the people API (`/api/2.0/engage`) supports pagination.
This class is designed to support the people API's implement... | <commit_before><commit_msg>Add ConcurrentPaginator (for the engage API)<commit_after> | import math
import itertools
from multiprocessing.pool import ThreadPool
class ConcurrentPaginator(object):
"""
Concurrently fetches all pages in a paginated collection.
Currently, only the people API (`/api/2.0/engage`) supports pagination.
This class is designed to support the people API's implement... | Add ConcurrentPaginator (for the engage API)import math
import itertools
from multiprocessing.pool import ThreadPool
class ConcurrentPaginator(object):
"""
Concurrently fetches all pages in a paginated collection.
Currently, only the people API (`/api/2.0/engage`) supports pagination.
This class is de... | <commit_before><commit_msg>Add ConcurrentPaginator (for the engage API)<commit_after>import math
import itertools
from multiprocessing.pool import ThreadPool
class ConcurrentPaginator(object):
"""
Concurrently fetches all pages in a paginated collection.
Currently, only the people API (`/api/2.0/engage`) ... | |
b224a281a11cdc4d0c632e451c217be34dccadcc | app/main/forms/suppliers.py | app/main/forms/suppliers.py | from flask.ext.wtf import Form
from wtforms import IntegerField, StringField, FieldList
from wtforms.validators import DataRequired, Email
class EditSupplierForm(Form):
description = StringField()
clients = FieldList(StringField(), max_entries=10)
class EditContactInformationForm(Form):
id = IntegerFiel... | Add forms for supplier and contact information | Add forms for supplier and contact information
Forms allow us to validate basic field requirements before making
a request to the API, since the API doesn't return per-field errors
at the moment.
Supplier and contact information have separate forms, since they
require separate API requests to save the data and might ... | Python | mit | mtekel/digitalmarketplace-supplier-frontend,mtekel/digitalmarketplace-supplier-frontend,mtekel/digitalmarketplace-supplier-frontend,mtekel/digitalmarketplace-supplier-frontend,alphagov/digitalmarketplace-supplier-frontend,alphagov/digitalmarketplace-supplier-frontend,alphagov/digitalmarketplace-supplier-frontend,alphag... | Add forms for supplier and contact information
Forms allow us to validate basic field requirements before making
a request to the API, since the API doesn't return per-field errors
at the moment.
Supplier and contact information have separate forms, since they
require separate API requests to save the data and might ... | from flask.ext.wtf import Form
from wtforms import IntegerField, StringField, FieldList
from wtforms.validators import DataRequired, Email
class EditSupplierForm(Form):
description = StringField()
clients = FieldList(StringField(), max_entries=10)
class EditContactInformationForm(Form):
id = IntegerFiel... | <commit_before><commit_msg>Add forms for supplier and contact information
Forms allow us to validate basic field requirements before making
a request to the API, since the API doesn't return per-field errors
at the moment.
Supplier and contact information have separate forms, since they
require separate API requests ... | from flask.ext.wtf import Form
from wtforms import IntegerField, StringField, FieldList
from wtforms.validators import DataRequired, Email
class EditSupplierForm(Form):
description = StringField()
clients = FieldList(StringField(), max_entries=10)
class EditContactInformationForm(Form):
id = IntegerFiel... | Add forms for supplier and contact information
Forms allow us to validate basic field requirements before making
a request to the API, since the API doesn't return per-field errors
at the moment.
Supplier and contact information have separate forms, since they
require separate API requests to save the data and might ... | <commit_before><commit_msg>Add forms for supplier and contact information
Forms allow us to validate basic field requirements before making
a request to the API, since the API doesn't return per-field errors
at the moment.
Supplier and contact information have separate forms, since they
require separate API requests ... | |
48a03c6e593e647e52a96aabcfc8ac02d7c93773 | tfx/examples/imdb/imdb_fetch_data.py | tfx/examples/imdb/imdb_fetch_data.py | import os
import pandas as pd
import tensorflow_datasets as tfds
# Example use in another file of this directory:
# import imdb_fetch_data as full_data
# full_data.fetch_data()
def fetch_data():
"""This downloads the full dataset to pwd/data/imdb.csv"""
ds = tfds.load('imdb_reviews', split='train+... | Add fetch data py script | Add fetch data py script
Add .py file with code to fetch data defined in function fetch_data() | Python | apache-2.0 | tensorflow/tfx,tensorflow/tfx | Add fetch data py script
Add .py file with code to fetch data defined in function fetch_data() | import os
import pandas as pd
import tensorflow_datasets as tfds
# Example use in another file of this directory:
# import imdb_fetch_data as full_data
# full_data.fetch_data()
def fetch_data():
"""This downloads the full dataset to pwd/data/imdb.csv"""
ds = tfds.load('imdb_reviews', split='train+... | <commit_before><commit_msg>Add fetch data py script
Add .py file with code to fetch data defined in function fetch_data()<commit_after> | import os
import pandas as pd
import tensorflow_datasets as tfds
# Example use in another file of this directory:
# import imdb_fetch_data as full_data
# full_data.fetch_data()
def fetch_data():
"""This downloads the full dataset to pwd/data/imdb.csv"""
ds = tfds.load('imdb_reviews', split='train+... | Add fetch data py script
Add .py file with code to fetch data defined in function fetch_data()import os
import pandas as pd
import tensorflow_datasets as tfds
# Example use in another file of this directory:
# import imdb_fetch_data as full_data
# full_data.fetch_data()
def fetch_data():
"""This downl... | <commit_before><commit_msg>Add fetch data py script
Add .py file with code to fetch data defined in function fetch_data()<commit_after>import os
import pandas as pd
import tensorflow_datasets as tfds
# Example use in another file of this directory:
# import imdb_fetch_data as full_data
# full_data.fetch_data(... | |
4d14acb3f805c26f1d917639f82974a006df9d4a | 2048/test_2048.py | 2048/test_2048.py | from __future__ import print_function
import numpy as np
import math
np.random.seed(1337) # for reproducibility
from keras.datasets import mnist
from keras.models import Sequential, model_from_json
from keras.layers.core import Dense, Dropout, Activation, Flatten
from keras.layers.convolutional import Convolution2D, ... | Test qlearning4k result for 2048 | Test qlearning4k result for 2048
| Python | mit | choupi/NDHUDLWorkshop | Test qlearning4k result for 2048 | from __future__ import print_function
import numpy as np
import math
np.random.seed(1337) # for reproducibility
from keras.datasets import mnist
from keras.models import Sequential, model_from_json
from keras.layers.core import Dense, Dropout, Activation, Flatten
from keras.layers.convolutional import Convolution2D, ... | <commit_before><commit_msg>Test qlearning4k result for 2048<commit_after> | from __future__ import print_function
import numpy as np
import math
np.random.seed(1337) # for reproducibility
from keras.datasets import mnist
from keras.models import Sequential, model_from_json
from keras.layers.core import Dense, Dropout, Activation, Flatten
from keras.layers.convolutional import Convolution2D, ... | Test qlearning4k result for 2048from __future__ import print_function
import numpy as np
import math
np.random.seed(1337) # for reproducibility
from keras.datasets import mnist
from keras.models import Sequential, model_from_json
from keras.layers.core import Dense, Dropout, Activation, Flatten
from keras.layers.conv... | <commit_before><commit_msg>Test qlearning4k result for 2048<commit_after>from __future__ import print_function
import numpy as np
import math
np.random.seed(1337) # for reproducibility
from keras.datasets import mnist
from keras.models import Sequential, model_from_json
from keras.layers.core import Dense, Dropout, A... | |
70c3fbbd4aee9a5dfd75c6891fe479ba894ace38 | strategic_form.py | strategic_form.py | from numpy import array
def transpose(payoff_matrix):
return array(payoff_matrix).transpose().tolist()
def get_best_responses(payoff_matrix):
# Select argmax from each row, and return the result as a list
return list(map(lambda x: (payoff_matrix.index(x), x.index(max(x))), payoff_matrix))
def solve_psne_2(payo... | Add algorithm for finding PSNE in two-player simultaneous-move games | Add algorithm for finding PSNE in two-player simultaneous-move games
| Python | mit | kubkon/py-game-theory | Add algorithm for finding PSNE in two-player simultaneous-move games | from numpy import array
def transpose(payoff_matrix):
return array(payoff_matrix).transpose().tolist()
def get_best_responses(payoff_matrix):
# Select argmax from each row, and return the result as a list
return list(map(lambda x: (payoff_matrix.index(x), x.index(max(x))), payoff_matrix))
def solve_psne_2(payo... | <commit_before><commit_msg>Add algorithm for finding PSNE in two-player simultaneous-move games<commit_after> | from numpy import array
def transpose(payoff_matrix):
return array(payoff_matrix).transpose().tolist()
def get_best_responses(payoff_matrix):
# Select argmax from each row, and return the result as a list
return list(map(lambda x: (payoff_matrix.index(x), x.index(max(x))), payoff_matrix))
def solve_psne_2(payo... | Add algorithm for finding PSNE in two-player simultaneous-move gamesfrom numpy import array
def transpose(payoff_matrix):
return array(payoff_matrix).transpose().tolist()
def get_best_responses(payoff_matrix):
# Select argmax from each row, and return the result as a list
return list(map(lambda x: (payoff_matri... | <commit_before><commit_msg>Add algorithm for finding PSNE in two-player simultaneous-move games<commit_after>from numpy import array
def transpose(payoff_matrix):
return array(payoff_matrix).transpose().tolist()
def get_best_responses(payoff_matrix):
# Select argmax from each row, and return the result as a list
... | |
1c10d27733c5448aaf9aa47c19ff3b279b3f0174 | yahoo_historical/tests/test_fetch.py | yahoo_historical/tests/test_fetch.py | from yahoo_historical import Fetcher
def test_get_historical():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_historical()
assert len(data) > 0
def test_get_dividends():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_dividends()
assert len(data) > 0
def test_get_splits():
data... | Add basic unit tests for CI job | Add basic unit tests for CI job
| Python | mit | AndrewRPorter/yahoo-historical | Add basic unit tests for CI job | from yahoo_historical import Fetcher
def test_get_historical():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_historical()
assert len(data) > 0
def test_get_dividends():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_dividends()
assert len(data) > 0
def test_get_splits():
data... | <commit_before><commit_msg>Add basic unit tests for CI job<commit_after> | from yahoo_historical import Fetcher
def test_get_historical():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_historical()
assert len(data) > 0
def test_get_dividends():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_dividends()
assert len(data) > 0
def test_get_splits():
data... | Add basic unit tests for CI jobfrom yahoo_historical import Fetcher
def test_get_historical():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_historical()
assert len(data) > 0
def test_get_dividends():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_dividends()
assert len(data) > 0
... | <commit_before><commit_msg>Add basic unit tests for CI job<commit_after>from yahoo_historical import Fetcher
def test_get_historical():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).get_historical()
assert len(data) > 0
def test_get_dividends():
data = Fetcher("AAPL", [2007, 1, 1], [2017, 1, 1]).ge... | |
da005126cdc6ac78a464ac55ae50f15c99c90b4a | bin/crosswalk_msa_county.py | bin/crosswalk_msa_county.py | """build_2000_msa.py
Extract a usable crosswalk between 2000 Metropolitan areas and counties.
Reconstitute a shapefile for the 2000 MSAs from the counties.
Parameters
----------
99mfips.txt: Delineation of the 2000 Metropolitan Statistical Areas
Returns
-------
crosswalk_msa_county.csv
msa.shp
"""
#
# Parse the ... | Add script to extract the 2000 MSA to county correspondence | Add script to extract the 2000 MSA to county correspondence
| Python | bsd-3-clause | rlouf/patterns-of-segregation | Add script to extract the 2000 MSA to county correspondence | """build_2000_msa.py
Extract a usable crosswalk between 2000 Metropolitan areas and counties.
Reconstitute a shapefile for the 2000 MSAs from the counties.
Parameters
----------
99mfips.txt: Delineation of the 2000 Metropolitan Statistical Areas
Returns
-------
crosswalk_msa_county.csv
msa.shp
"""
#
# Parse the ... | <commit_before><commit_msg>Add script to extract the 2000 MSA to county correspondence<commit_after> | """build_2000_msa.py
Extract a usable crosswalk between 2000 Metropolitan areas and counties.
Reconstitute a shapefile for the 2000 MSAs from the counties.
Parameters
----------
99mfips.txt: Delineation of the 2000 Metropolitan Statistical Areas
Returns
-------
crosswalk_msa_county.csv
msa.shp
"""
#
# Parse the ... | Add script to extract the 2000 MSA to county correspondence"""build_2000_msa.py
Extract a usable crosswalk between 2000 Metropolitan areas and counties.
Reconstitute a shapefile for the 2000 MSAs from the counties.
Parameters
----------
99mfips.txt: Delineation of the 2000 Metropolitan Statistical Areas
Returns
---... | <commit_before><commit_msg>Add script to extract the 2000 MSA to county correspondence<commit_after>"""build_2000_msa.py
Extract a usable crosswalk between 2000 Metropolitan areas and counties.
Reconstitute a shapefile for the 2000 MSAs from the counties.
Parameters
----------
99mfips.txt: Delineation of the 2000 Me... | |
8054c9e06511041c0834f901e2e515e58100e8f7 | bindings/const_generator.py | bindings/const_generator.py | import sys, re
INCL_DIR = '../include'
include = [
('/arm.h', 'ARM_'),
('/arm64.h', 'ARM64_'),
('/x86.h', 'X86_'),
('/mips.h', 'MIPS_'),
]
template = {
'java': {
'header': "// AUTOGENRATED FILE, DO NOT EDIT\npackage capstone;\n\npublic class %sconst {\n",
'footer': "}",
... | Add a script to generate constant for binding | Add a script to generate constant for binding
| Python | bsd-3-clause | dynm/capstone,07151129/capstone,nplanel/capstone,nplanel/capstone,bughoho/capstone,krytarowski/capstone,angelabier1/capstone,nplanel/capstone,07151129/capstone,bigendiansmalls/capstone,bowlofstew/capstone,pranith/capstone,code4bones/capstone,sigma-random/capstone,sephiroth99/capstone,NeilBryant/capstone,fvrmatteo/capst... | Add a script to generate constant for binding | import sys, re
INCL_DIR = '../include'
include = [
('/arm.h', 'ARM_'),
('/arm64.h', 'ARM64_'),
('/x86.h', 'X86_'),
('/mips.h', 'MIPS_'),
]
template = {
'java': {
'header': "// AUTOGENRATED FILE, DO NOT EDIT\npackage capstone;\n\npublic class %sconst {\n",
'footer': "}",
... | <commit_before><commit_msg>Add a script to generate constant for binding<commit_after> | import sys, re
INCL_DIR = '../include'
include = [
('/arm.h', 'ARM_'),
('/arm64.h', 'ARM64_'),
('/x86.h', 'X86_'),
('/mips.h', 'MIPS_'),
]
template = {
'java': {
'header': "// AUTOGENRATED FILE, DO NOT EDIT\npackage capstone;\n\npublic class %sconst {\n",
'footer': "}",
... | Add a script to generate constant for bindingimport sys, re
INCL_DIR = '../include'
include = [
('/arm.h', 'ARM_'),
('/arm64.h', 'ARM64_'),
('/x86.h', 'X86_'),
('/mips.h', 'MIPS_'),
]
template = {
'java': {
'header': "// AUTOGENRATED FILE, DO NOT EDIT\npackage capstone;\n\npublic clas... | <commit_before><commit_msg>Add a script to generate constant for binding<commit_after>import sys, re
INCL_DIR = '../include'
include = [
('/arm.h', 'ARM_'),
('/arm64.h', 'ARM64_'),
('/x86.h', 'X86_'),
('/mips.h', 'MIPS_'),
]
template = {
'java': {
'header': "// AUTOGENRATED FILE, DO N... | |
0eac761535f959a07acb0e611b415e2cbccd9a97 | tests/sentry/web/frontend/test_organization_settings.py | tests/sentry/web/frontend/test_organization_settings.py | from __future__ import absolute_import
from django.core.urlresolvers import reverse
from sentry.models import Organization
from sentry.testutils import TestCase
class OrganizationSettingsTest(TestCase):
def test_renders_with_context(self):
organization = self.create_organization(name='foo', owner=self.u... | Add test for organization settings | Add test for organization settings
| Python | bsd-3-clause | songyi199111/sentry,JackDanger/sentry,mvaled/sentry,jokey2k/sentry,felixbuenemann/sentry,drcapulet/sentry,Kryz/sentry,fotinakis/sentry,kevinastone/sentry,gencer/sentry,looker/sentry,kevinlondon/sentry,boneyao/sentry,mvaled/sentry,jean/sentry,llonchj/sentry,BuildingLink/sentry,gencer/sentry,korealerts1/sentry,vperron/se... | Add test for organization settings | from __future__ import absolute_import
from django.core.urlresolvers import reverse
from sentry.models import Organization
from sentry.testutils import TestCase
class OrganizationSettingsTest(TestCase):
def test_renders_with_context(self):
organization = self.create_organization(name='foo', owner=self.u... | <commit_before><commit_msg>Add test for organization settings<commit_after> | from __future__ import absolute_import
from django.core.urlresolvers import reverse
from sentry.models import Organization
from sentry.testutils import TestCase
class OrganizationSettingsTest(TestCase):
def test_renders_with_context(self):
organization = self.create_organization(name='foo', owner=self.u... | Add test for organization settingsfrom __future__ import absolute_import
from django.core.urlresolvers import reverse
from sentry.models import Organization
from sentry.testutils import TestCase
class OrganizationSettingsTest(TestCase):
def test_renders_with_context(self):
organization = self.create_org... | <commit_before><commit_msg>Add test for organization settings<commit_after>from __future__ import absolute_import
from django.core.urlresolvers import reverse
from sentry.models import Organization
from sentry.testutils import TestCase
class OrganizationSettingsTest(TestCase):
def test_renders_with_context(self... | |
e29962ede32106984b41454b7cbfbf0217bef7fe | tools/rebuild_all.py | tools/rebuild_all.py | import os
import shutil
import subprocess
import sys
from contextlib import contextmanager
@contextmanager
def ignore_no_file():
try:
yield
except FileNotFoundError:
pass
os.chdir(os.path.dirname(os.path.dirname(os.path.realpath(__file__))))
with ignore_no_file():
print('Removing @Pythi... | Add a script to simplify building releases | Add a script to simplify building releases
| Python | mit | overfl0/Pythia,overfl0/Pythia,overfl0/Pythia,overfl0/Pythia | Add a script to simplify building releases | import os
import shutil
import subprocess
import sys
from contextlib import contextmanager
@contextmanager
def ignore_no_file():
try:
yield
except FileNotFoundError:
pass
os.chdir(os.path.dirname(os.path.dirname(os.path.realpath(__file__))))
with ignore_no_file():
print('Removing @Pythi... | <commit_before><commit_msg>Add a script to simplify building releases<commit_after> | import os
import shutil
import subprocess
import sys
from contextlib import contextmanager
@contextmanager
def ignore_no_file():
try:
yield
except FileNotFoundError:
pass
os.chdir(os.path.dirname(os.path.dirname(os.path.realpath(__file__))))
with ignore_no_file():
print('Removing @Pythi... | Add a script to simplify building releasesimport os
import shutil
import subprocess
import sys
from contextlib import contextmanager
@contextmanager
def ignore_no_file():
try:
yield
except FileNotFoundError:
pass
os.chdir(os.path.dirname(os.path.dirname(os.path.realpath(__file__))))
with ig... | <commit_before><commit_msg>Add a script to simplify building releases<commit_after>import os
import shutil
import subprocess
import sys
from contextlib import contextmanager
@contextmanager
def ignore_no_file():
try:
yield
except FileNotFoundError:
pass
os.chdir(os.path.dirname(os.path.dirna... | |
1621e19c5bf6deacc54a017115c47175a6e52765 | tests/test_api.py | tests/test_api.py | """Tests the isort API module"""
import pytest
from isort import api, exceptions
def test_sort_file_invalid_syntax(tmpdir) -> None:
"""Test to ensure file encoding is respected"""
tmp_file = tmpdir.join(f"test_bad_syntax.py")
tmp_file.write_text("""print('mismathing quotes")""", "utf8")
with pytest.w... | Add initial testing module for api.py | Add initial testing module for api.py
| Python | mit | PyCQA/isort,PyCQA/isort | Add initial testing module for api.py | """Tests the isort API module"""
import pytest
from isort import api, exceptions
def test_sort_file_invalid_syntax(tmpdir) -> None:
"""Test to ensure file encoding is respected"""
tmp_file = tmpdir.join(f"test_bad_syntax.py")
tmp_file.write_text("""print('mismathing quotes")""", "utf8")
with pytest.w... | <commit_before><commit_msg>Add initial testing module for api.py<commit_after> | """Tests the isort API module"""
import pytest
from isort import api, exceptions
def test_sort_file_invalid_syntax(tmpdir) -> None:
"""Test to ensure file encoding is respected"""
tmp_file = tmpdir.join(f"test_bad_syntax.py")
tmp_file.write_text("""print('mismathing quotes")""", "utf8")
with pytest.w... | Add initial testing module for api.py"""Tests the isort API module"""
import pytest
from isort import api, exceptions
def test_sort_file_invalid_syntax(tmpdir) -> None:
"""Test to ensure file encoding is respected"""
tmp_file = tmpdir.join(f"test_bad_syntax.py")
tmp_file.write_text("""print('mismathing q... | <commit_before><commit_msg>Add initial testing module for api.py<commit_after>"""Tests the isort API module"""
import pytest
from isort import api, exceptions
def test_sort_file_invalid_syntax(tmpdir) -> None:
"""Test to ensure file encoding is respected"""
tmp_file = tmpdir.join(f"test_bad_syntax.py")
t... | |
51ab60836c58c7d0d10c637482919a9acbc31efe | scripts/create_database.py | scripts/create_database.py | #!/usr/bin/env python
"""Create the initial database structure.
Existing tables will be ignored, and those not existing will be created.
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
import click
from byceps.database import db
from byceps.util.system import get_con... | Add script to create initial database structure | Add script to create initial database structure
| Python | bsd-3-clause | homeworkprod/byceps,m-ober/byceps,homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps | Add script to create initial database structure | #!/usr/bin/env python
"""Create the initial database structure.
Existing tables will be ignored, and those not existing will be created.
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
import click
from byceps.database import db
from byceps.util.system import get_con... | <commit_before><commit_msg>Add script to create initial database structure<commit_after> | #!/usr/bin/env python
"""Create the initial database structure.
Existing tables will be ignored, and those not existing will be created.
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
import click
from byceps.database import db
from byceps.util.system import get_con... | Add script to create initial database structure#!/usr/bin/env python
"""Create the initial database structure.
Existing tables will be ignored, and those not existing will be created.
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
import click
from byceps.database i... | <commit_before><commit_msg>Add script to create initial database structure<commit_after>#!/usr/bin/env python
"""Create the initial database structure.
Existing tables will be ignored, and those not existing will be created.
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
... | |
7855f7ca4f7ea96bf8bf85a5d3fb49a666ddc6cc | src/python/QueryParuetFile.py | src/python/QueryParuetFile.py | # Finds the names of people who like pandas from a parquet file
# consisting of name & favouriteAnimal.
# For input you can use the result of MakeParquetFile
from pyspark import SparkContext
from pyspark.sql import SQLContext
import json
import sys
if __name__ == "__main__":
if len(sys.argv) != 4:
print "E... | Add a python parquet file query example | Add a python parquet file query example
| Python | mit | zaxliu/learning-spark,bhagatsingh/learning-spark,bhagatsingh/learning-spark,XiaoqingWang/learning-spark,kpraveen420/learning-spark,holdenk/learning-spark-examples,zaxliu/learning-spark,diogoaurelio/learning-spark,ellis429/learning-spark-examples,huixiang/learning-spark,concerned3rdparty/learning-spark,zaxliu/learning-s... | Add a python parquet file query example | # Finds the names of people who like pandas from a parquet file
# consisting of name & favouriteAnimal.
# For input you can use the result of MakeParquetFile
from pyspark import SparkContext
from pyspark.sql import SQLContext
import json
import sys
if __name__ == "__main__":
if len(sys.argv) != 4:
print "E... | <commit_before><commit_msg>Add a python parquet file query example<commit_after> | # Finds the names of people who like pandas from a parquet file
# consisting of name & favouriteAnimal.
# For input you can use the result of MakeParquetFile
from pyspark import SparkContext
from pyspark.sql import SQLContext
import json
import sys
if __name__ == "__main__":
if len(sys.argv) != 4:
print "E... | Add a python parquet file query example# Finds the names of people who like pandas from a parquet file
# consisting of name & favouriteAnimal.
# For input you can use the result of MakeParquetFile
from pyspark import SparkContext
from pyspark.sql import SQLContext
import json
import sys
if __name__ == "__main__":
... | <commit_before><commit_msg>Add a python parquet file query example<commit_after># Finds the names of people who like pandas from a parquet file
# consisting of name & favouriteAnimal.
# For input you can use the result of MakeParquetFile
from pyspark import SparkContext
from pyspark.sql import SQLContext
import json
im... | |
65eaffabe9927d17847a6a8ce9f39e92e130ab6b | Challenge2.py | Challenge2.py | #Program to find the n largest lines in a file
import operator
import sys
#Read the file from terminal
test_cases = open(sys_argv[1],'r')
#Remove white spaces and empty lines and make a list for all the test in test_cases
tests = [test_cases.strip() for test in test_cases if not test == '\n']
#The first element in t... | Read file and print n largest lines | Read file and print n largest lines
| Python | mit | mailpraveens/Python-Experiments | Read file and print n largest lines | #Program to find the n largest lines in a file
import operator
import sys
#Read the file from terminal
test_cases = open(sys_argv[1],'r')
#Remove white spaces and empty lines and make a list for all the test in test_cases
tests = [test_cases.strip() for test in test_cases if not test == '\n']
#The first element in t... | <commit_before><commit_msg>Read file and print n largest lines<commit_after> | #Program to find the n largest lines in a file
import operator
import sys
#Read the file from terminal
test_cases = open(sys_argv[1],'r')
#Remove white spaces and empty lines and make a list for all the test in test_cases
tests = [test_cases.strip() for test in test_cases if not test == '\n']
#The first element in t... | Read file and print n largest lines#Program to find the n largest lines in a file
import operator
import sys
#Read the file from terminal
test_cases = open(sys_argv[1],'r')
#Remove white spaces and empty lines and make a list for all the test in test_cases
tests = [test_cases.strip() for test in test_cases if not tes... | <commit_before><commit_msg>Read file and print n largest lines<commit_after>#Program to find the n largest lines in a file
import operator
import sys
#Read the file from terminal
test_cases = open(sys_argv[1],'r')
#Remove white spaces and empty lines and make a list for all the test in test_cases
tests = [test_cases.... | |
09eea0de71ac2f2c4f9cee040a14874638dcb097 | bin/neighbourhoods_categories.py | bin/neighbourhoods_categories.py | """neighbourhoods_categories.py
Find the tracts where each category is over-represented or all cities in the
dataset.
"""
import csv
import marble as mb
#
# Import a list of MSA
#
msa = {}
with open('data/names/msa.csv', 'r') as source:
reader = csv.reader(source, delimiter='\t')
reader.next()
for rows i... | Add script to find areal units where each category is represented (and test marble) | Add script to find areal units where each category is represented (and test marble)
| Python | bsd-3-clause | rlouf/patterns-of-segregation | Add script to find areal units where each category is represented (and test marble) | """neighbourhoods_categories.py
Find the tracts where each category is over-represented or all cities in the
dataset.
"""
import csv
import marble as mb
#
# Import a list of MSA
#
msa = {}
with open('data/names/msa.csv', 'r') as source:
reader = csv.reader(source, delimiter='\t')
reader.next()
for rows i... | <commit_before><commit_msg>Add script to find areal units where each category is represented (and test marble)<commit_after> | """neighbourhoods_categories.py
Find the tracts where each category is over-represented or all cities in the
dataset.
"""
import csv
import marble as mb
#
# Import a list of MSA
#
msa = {}
with open('data/names/msa.csv', 'r') as source:
reader = csv.reader(source, delimiter='\t')
reader.next()
for rows i... | Add script to find areal units where each category is represented (and test marble)"""neighbourhoods_categories.py
Find the tracts where each category is over-represented or all cities in the
dataset.
"""
import csv
import marble as mb
#
# Import a list of MSA
#
msa = {}
with open('data/names/msa.csv', 'r') as sourc... | <commit_before><commit_msg>Add script to find areal units where each category is represented (and test marble)<commit_after>"""neighbourhoods_categories.py
Find the tracts where each category is over-represented or all cities in the
dataset.
"""
import csv
import marble as mb
#
# Import a list of MSA
#
msa = {}
with... | |
3777358287dd4f2ec485d7d53c29d83bf6f56974 | migrations/versions/0368_move_orgs_to_nhs_branding_.py | migrations/versions/0368_move_orgs_to_nhs_branding_.py | """
Revision ID: 0368_move_orgs_to_nhs_branding
Revises: 0367_add_reach
Create Date: 2022-04-12 18:22:12.069016
"""
from alembic import op
revision = '0368_move_orgs_to_nhs_branding'
down_revision = '0367_add_reach'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.execute("""... | Move existing nhs orgs without branding onto nhs branding | Move existing nhs orgs without branding onto nhs branding
This is done to make self-service branding easier to implement,
and also because NHS branding makes much more sense for services
in those orgs than GOV.UK branding.
| Python | mit | alphagov/notifications-api,alphagov/notifications-api | Move existing nhs orgs without branding onto nhs branding
This is done to make self-service branding easier to implement,
and also because NHS branding makes much more sense for services
in those orgs than GOV.UK branding. | """
Revision ID: 0368_move_orgs_to_nhs_branding
Revises: 0367_add_reach
Create Date: 2022-04-12 18:22:12.069016
"""
from alembic import op
revision = '0368_move_orgs_to_nhs_branding'
down_revision = '0367_add_reach'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.execute("""... | <commit_before><commit_msg>Move existing nhs orgs without branding onto nhs branding
This is done to make self-service branding easier to implement,
and also because NHS branding makes much more sense for services
in those orgs than GOV.UK branding.<commit_after> | """
Revision ID: 0368_move_orgs_to_nhs_branding
Revises: 0367_add_reach
Create Date: 2022-04-12 18:22:12.069016
"""
from alembic import op
revision = '0368_move_orgs_to_nhs_branding'
down_revision = '0367_add_reach'
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.execute("""... | Move existing nhs orgs without branding onto nhs branding
This is done to make self-service branding easier to implement,
and also because NHS branding makes much more sense for services
in those orgs than GOV.UK branding."""
Revision ID: 0368_move_orgs_to_nhs_branding
Revises: 0367_add_reach
Create Date: 2022-04-12 ... | <commit_before><commit_msg>Move existing nhs orgs without branding onto nhs branding
This is done to make self-service branding easier to implement,
and also because NHS branding makes much more sense for services
in those orgs than GOV.UK branding.<commit_after>"""
Revision ID: 0368_move_orgs_to_nhs_branding
Revises... | |
87e3e6b4930b73563027cb0e125ddd9b9c787d6d | api_tests/requests/views/test_request_action_list.py | api_tests/requests/views/test_request_action_list.py | import pytest
from api.base.settings.defaults import API_BASE
from api_tests.requests.mixins import PreprintRequestTestMixin
@pytest.mark.django_db
class TestPreprintRequestActionList(PreprintRequestTestMixin):
def url(self, request):
return '/{}requests/{}/actions/'.format(API_BASE, request._id)
def... | Add action list permissions tests | Add action list permissions tests
| Python | apache-2.0 | erinspace/osf.io,caseyrollins/osf.io,pattisdr/osf.io,erinspace/osf.io,aaxelb/osf.io,CenterForOpenScience/osf.io,HalcyonChimera/osf.io,CenterForOpenScience/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,felliott/osf.io,Johnetordoff/osf.io,pattisdr/osf.io,brianjgeiger/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,Johnetordoff... | Add action list permissions tests | import pytest
from api.base.settings.defaults import API_BASE
from api_tests.requests.mixins import PreprintRequestTestMixin
@pytest.mark.django_db
class TestPreprintRequestActionList(PreprintRequestTestMixin):
def url(self, request):
return '/{}requests/{}/actions/'.format(API_BASE, request._id)
def... | <commit_before><commit_msg>Add action list permissions tests<commit_after> | import pytest
from api.base.settings.defaults import API_BASE
from api_tests.requests.mixins import PreprintRequestTestMixin
@pytest.mark.django_db
class TestPreprintRequestActionList(PreprintRequestTestMixin):
def url(self, request):
return '/{}requests/{}/actions/'.format(API_BASE, request._id)
def... | Add action list permissions testsimport pytest
from api.base.settings.defaults import API_BASE
from api_tests.requests.mixins import PreprintRequestTestMixin
@pytest.mark.django_db
class TestPreprintRequestActionList(PreprintRequestTestMixin):
def url(self, request):
return '/{}requests/{}/actions/'.forma... | <commit_before><commit_msg>Add action list permissions tests<commit_after>import pytest
from api.base.settings.defaults import API_BASE
from api_tests.requests.mixins import PreprintRequestTestMixin
@pytest.mark.django_db
class TestPreprintRequestActionList(PreprintRequestTestMixin):
def url(self, request):
... | |
8aa9ea4d0c19a0ec5eb0312bc29a0fa00439c335 | tests/formatter/test_xmler.py | tests/formatter/test_xmler.py | import unittest, argparse
from echolalia.formatter.xmler import Formatter
class XmlerTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 100)]
self.formatter = Formatter()
def test_add_args(self):
... | Add tests for formatter xml | Add tests for formatter xml
| Python | mit | eiri/echolalia-prototype | Add tests for formatter xml | import unittest, argparse
from echolalia.formatter.xmler import Formatter
class XmlerTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 100)]
self.formatter = Formatter()
def test_add_args(self):
... | <commit_before><commit_msg>Add tests for formatter xml<commit_after> | import unittest, argparse
from echolalia.formatter.xmler import Formatter
class XmlerTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 100)]
self.formatter = Formatter()
def test_add_args(self):
... | Add tests for formatter xmlimport unittest, argparse
from echolalia.formatter.xmler import Formatter
class XmlerTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 100)]
self.formatter = Formatter()
d... | <commit_before><commit_msg>Add tests for formatter xml<commit_after>import unittest, argparse
from echolalia.formatter.xmler import Formatter
class XmlerTestCase(unittest.TestCase):
def setUp(self):
self.parser = argparse.ArgumentParser()
self.data = [{'char': chr(i), 'order': i - 96} for i in xrange(97, 10... | |
96d556edf9c736f6c7e9deb70753762230933bc0 | kargtom/twodim/LongestChain/longestChain_001.py | kargtom/twodim/LongestChain/longestChain_001.py | def longestChain(words):
if words == []:
return 0
lendict = {}
lenlist = []
# build the dictory where
# the key is the length of a word,
# and the value is the set of words with the length
for word in words:
l = len(word)
if len(word) in lendict:
lendict[... | Create the bottom-up version of longestPath | Create the bottom-up version of longestPath | Python | mit | Chasego/codi,cc13ny/algo,cc13ny/Allin,Chasego/cod,Chasego/codirit,Chasego/cod,cc13ny/algo,Chasego/cod,Chasego/codi,Chasego/codirit,cc13ny/Allin,Chasego/cod,Chasego/cod,cc13ny/Allin,cc13ny/algo,Chasego/codi,Chasego/codi,cc13ny/Allin,cc13ny/algo,Chasego/codirit,Chasego/codirit,Chasego/codi,cc13ny/algo,cc13ny/Allin,Chaseg... | Create the bottom-up version of longestPath | def longestChain(words):
if words == []:
return 0
lendict = {}
lenlist = []
# build the dictory where
# the key is the length of a word,
# and the value is the set of words with the length
for word in words:
l = len(word)
if len(word) in lendict:
lendict[... | <commit_before><commit_msg>Create the bottom-up version of longestPath<commit_after> | def longestChain(words):
if words == []:
return 0
lendict = {}
lenlist = []
# build the dictory where
# the key is the length of a word,
# and the value is the set of words with the length
for word in words:
l = len(word)
if len(word) in lendict:
lendict[... | Create the bottom-up version of longestPathdef longestChain(words):
if words == []:
return 0
lendict = {}
lenlist = []
# build the dictory where
# the key is the length of a word,
# and the value is the set of words with the length
for word in words:
l = len(word)
if... | <commit_before><commit_msg>Create the bottom-up version of longestPath<commit_after>def longestChain(words):
if words == []:
return 0
lendict = {}
lenlist = []
# build the dictory where
# the key is the length of a word,
# and the value is the set of words with the length
for word i... | |
f3c4cb70d1f9fab3b7cb497faa2cbe2350e2f116 | axelrod/tests/unit/test_deterministic_cache.py | axelrod/tests/unit/test_deterministic_cache.py | import unittest
class TestDeterministicCache(unittest.TestCase):
def test_init(self):
pass
def test_setitem(self):
pass
def test_save(self):
pass
def test_load(self):
pass
| Add stubs for deterministic cache tests | Add stubs for deterministic cache tests
| Python | mit | marcharper/Axelrod,ranjinidas/Axelrod,marcharper/Axelrod,ranjinidas/Axelrod | Add stubs for deterministic cache tests | import unittest
class TestDeterministicCache(unittest.TestCase):
def test_init(self):
pass
def test_setitem(self):
pass
def test_save(self):
pass
def test_load(self):
pass
| <commit_before><commit_msg>Add stubs for deterministic cache tests<commit_after> | import unittest
class TestDeterministicCache(unittest.TestCase):
def test_init(self):
pass
def test_setitem(self):
pass
def test_save(self):
pass
def test_load(self):
pass
| Add stubs for deterministic cache testsimport unittest
class TestDeterministicCache(unittest.TestCase):
def test_init(self):
pass
def test_setitem(self):
pass
def test_save(self):
pass
def test_load(self):
pass
| <commit_before><commit_msg>Add stubs for deterministic cache tests<commit_after>import unittest
class TestDeterministicCache(unittest.TestCase):
def test_init(self):
pass
def test_setitem(self):
pass
def test_save(self):
pass
def test_load(self):
pass
| |
cb2c7126e1682c38a4efc92a1b89c04498d1926f | clsimplex.py | clsimplex.py | import pyopencl as cl
import numpy
import sys
import datetime
class NoiseGenerator(object):
def __init__(self, block_dim=None):
self.ctx = cl.create_some_context()
self.queue = cl.CommandQueue(self.ctx)
if block_dim: self.block_dim = block_dim
else: self.block_dim = 256, 256, 256
... | Add python host code using pyopencl | Add python host code using pyopencl
| Python | mit | jakogut/clsimplex,jakogut/clsimplex | Add python host code using pyopencl | import pyopencl as cl
import numpy
import sys
import datetime
class NoiseGenerator(object):
def __init__(self, block_dim=None):
self.ctx = cl.create_some_context()
self.queue = cl.CommandQueue(self.ctx)
if block_dim: self.block_dim = block_dim
else: self.block_dim = 256, 256, 256
... | <commit_before><commit_msg>Add python host code using pyopencl<commit_after> | import pyopencl as cl
import numpy
import sys
import datetime
class NoiseGenerator(object):
def __init__(self, block_dim=None):
self.ctx = cl.create_some_context()
self.queue = cl.CommandQueue(self.ctx)
if block_dim: self.block_dim = block_dim
else: self.block_dim = 256, 256, 256
... | Add python host code using pyopenclimport pyopencl as cl
import numpy
import sys
import datetime
class NoiseGenerator(object):
def __init__(self, block_dim=None):
self.ctx = cl.create_some_context()
self.queue = cl.CommandQueue(self.ctx)
if block_dim: self.block_dim = block_dim
els... | <commit_before><commit_msg>Add python host code using pyopencl<commit_after>import pyopencl as cl
import numpy
import sys
import datetime
class NoiseGenerator(object):
def __init__(self, block_dim=None):
self.ctx = cl.create_some_context()
self.queue = cl.CommandQueue(self.ctx)
if block_di... | |
16b6a285e933313625da0b3c71455a3506081d0b | qthread.py | qthread.py | import sys
from queue import Queue
from PyQt4.QtCore import *
from PyQt4.QtGui import *
# The new Stream Object which replaces the default stream associated with sys.stdout
# This object just puts data in a queue!
class WriteStream(object):
def __init__(self,queue):
self.queue = queue
def write(self, ... | Copy QThread code from stackoverflow | Copy QThread code from stackoverflow
| Python | bsd-3-clause | torebutlin/cued_datalogger | Copy QThread code from stackoverflow | import sys
from queue import Queue
from PyQt4.QtCore import *
from PyQt4.QtGui import *
# The new Stream Object which replaces the default stream associated with sys.stdout
# This object just puts data in a queue!
class WriteStream(object):
def __init__(self,queue):
self.queue = queue
def write(self, ... | <commit_before><commit_msg>Copy QThread code from stackoverflow<commit_after> | import sys
from queue import Queue
from PyQt4.QtCore import *
from PyQt4.QtGui import *
# The new Stream Object which replaces the default stream associated with sys.stdout
# This object just puts data in a queue!
class WriteStream(object):
def __init__(self,queue):
self.queue = queue
def write(self, ... | Copy QThread code from stackoverflowimport sys
from queue import Queue
from PyQt4.QtCore import *
from PyQt4.QtGui import *
# The new Stream Object which replaces the default stream associated with sys.stdout
# This object just puts data in a queue!
class WriteStream(object):
def __init__(self,queue):
self... | <commit_before><commit_msg>Copy QThread code from stackoverflow<commit_after>import sys
from queue import Queue
from PyQt4.QtCore import *
from PyQt4.QtGui import *
# The new Stream Object which replaces the default stream associated with sys.stdout
# This object just puts data in a queue!
class WriteStream(object):
... | |
15c799f29652d85fd6957ac0166ffe8590680f33 | froide/account/migrations/0027_auto_20210412_1518.py | froide/account/migrations/0027_auto_20210412_1518.py | # Generated by Django 3.1.8 on 2021-04-12 13:18
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('account', '0026_auto_20201221_1953'),
]
operations = [
migrations.AddField(
model_name='application',
name='algorith... | Add migration for oauth application | Add migration for oauth application | Python | mit | fin/froide,fin/froide,fin/froide,fin/froide | Add migration for oauth application | # Generated by Django 3.1.8 on 2021-04-12 13:18
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('account', '0026_auto_20201221_1953'),
]
operations = [
migrations.AddField(
model_name='application',
name='algorith... | <commit_before><commit_msg>Add migration for oauth application<commit_after> | # Generated by Django 3.1.8 on 2021-04-12 13:18
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('account', '0026_auto_20201221_1953'),
]
operations = [
migrations.AddField(
model_name='application',
name='algorith... | Add migration for oauth application# Generated by Django 3.1.8 on 2021-04-12 13:18
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('account', '0026_auto_20201221_1953'),
]
operations = [
migrations.AddField(
model_name='appli... | <commit_before><commit_msg>Add migration for oauth application<commit_after># Generated by Django 3.1.8 on 2021-04-12 13:18
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('account', '0026_auto_20201221_1953'),
]
operations = [
migration... | |
173aba72cad0c6c3602b2ae4e1b8bd4e5773bd3b | pyservice/context.py | pyservice/context.py | """
RequestContext stores state relevant to the current request, as well as
keeping track of the plugin execution order and providing a simple method
`advance` for calling the next plugin in the chain.
"""
import collections
class Container(collections.defaultdict):
DEFAULT_FACTORY = lambda: None
def __init_... | Add building blocks Container, Context | Add building blocks Container, Context
These will be used for request/response dicts, as well as
containers for plugins to pass functions and values to the
operation function or other plugins.
| Python | mit | numberoverzero/pyservice | Add building blocks Container, Context
These will be used for request/response dicts, as well as
containers for plugins to pass functions and values to the
operation function or other plugins. | """
RequestContext stores state relevant to the current request, as well as
keeping track of the plugin execution order and providing a simple method
`advance` for calling the next plugin in the chain.
"""
import collections
class Container(collections.defaultdict):
DEFAULT_FACTORY = lambda: None
def __init_... | <commit_before><commit_msg>Add building blocks Container, Context
These will be used for request/response dicts, as well as
containers for plugins to pass functions and values to the
operation function or other plugins.<commit_after> | """
RequestContext stores state relevant to the current request, as well as
keeping track of the plugin execution order and providing a simple method
`advance` for calling the next plugin in the chain.
"""
import collections
class Container(collections.defaultdict):
DEFAULT_FACTORY = lambda: None
def __init_... | Add building blocks Container, Context
These will be used for request/response dicts, as well as
containers for plugins to pass functions and values to the
operation function or other plugins."""
RequestContext stores state relevant to the current request, as well as
keeping track of the plugin execution order and pro... | <commit_before><commit_msg>Add building blocks Container, Context
These will be used for request/response dicts, as well as
containers for plugins to pass functions and values to the
operation function or other plugins.<commit_after>"""
RequestContext stores state relevant to the current request, as well as
keeping tr... | |
b98fe743844973b2b03dd128fc906cfd4cfa9728 | solutions/uri/1036/1036.py | solutions/uri/1036/1036.py | import math
a, b, c = map(float, input().split())
delta = math.pow(b, 2) - 4 * a * c
if delta >= 0 and a != 0:
print(f'R1 = {((b * -1) + math.sqrt(delta)) / (2 * a):.5f}')
print(f'R2 = {((b * -1) - math.sqrt(delta)) / (2 * a):.5f}')
else:
print('Impossivel calcular')
| Solve Bhaskara's Formula in python | Solve Bhaskara's Formula in python
| Python | mit | deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playground,deniscostadsc/playgr... | Solve Bhaskara's Formula in python | import math
a, b, c = map(float, input().split())
delta = math.pow(b, 2) - 4 * a * c
if delta >= 0 and a != 0:
print(f'R1 = {((b * -1) + math.sqrt(delta)) / (2 * a):.5f}')
print(f'R2 = {((b * -1) - math.sqrt(delta)) / (2 * a):.5f}')
else:
print('Impossivel calcular')
| <commit_before><commit_msg>Solve Bhaskara's Formula in python<commit_after> | import math
a, b, c = map(float, input().split())
delta = math.pow(b, 2) - 4 * a * c
if delta >= 0 and a != 0:
print(f'R1 = {((b * -1) + math.sqrt(delta)) / (2 * a):.5f}')
print(f'R2 = {((b * -1) - math.sqrt(delta)) / (2 * a):.5f}')
else:
print('Impossivel calcular')
| Solve Bhaskara's Formula in pythonimport math
a, b, c = map(float, input().split())
delta = math.pow(b, 2) - 4 * a * c
if delta >= 0 and a != 0:
print(f'R1 = {((b * -1) + math.sqrt(delta)) / (2 * a):.5f}')
print(f'R2 = {((b * -1) - math.sqrt(delta)) / (2 * a):.5f}')
else:
print('Impossivel calcular')
| <commit_before><commit_msg>Solve Bhaskara's Formula in python<commit_after>import math
a, b, c = map(float, input().split())
delta = math.pow(b, 2) - 4 * a * c
if delta >= 0 and a != 0:
print(f'R1 = {((b * -1) + math.sqrt(delta)) / (2 * a):.5f}')
print(f'R2 = {((b * -1) - math.sqrt(delta)) / (2 * a):.5f}')
e... | |
dba4c700d0fbc68e853718d27471b4dd7ed21253 | scripts/set_product_info.py | scripts/set_product_info.py | #!/usr/bin/python
#
# Copyright (c) SAS Institute, Inc.
#
"""
Script to set product info on a group.
"""
import sys
import epdb
sys.excepthook = epdb.excepthook()
import json
from conary import trove
from conary import conarycfg
from conary import conaryclient
from conary.conaryclient import cmdline
def setProduct... | Add metadata example code APPENG-3048 | Add metadata example code APPENG-3048
| Python | apache-2.0 | sassoftware/mirrorball,sassoftware/mirrorball | Add metadata example code APPENG-3048 | #!/usr/bin/python
#
# Copyright (c) SAS Institute, Inc.
#
"""
Script to set product info on a group.
"""
import sys
import epdb
sys.excepthook = epdb.excepthook()
import json
from conary import trove
from conary import conarycfg
from conary import conaryclient
from conary.conaryclient import cmdline
def setProduct... | <commit_before><commit_msg>Add metadata example code APPENG-3048<commit_after> | #!/usr/bin/python
#
# Copyright (c) SAS Institute, Inc.
#
"""
Script to set product info on a group.
"""
import sys
import epdb
sys.excepthook = epdb.excepthook()
import json
from conary import trove
from conary import conarycfg
from conary import conaryclient
from conary.conaryclient import cmdline
def setProduct... | Add metadata example code APPENG-3048#!/usr/bin/python
#
# Copyright (c) SAS Institute, Inc.
#
"""
Script to set product info on a group.
"""
import sys
import epdb
sys.excepthook = epdb.excepthook()
import json
from conary import trove
from conary import conarycfg
from conary import conaryclient
from conary.conary... | <commit_before><commit_msg>Add metadata example code APPENG-3048<commit_after>#!/usr/bin/python
#
# Copyright (c) SAS Institute, Inc.
#
"""
Script to set product info on a group.
"""
import sys
import epdb
sys.excepthook = epdb.excepthook()
import json
from conary import trove
from conary import conarycfg
from cona... | |
1c856409149ce1523dbe080717465e1985f778f7 | scripts/find-skeletons-with-no-treenodes.py | scripts/find-skeletons-with-no-treenodes.py | #!/usr/bin/python
# This script checks your database for some common
# inconsistencies or errors that may have arisen from
# past bugs.
# You may need to install psycopg2, e.g. with:
# sudo apt-get install python-psycopg2
# Requires the file .catmaid-db to be present in your
# home directory, with the following fo... | Add a script to find "empty" skeletons | Add a script to find "empty" skeletons
| Python | agpl-3.0 | htem/CATMAID,fzadow/CATMAID,fzadow/CATMAID,fzadow/CATMAID,htem/CATMAID,fzadow/CATMAID,htem/CATMAID,htem/CATMAID | Add a script to find "empty" skeletons | #!/usr/bin/python
# This script checks your database for some common
# inconsistencies or errors that may have arisen from
# past bugs.
# You may need to install psycopg2, e.g. with:
# sudo apt-get install python-psycopg2
# Requires the file .catmaid-db to be present in your
# home directory, with the following fo... | <commit_before><commit_msg>Add a script to find "empty" skeletons<commit_after> | #!/usr/bin/python
# This script checks your database for some common
# inconsistencies or errors that may have arisen from
# past bugs.
# You may need to install psycopg2, e.g. with:
# sudo apt-get install python-psycopg2
# Requires the file .catmaid-db to be present in your
# home directory, with the following fo... | Add a script to find "empty" skeletons#!/usr/bin/python
# This script checks your database for some common
# inconsistencies or errors that may have arisen from
# past bugs.
# You may need to install psycopg2, e.g. with:
# sudo apt-get install python-psycopg2
# Requires the file .catmaid-db to be present in your
#... | <commit_before><commit_msg>Add a script to find "empty" skeletons<commit_after>#!/usr/bin/python
# This script checks your database for some common
# inconsistencies or errors that may have arisen from
# past bugs.
# You may need to install psycopg2, e.g. with:
# sudo apt-get install python-psycopg2
# Requires the... | |
6276e1ad6e74874cdc2ab8e84608d02edbf6b587 | unify/framework/jasyscript.py | unify/framework/jasyscript.py | # Unify project builder
# Copyright 2012 Sebastian Fastner, Mainz, Germany
import webbrowser
@task("Open help in browser")
def help():
# Clearing cache
webbrowser.open("http://unify-training.com/")
| Add initial unify bould script | Add initial unify bould script
| Python | mit | unify/unify,unify/unify,unify/unify,unify/unify,unify/unify,unify/unify | Add initial unify bould script | # Unify project builder
# Copyright 2012 Sebastian Fastner, Mainz, Germany
import webbrowser
@task("Open help in browser")
def help():
# Clearing cache
webbrowser.open("http://unify-training.com/")
| <commit_before><commit_msg>Add initial unify bould script<commit_after> | # Unify project builder
# Copyright 2012 Sebastian Fastner, Mainz, Germany
import webbrowser
@task("Open help in browser")
def help():
# Clearing cache
webbrowser.open("http://unify-training.com/")
| Add initial unify bould script# Unify project builder
# Copyright 2012 Sebastian Fastner, Mainz, Germany
import webbrowser
@task("Open help in browser")
def help():
# Clearing cache
webbrowser.open("http://unify-training.com/")
| <commit_before><commit_msg>Add initial unify bould script<commit_after># Unify project builder
# Copyright 2012 Sebastian Fastner, Mainz, Germany
import webbrowser
@task("Open help in browser")
def help():
# Clearing cache
webbrowser.open("http://unify-training.com/")
| |
be1430571bb043bc397e73873a61908d40a2bda1 | scratchpad/nano-i2c-test.py | scratchpad/nano-i2c-test.py | #!/usr/bin/env python3
import smbus
import time
# for RPI version 1, use “bus = smbus.SMBus(0)”
bus = smbus.SMBus(1)
# This is the address we setup in the Arduino Program
address = 0x04
cmd_read_analog = 1
while True:
number = bus.read_word_data(address, cmd_read_analog)
print("analog value =", "{0:4X}".f... | Read value from nano via i2c | Read value from nano via i2c
| Python | mit | gizmo-cda/g2x,gizmo-cda/g2x,gizmo-cda/g2x,thelonious/g2x,gizmo-cda/g2x,thelonious/g2x | Read value from nano via i2c | #!/usr/bin/env python3
import smbus
import time
# for RPI version 1, use “bus = smbus.SMBus(0)”
bus = smbus.SMBus(1)
# This is the address we setup in the Arduino Program
address = 0x04
cmd_read_analog = 1
while True:
number = bus.read_word_data(address, cmd_read_analog)
print("analog value =", "{0:4X}".f... | <commit_before><commit_msg>Read value from nano via i2c<commit_after> | #!/usr/bin/env python3
import smbus
import time
# for RPI version 1, use “bus = smbus.SMBus(0)”
bus = smbus.SMBus(1)
# This is the address we setup in the Arduino Program
address = 0x04
cmd_read_analog = 1
while True:
number = bus.read_word_data(address, cmd_read_analog)
print("analog value =", "{0:4X}".f... | Read value from nano via i2c#!/usr/bin/env python3
import smbus
import time
# for RPI version 1, use “bus = smbus.SMBus(0)”
bus = smbus.SMBus(1)
# This is the address we setup in the Arduino Program
address = 0x04
cmd_read_analog = 1
while True:
number = bus.read_word_data(address, cmd_read_analog)
print(... | <commit_before><commit_msg>Read value from nano via i2c<commit_after>#!/usr/bin/env python3
import smbus
import time
# for RPI version 1, use “bus = smbus.SMBus(0)”
bus = smbus.SMBus(1)
# This is the address we setup in the Arduino Program
address = 0x04
cmd_read_analog = 1
while True:
number = bus.read_word_d... | |
89f4f16e3ec8b6018c9ef487ae4059271e1829df | tests/functional/test_new_resolver.py | tests/functional/test_new_resolver.py | import json
from tests.lib import create_basic_wheel_for_package
def assert_installed(script, **kwargs):
ret = script.pip('list', '--format=json')
installed = set(
(val['name'], val['version'])
for val in json.loads(ret.stdout)
)
assert set(kwargs.items()) <= installed
def te... | Add some functional tests for the new resolver | Add some functional tests for the new resolver
| Python | mit | pfmoore/pip,pypa/pip,sbidoul/pip,pypa/pip,pfmoore/pip,pradyunsg/pip,sbidoul/pip,pradyunsg/pip | Add some functional tests for the new resolver | import json
from tests.lib import create_basic_wheel_for_package
def assert_installed(script, **kwargs):
ret = script.pip('list', '--format=json')
installed = set(
(val['name'], val['version'])
for val in json.loads(ret.stdout)
)
assert set(kwargs.items()) <= installed
def te... | <commit_before><commit_msg>Add some functional tests for the new resolver<commit_after> | import json
from tests.lib import create_basic_wheel_for_package
def assert_installed(script, **kwargs):
ret = script.pip('list', '--format=json')
installed = set(
(val['name'], val['version'])
for val in json.loads(ret.stdout)
)
assert set(kwargs.items()) <= installed
def te... | Add some functional tests for the new resolverimport json
from tests.lib import create_basic_wheel_for_package
def assert_installed(script, **kwargs):
ret = script.pip('list', '--format=json')
installed = set(
(val['name'], val['version'])
for val in json.loads(ret.stdout)
)
assert set... | <commit_before><commit_msg>Add some functional tests for the new resolver<commit_after>import json
from tests.lib import create_basic_wheel_for_package
def assert_installed(script, **kwargs):
ret = script.pip('list', '--format=json')
installed = set(
(val['name'], val['version'])
for val in js... | |
5c45586ed1c90c72620738420d37a444647fe186 | code/supervise_acronym_gene_mentions.py | code/supervise_acronym_gene_mentions.py | #! /usr/bin/env pyton3
import fileinput
import json
# Process the input
with fileinput.input() as input_files:
for line in input_files:
mention = json.loads(line)
mention.is_correct = False
print(json.dumps(mention))
| Add script to supervise acronyms | Add script to supervise acronyms
| Python | apache-2.0 | HazyResearch/dd-genomics,HazyResearch/dd-genomics,amwenger/dd-genomics,rionda/dd-genomics,rionda/dd-genomics,HazyResearch/dd-genomics,HazyResearch/dd-genomics,HazyResearch/dd-genomics,amwenger/dd-genomics,amwenger/dd-genomics | Add script to supervise acronyms | #! /usr/bin/env pyton3
import fileinput
import json
# Process the input
with fileinput.input() as input_files:
for line in input_files:
mention = json.loads(line)
mention.is_correct = False
print(json.dumps(mention))
| <commit_before><commit_msg>Add script to supervise acronyms<commit_after> | #! /usr/bin/env pyton3
import fileinput
import json
# Process the input
with fileinput.input() as input_files:
for line in input_files:
mention = json.loads(line)
mention.is_correct = False
print(json.dumps(mention))
| Add script to supervise acronyms#! /usr/bin/env pyton3
import fileinput
import json
# Process the input
with fileinput.input() as input_files:
for line in input_files:
mention = json.loads(line)
mention.is_correct = False
print(json.dumps(mention))
| <commit_before><commit_msg>Add script to supervise acronyms<commit_after>#! /usr/bin/env pyton3
import fileinput
import json
# Process the input
with fileinput.input() as input_files:
for line in input_files:
mention = json.loads(line)
mention.is_correct = False
print(json.dumps(mention))
... | |
fef62c34be24ac2027d89ffc0c3926b89ad7625d | tools/data/change_window_file_nums.py | tools/data/change_window_file_nums.py | #!/usr/bin/env python
import argparse
import scipy.io as sio
import os
import os.path as osp
import numpy as np
from vdetlib.vdet.dataset import index_det_to_vdet
if __name__ == '__main__':
parser = argparse.ArgumentParser('Change window file numbers.')
parser.add_argument('window_file')
parser.add_argumen... | Add script to change starting index in window file. | Add script to change starting index in window file.
| Python | mit | myfavouritekk/TPN | Add script to change starting index in window file. | #!/usr/bin/env python
import argparse
import scipy.io as sio
import os
import os.path as osp
import numpy as np
from vdetlib.vdet.dataset import index_det_to_vdet
if __name__ == '__main__':
parser = argparse.ArgumentParser('Change window file numbers.')
parser.add_argument('window_file')
parser.add_argumen... | <commit_before><commit_msg>Add script to change starting index in window file.<commit_after> | #!/usr/bin/env python
import argparse
import scipy.io as sio
import os
import os.path as osp
import numpy as np
from vdetlib.vdet.dataset import index_det_to_vdet
if __name__ == '__main__':
parser = argparse.ArgumentParser('Change window file numbers.')
parser.add_argument('window_file')
parser.add_argumen... | Add script to change starting index in window file.#!/usr/bin/env python
import argparse
import scipy.io as sio
import os
import os.path as osp
import numpy as np
from vdetlib.vdet.dataset import index_det_to_vdet
if __name__ == '__main__':
parser = argparse.ArgumentParser('Change window file numbers.')
parser... | <commit_before><commit_msg>Add script to change starting index in window file.<commit_after>#!/usr/bin/env python
import argparse
import scipy.io as sio
import os
import os.path as osp
import numpy as np
from vdetlib.vdet.dataset import index_det_to_vdet
if __name__ == '__main__':
parser = argparse.ArgumentParser(... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.