repo_name stringlengths 7 94 | repo_path stringlengths 4 237 | repo_head_hexsha stringlengths 40 40 | content stringlengths 10 680k | apis stringlengths 2 680k |
|---|---|---|---|---|
mujtahidalam/pandas | pandas/core/indexes/range.py | 526468c8fe6fc5157aaf2fce327c5ab2a3350f49 | from __future__ import annotations
from datetime import timedelta
import operator
from sys import getsizeof
from typing import (
TYPE_CHECKING,
Any,
Callable,
Hashable,
List,
cast,
)
import warnings
import numpy as np
from pandas._libs import index as libindex
from pandas._libs.lib import no_... | [((10265, 10288), 'pandas.util._decorators.doc', 'doc', (['Int64Index.get_loc'], {}), '(Int64Index.get_loc)\n', (10268, 10288), False, 'from pandas.util._decorators import cache_readonly, doc\n'), ((12844, 12868), 'pandas.util._decorators.doc', 'doc', (['Int64Index.__iter__'], {}), '(Int64Index.__iter__)\n', (12847, 12... |
Hasanweight/pytorch-chatbot-master | model.py | 7a3b58af7e5284f1f3f7f7b0aeb3f19d9ee3cbc1 | import torch
import torch.nn as nn
class NeuralNet(nn.Module):
def __init__(self, input_size, hidden_size, num_classes):
super(NeuralNet, self).__init__()
self.l1 = nn.Linear(input_size, hidden_size)
self.l2 = nn.Linear(hidden_size, hidden_size)
self.l3 = nn.Linear(hidden_size, h... | [((187, 221), 'torch.nn.Linear', 'nn.Linear', (['input_size', 'hidden_size'], {}), '(input_size, hidden_size)\n', (196, 221), True, 'import torch.nn as nn\n'), ((241, 276), 'torch.nn.Linear', 'nn.Linear', (['hidden_size', 'hidden_size'], {}), '(hidden_size, hidden_size)\n', (250, 276), True, 'import torch.nn as nn\n'),... |
hover2pi/jwql | jwql/utils/logging_functions.py | 0a97fe618c007883ffbced88ac1cb45a667fcb3c |
""" Logging functions for the ``jwql`` automation platform.
This module provides decorators to log the execution of modules. Log
files are written to the ``logs/`` directory in the ``jwql`` central
storage area, named by module name and timestamp, e.g.
``monitor_filesystem/monitor_filesystem_2018-06-20-15:22:51.log`... | [((2403, 2552), 'logging.basicConfig', 'logging.basicConfig', ([], {'filename': 'log_file', 'format': '"""%(asctime)s %(levelname)s: %(message)s"""', 'datefmt': '"""%m/%d/%Y %H:%M:%S %p"""', 'level': 'logging.INFO'}), "(filename=log_file, format=\n '%(asctime)s %(levelname)s: %(message)s', datefmt=\n '%m/%d/%Y %H... |
takos22/API-1 | api/services/http.py | 261ecd34648d610169caf27b3712256f757b100d | from aiohttp import ClientSession
from typing import Optional
session: Optional[ClientSession] = None
__all__ = (session,)
| [] |
jiaxiaolei/my_snap_demo | bcloud-snap/bcloud-3.9.1/bcloud/hasher.py | 0444077c763e029eb67af7242537cebb3c3d6aa4 |
# Copyright (C) 2014-2015 LiuLang <[email protected]>
# Use of this source code is governed by GPLv3 license that can be found
# in http://www.gnu.org/licenses/gpl-3.0.html
import hashlib
import os
import zlib
CHUNK = 2 ** 20
def crc(path):
_crc = 0
fh = open(path, 'rb')
while True:
chunk... | [((515, 528), 'hashlib.md5', 'hashlib.md5', ([], {}), '()\n', (526, 528), False, 'import hashlib\n'), ((923, 937), 'hashlib.sha1', 'hashlib.sha1', ([], {}), '()\n', (935, 937), False, 'import hashlib\n'), ((1156, 1172), 'hashlib.sha224', 'hashlib.sha224', ([], {}), '()\n', (1170, 1172), False, 'import hashlib\n'), ((13... |
ariesdevil/beansdbadmin | beansdbadmin/core/client.py | 3165087ef57b7511ab84fbc50cf16eb8f54d83cd | #!/usr/bin/python
# encoding: utf-8
'''a rich client
1. for one server (instead of multi like in libmc.Client)
2. encapsulate @, ?, gc ...
use is instead of libmc.Client
'''
import telnetlib
import logging
import libmc
import string
import urllib
import itertools
import warnings
from collections import defaul... | [] |
cuixiongyi/RBE595 | Simulator/Geometry/RectOverlap.py | fc5c6aa6c479eb14186a9168e47724b7b3d06cde | import matplotlib.pyplot
__author__ = 'xiongyi'
line1 = [(200, 100), (200, 400)]
line2 = [(190, 190), (210, 210)]
def overlap():
l1p1x = line1[0][0]
l1p1y = line1[0][1]
l1p2x = line1[1][0]
l1p2y = line1[1][1]
# make sure p1x < p2x
if l1p1x > l1p2x:
tmp = l1p1x
l1p1x = l1p2x
... | [] |
p4l1ly/gino | gino/loader.py | bbe63ed841bf989a0f47b6cae64db85b0b606794 | from sqlalchemy import select
from sqlalchemy.schema import Column
from .declarative import Model
class Loader:
@classmethod
def get(cls, value):
from .crud import Alias
if isinstance(value, Loader):
rv = value
elif isinstance(value, type) and issubclass(value, Model):
... | [] |
Andrew-Tan/e-mission-server | emission/clients/choice/choice.py | 91d59bee86e63d803e401f10f4b6a2502effedda | # Standard imports
import logging
import math
import json
from uuid import UUID
from datetime import datetime, timedelta
import time
# Our imports
from emission.core.get_database import get_trip_db, get_section_db
import emission.analysis.result.carbon as carbon
import emission.core.common as common
import emission.ne... | [] |
ylafon/redbot | lib/formatter/text.py | 87f4edcc8ccda35f556331abd1e76d5e9b79cdd0 | #!/usr/bin/env python
"""
HAR Formatter for REDbot.
"""
__author__ = "Jerome Renard <[email protected]>"
__copyright__ = """\
Copyright (c) 2008-2010 Mark Nottingham
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software")... | [] |
ifurusato/ros | lib/ioe_pot.py | 77b1361e78f68f00ba2d3e3db908bb5ce0f973f5 | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
#
# Copyright 2020-2021 by Murray Altheim. All rights reserved. This file is part
# of the Robot Operating System project, released under the MIT License. Please
# see the LICENSE file included as part of this package.
#
# author: Murray Altheim
# created: 2020-09-19
# ... | [((428, 434), 'colorama.init', 'init', ([], {}), '()\n', (432, 434), False, 'from colorama import init, Fore, Style\n'), ((872, 892), 'lib.logger.Logger', 'Logger', (['"""ioe"""', 'level'], {}), "('ioe', level)\n", (878, 892), False, 'from lib.logger import Logger\n'), ((2345, 2376), 'ioexpander.IOE', 'io.IOE', ([], {'... |
RonaldHiemstra/micropython-stubs | stubs/micropython-esp32-1_12/urequests.py | d97f879b01f6687baaebef1c7e26a80909c3cff3 | """
Module: 'urequests' on esp32 1.12.0
"""
# MCU: (sysname='esp32', nodename='esp32', release='1.12.0', version='v1.12 on 2019-12-20', machine='ESP32 module (spiram) with ESP32')
# Stubber: 1.3.2
class Response:
''
def close():
pass
content = None
def json():
pass
text = None
def... | [] |
seckcoder/lang-learn | python/re_user.py | 1e0d6f412bbd7f89b1af00293fd907ddb3c1b571 | #!/usr/bin/env python
#-*- coding=utf-8 -*-
#
# Copyright 2012 Jike Inc. All Rights Reserved.
# Author: [email protected]
import re
from urlparse import urlparse
def parse1():
p = re.compile(r"/(?P<uid>\d+)/(?P<mid>\w+)")
o = urlparse("http://weibo.com/2827699110/yz62AlEjF")
m = p.search(o.path)
print m.... | [] |
LukHad/AccountBook | TransactionBook/gui_kivy/generic/MultiSelectPopUp.py | 8da3ebbd2a824efb9d50f7695ceaaa6cf2370cd8 | from kivy.uix.gridlayout import GridLayout
from kivy.uix.label import Label
from kivy.uix.textinput import TextInput
from kivy.garden.matplotlib.backend_kivyagg import FigureCanvasKivyAgg
from kivy.uix.anchorlayout import AnchorLayout
from kivy.uix.boxlayout import BoxLayout
from kivy.uix.button import Button
import ma... | [((677, 697), 'kivy.properties.NumericProperty', 'NumericProperty', (['(0.7)'], {}), '(0.7)\n', (692, 697), False, 'from kivy.properties import NumericProperty, ReferenceListProperty\n'), ((712, 732), 'kivy.properties.NumericProperty', 'NumericProperty', (['(0.7)'], {}), '(0.7)\n', (727, 732), False, 'from kivy.propert... |
MormonJesus69420/Knowledge-Based-Systems-Project | Assignment1/Part2/Bridge2.py | 8b1e330c64dd58743513f3e48efb6569457beb94 | from dataclasses import dataclass, field
from typing import List
from Car2 import Car
@dataclass
class Bridge:
"""Bridge class simulating the behaviour of bridge in simulation.
On can set specific length and capacity for the bridge to change the overall
behaviour of bridge in the simulation and see how i... | [((384, 400), 'dataclasses.field', 'field', ([], {'default': '(5)'}), '(default=5)\n', (389, 400), False, 'from dataclasses import dataclass, field\n'), ((500, 517), 'dataclasses.field', 'field', ([], {'default': '(10)'}), '(default=10)\n', (505, 517), False, 'from dataclasses import dataclass, field\n'), ((619, 670), ... |
jefflundberg/locast2plex | lib/tuner_interface.py | 3ab747a13c47888507c08f17d0afacad09894019 | import subprocess
import threading
import time
import errno
import socket
import urllib
import pathlib
from io import StringIO
from http.server import BaseHTTPRequestHandler, HTTPServer
import lib.stations as stations
import lib.epg2xml as epg2xml
import lib.channels_m3u as channels_m3u
from lib.templates import templ... | [((18747, 18796), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (18760, 18796), False, 'import socket\n'), ((14506, 14572), 'lib.stations.get_dma_stations_and_channels', 'stations.get_dma_stations_and_channels', (['self.config', 'self.locatio... |
dcs3spp/cookiecutter-django-api | {{ cookiecutter.project_name }}/{{ cookiecutter.project_name }}/local/pages/views.py | d575dda07930743c05a27eb968489867831d97de | from django import template
from django.contrib.auth.decorators import login_required
from django.http import HttpResponse
from django.template import loader
@login_required(login_url="/login/")
def index(request):
context = {}
context["segment"] = "index"
html_template = loader.get_template("index.html"... | [((161, 196), 'django.contrib.auth.decorators.login_required', 'login_required', ([], {'login_url': '"""/login/"""'}), "(login_url='/login/')\n", (175, 196), False, 'from django.contrib.auth.decorators import login_required\n'), ((389, 424), 'django.contrib.auth.decorators.login_required', 'login_required', ([], {'logi... |
jeongyoonlee/audiomentations | audiomentations/core/utils.py | 7f0112ae310989430e0ef7eb32c4116114810966 | import os
from pathlib import Path
import numpy as np
AUDIO_FILENAME_ENDINGS = (".aiff", ".flac", ".m4a", ".mp3", ".ogg", ".opus", ".wav")
def get_file_paths(
root_path, filename_endings=AUDIO_FILENAME_ENDINGS, traverse_subdirectories=True
):
"""Return a list of paths to all files with the given filename ex... | [((453, 471), 'os.walk', 'os.walk', (['root_path'], {}), '(root_path)\n', (460, 471), False, 'import os\n'), ((571, 592), 'os.path.abspath', 'os.path.abspath', (['root'], {}), '(root)\n', (586, 592), False, 'import os\n'), ((617, 651), 'os.path.join', 'os.path.join', (['input_path', 'filename'], {}), '(input_path, file... |
HoneyS2/meaningful | algorithm/python/LeetCode/isValid.py | 78659de1ed74121db4ade211f6565ddc6d117041 | s = "([}}])"
stack = []
if len(s) % 2 == 1:
print(False)
exit()
for i in s:
if i == "(":
stack.append("(")
elif i == "[":
stack.append("[")
elif i == "{":
stack.append("{")
elif i == ")":
if len(stack) < 1:
print(False)
exit()
if... | [] |
williamegomezo/JupyterSlides | JupyterHTMLSlides/core.py | 403fe15e360eb1d79bf813b923eb569a81ab0934 | import random
import string
import os
from IPython.display import display, HTML
from .utils import html_loader
from .utils import get_content
from jinja2 import Template
class JupyterSlides:
def __init__(
self,
content_path='./content.yaml',
table_contents=False
):
self.set_ba... | [((870, 881), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (879, 881), False, 'import os\n'), ((1373, 1421), 'os.path.isfile', 'os.path.isfile', (['f"""{self.template_dir}/init.html"""'], {}), "(f'{self.template_dir}/init.html')\n", (1387, 1421), False, 'import os\n'), ((1782, 1840), 'os.path.isfile', 'os.path.isfile', ... |
NickeZ/lightning | tests/test_wallet.py | f376a9c24cc71d139393196dea86b5a39aee7db8 | from decimal import Decimal
from fixtures import * # noqa: F401,F403
from fixtures import TEST_NETWORK
from flaky import flaky # noqa: F401
from pyln.client import RpcError, Millisatoshi
from utils import (
only_one, wait_for, sync_blockheight, EXPERIMENTAL_FEATURES, COMPAT,
VALGRIND
)
import os
import pytes... | [((371, 487), 'unittest.skipIf', 'unittest.skipIf', (["(TEST_NETWORK != 'regtest')", '"""Test relies on a number of example addresses valid only in regtest"""'], {}), "(TEST_NETWORK != 'regtest',\n 'Test relies on a number of example addresses valid only in regtest')\n", (386, 487), False, 'import unittest\n'), ((89... |
ScienceStacks/MicrobEPy | microbepy/plot/mutation_plot.py | 704435e66c58677bab24f27820458870092924e2 | """Provides plots of mutations for Isolates and Lines."""
from microbepy.common import constants as cn
from microbepy.common.dataframe_sorter import DataframeSorter
from microbepy.common.isolate import Isolate
from microbepy.common import util
from microbepy.correlation import genome_correlation
from microbepy.data.m... | [((1254, 1328), 'microbepy.plot.mutation_cofraction.MutationCofraction', 'MutationCofraction', ([], {'species': 'self._species', 'mutation_column': 'mutation_column'}), '(species=self._species, mutation_column=mutation_column)\n', (1272, 1328), False, 'from microbepy.plot.mutation_cofraction import MutationCofraction\n... |
shikisawamura/nnabla-examples | semantic-segmentation/deeplabv3plus/dataset_utils.py | baf4e4cc620dedbf4368683325c0fb868676850d | # Copyright (c) 2017 Sony Corporation. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | [((771, 1112), 'numpy.array', 'np.array', (['[[0, 0, 0], [128, 0, 0], [0, 128, 0], [128, 128, 0], [0, 0, 128], [120, 0, \n 128], [0, 128, 128], [128, 128, 128], [64, 0, 0], [192, 0, 0], [64, 128,\n 0], [192, 128, 0], [64, 0, 128], [192, 0, 128], [64, 128, 128], [192, \n 128, 128], [0, 64, 0], [128, 64, 0], [0,... |
shamilcm/fairseq-py | train.py | ceb2f1200c9e5b8bf42a1033e7638d3e8586609a | # Copyright (c) 2017-present, Facebook, Inc.
# All rights reserved.
#
# This source code is licensed under the license found in the LICENSE file in
# the root directory of this source tree. An additional grant of patent rights
# can be found in the PATENTS file in the same directory.
#
import collections
import os
imp... | [((654, 683), 'fairseq.options.get_parser', 'options.get_parser', (['"""Trainer"""'], {}), "('Trainer')\n", (672, 683), False, 'from fairseq import bleu, data, options, utils\n'), ((703, 735), 'fairseq.options.add_dataset_args', 'options.add_dataset_args', (['parser'], {}), '(parser)\n', (727, 735), False, 'from fairse... |
jeffmoffitt/storm-control | storm_control/sc_library/log_timing.py | 522add1e196e0b7964f574481fd90c20a74b575e | #!/usr/bin/env python
"""
This parses a log file series (i.e. log, log.1, log.2, etc..) and
outputs timing and call frequency information for HAL messages.
Hazen 5/18
"""
from datetime import datetime
import os
pattern = '%Y-%m-%d %H:%M:%S,%f'
class Message(object):
"""
Storage for the timing of a single m... | [((2042, 2074), 'datetime.datetime.strptime', 'datetime.strptime', (['time', 'pattern'], {}), '(time, pattern)\n', (2059, 2074), False, 'from datetime import datetime\n'), ((3998, 4019), 'os.path.exists', 'os.path.exists', (['fname'], {}), '(fname)\n', (4012, 4019), False, 'import os\n')] |
38elements/django-simple-jsonschema | django_simple_jsonschema/management/commands/check_schema.py | ab08aaa3453c40a41d443869643113f23eb40db6 | from django.core.management.base import BaseCommand
from django.utils import termcolors
from jsonschema import Draft4Validator
from jsonschema.exceptions import SchemaError
import json
class Command(BaseCommand):
can_import_settings = True
@property
def _jsonschema_exist(self):
from django.conf ... | [((1013, 1046), 'django.utils.termcolors.make_style', 'termcolors.make_style', ([], {'fg': '"""green"""'}), "(fg='green')\n", (1034, 1046), False, 'from django.utils import termcolors\n'), ((1063, 1094), 'django.utils.termcolors.make_style', 'termcolors.make_style', ([], {'fg': '"""red"""'}), "(fg='red')\n", (1084, 109... |
keke185321/combine-copy- | lib/interface.py | de2eba77d8db5c9c1908aac1262590b80c2348ce | import cv2, time
import numpy as np
import Tkinter
"""
Wraps up some interfaces to opencv user interface methods (displaying
image frames, event handling, etc).
If desired, an alternative UI could be built and imported into get_pulse.py
instead. Opencv is used to perform much of the data analysis, but there is no
re... | [((468, 495), 'cv2.resize', 'cv2.resize', (['*args'], {}), '(*args, **kwargs)\n', (478, 495), False, 'import cv2, time\n'), ((583, 628), 'cv2.cvtColor', 'cv2.cvtColor', (['output_frame', 'cv2.COLOR_BGR2RGB'], {}), '(output_frame, cv2.COLOR_BGR2RGB)\n', (595, 628), False, 'import cv2, time\n'), ((845, 879), 'cv2.destroy... |
FGDBTKD/nltk | nltk/tag/brill.py | 384e46e82789c7f47a7fb521ef976f82c3c4c3fb | # -*- coding: utf-8 -*-
# Natural Language Toolkit: Transformation-based learning
#
# Copyright (C) 2001-2018 NLTK Project
# Author: Marcus Uneson <[email protected]>
# based on previous (nltk2) version by
# Christopher Maloof, Edward Loper, Steven Bird
# URL: <http://nltk.org/>
# For license information, see... | [((5349, 5410), 'inspect.getmembers', 'inspect.getmembers', (['sys.modules[__name__]', 'inspect.isfunction'], {}), '(sys.modules[__name__], inspect.isfunction)\n', (5367, 5410), False, 'import inspect\n'), ((8431, 8447), 'collections.defaultdict', 'defaultdict', (['set'], {}), '(set)\n', (8442, 8447), False, 'from coll... |
agentsolaris/xlnn | superglue_parsers/wsc.py | 0ab07d1ac526cadc2964379aef0a44927e0618eb | import json
import logging
import sys
import numpy as np
import torch
from task_config import SuperGLUE_LABEL_MAPPING
from snorkel.mtl.data import MultitaskDataset
sys.path.append("..") # Adds higher directory to python modules path.
logger = logging.getLogger(__name__)
TASK_NAME = "WSC"
def get_char_index(tex... | [((167, 188), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (182, 188), False, 'import sys\n'), ((249, 276), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (266, 276), False, 'import logging\n'), ((7278, 7658), 'snorkel.mtl.data.MultitaskDataset', 'MultitaskDataset... |
TimmyCarbone/simplimental | simplimental/simplimental.py | e46a0e63ce33e36b1e4ca3a473ad15d0732614ed | import re
import json
__all__ = ["Simplimental"]
class Simplimental:
def __init__(self, text="This is not a bad idea"):
self.text = text
with open('simplimental/data/afinn.json') as data_file:
self.dictionary = json.load(data_file)
no_punctunation = re.sub(r"[^a-zA-Z ]+", " ", self.text)
self.toke... | [((270, 307), 're.sub', 're.sub', (['"""[^a-zA-Z ]+"""', '""" """', 'self.text'], {}), "('[^a-zA-Z ]+', ' ', self.text)\n", (276, 307), False, 'import re\n'), ((228, 248), 'json.load', 'json.load', (['data_file'], {}), '(data_file)\n', (237, 248), False, 'import json\n')] |
archformco/RoboDK-API | Python/Examples/Macros/SettingsAxesOptimization.py | b3d0cad6a83f505811e2be273453ccb4579324f1 | # This example shows how to read or modify the Axes Optimization settings using the RoboDK API and a JSON string.
# You can select "Axes optimization" in a robot machining menu or the robot parameters to view the axes optimization settings.
# It is possible to update the axes optimization settings attached to a robot o... | [((2415, 2435), 'json.dumps', 'json.dumps', (['ToUpdate'], {}), '(ToUpdate)\n', (2425, 2435), False, 'import json\n'), ((3335, 3356), 'json.loads', 'json.loads', (['json_data'], {}), '(json_data)\n', (3345, 3356), False, 'import json\n'), ((3555, 3576), 'json.loads', 'json.loads', (['json_data'], {}), '(json_data)\n', ... |
Vipul97/SLR-Parser | tests/test_grammar.py | 3de5609235d173d29ad9bd9ed7bdfe2a813ab1bd | from slr_parser.grammar import Grammar
import unittest
class TestGrammar(unittest.TestCase):
def test_grammar(self):
with open('tests/test_grammar.txt') as grammar_file:
self.G = Grammar(grammar_file.read())
self.assertDictEqual(
{'E': {('E', '+', 'T'), ('T',)}, 'T'... | [((1056, 1071), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1069, 1071), False, 'import unittest\n'), ((908, 936), 'slr_parser.grammar.Grammar', 'Grammar', (['self.grammar_str[0]'], {}), '(self.grammar_str[0])\n', (915, 936), False, 'from slr_parser.grammar import Grammar\n'), ((994, 1022), 'slr_parser.grammar... |
leonhead/chess-insight | insight/migrations/0001_initial.py | b893295719df21b4fee10d4e7b01639ded8b42b4 | # Generated by Django 3.1 on 2020-09-08 07:43
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='OpeningSystem',
fields=[
... | [((340, 433), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (356, 433), False, 'from django.db import migrations, models\... |
jmerizia/parallel-pytorch | tests/test_merge.py | d27b2fd145d25f1329a039c99b8895783bfc71e5 | import torch
import numpy as np
from mpi4py import MPI
from parallel_pytorch.ops import tensor_merge
from parallel_pytorch.utils import abort_on_exception
@abort_on_exception
def test_1():
worker_shape = [2, 2]
world = MPI.COMM_WORLD
num_workers = np.array(worker_shape).prod()
comm = MPI.COMM_WORLD.... | [((781, 834), 'parallel_pytorch.ops.tensor_merge', 'tensor_merge', (['x'], {'comm': 'comm', 'worker_shape': 'worker_shape'}), '(x, comm=comm, worker_shape=worker_shape)\n', (793, 834), False, 'from parallel_pytorch.ops import tensor_merge\n'), ((1498, 1551), 'parallel_pytorch.ops.tensor_merge', 'tensor_merge', (['x'], ... |
tebriel/aoc2021 | day07/main.py | 65ca19be3ad66dc52eee9ca31cf12306695a24e9 | """Day 07"""
def process(filename):
with open(filename) as infile:
positions = [int(x) for x in infile.readline().strip().split(',')]
min_x = min(positions)
max_x = max(positions)
costs = {x: 0 for x in range(min_x, max_x + 1)}
for pos in costs.keys():
for crab in positions:
... | [] |
pritchardlabatpsu/cga | src/ceres_infer/utils.py | 0a71c672b1348cebc724560643fd908d636fc133 | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
utilities
@author: boyangzhao
"""
import pandas as pd
import re
def int2ordinal(n):
# partially based on https://stackoverflow.com/questions/9647202/ordinal-numbers-replacement
if (type(n) is int) or n.isdigit():
if type(n) is not int:
n =... | [((500, 512), 'pandas.isnull', 'pd.isnull', (['x'], {}), '(x)\n', (509, 512), True, 'import pandas as pd\n'), ((545, 608), 're.findall', 're.findall', (['"""([^,\\\\()]*)\\\\s(\\\\(([^,]*)\\\\)\\\\s)*\\\\[([^,]*)\\\\]"""', 'x'], {}), "('([^,\\\\()]*)\\\\s(\\\\(([^,]*)\\\\)\\\\s)*\\\\[([^,]*)\\\\]', x)\n", (555, 608), F... |
aragubas/fogoso | MAIN/Screens/Settings/category_2/__init__.py | bd24e049ee994410320e87fb3706c95bd8c9801f | #!/usr/bin/python3.7
# Copyright 2020 Aragubas
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | [((1267, 1298), 'Fogoso.MAIN.ClassesUtils.UpDownButton', 'gameObjs.UpDownButton', (['(0)', '(0)', '(14)'], {}), '(0, 0, 14)\n', (1288, 1298), True, 'from Fogoso.MAIN import ClassesUtils as gameObjs\n'), ((1335, 1366), 'Fogoso.MAIN.ClassesUtils.UpDownButton', 'gameObjs.UpDownButton', (['(0)', '(0)', '(14)'], {}), '(0, 0... |
ConnectionMaster/robotpy-wpilib-utilities | tests/test_toggle.py | b62e563c7df113e9e513a36b9039f47f34157be1 | from robotpy_ext.control.toggle import Toggle
from robotpy_ext.misc.precise_delay import NotifierDelay
class FakeJoystick:
def __init__(self):
self._pressed = [False] * 2
def getRawButton(self, num):
return self._pressed[num]
def press(self, num):
self._pressed[num] = True
d... | [((448, 467), 'robotpy_ext.control.toggle.Toggle', 'Toggle', (['joystick', '(0)'], {}), '(joystick, 0)\n', (454, 467), False, 'from robotpy_ext.control.toggle import Toggle\n'), ((488, 507), 'robotpy_ext.control.toggle.Toggle', 'Toggle', (['joystick', '(1)'], {}), '(joystick, 1)\n', (494, 507), False, 'from robotpy_ext... |
kjanik70/tflearn | tests/test.py | db5176773299b67a2a75c5889fb2aba7fd0fea8a | '''
This file contains test cases for tflearn
'''
import tensorflow.compat.v1 as tf
import tflearn
import unittest
class TestActivations(unittest.TestCase):
'''
This class contains test cases for the functions in tflearn/activations.py
'''
PLACES = 4 # Number of places to match when testing fl... | [((2857, 2872), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2870, 2872), False, 'import unittest\n'), ((425, 461), 'tensorflow.compat.v1.placeholder', 'tf.placeholder', (['tf.float32'], {'shape': '()'}), '(tf.float32, shape=())\n', (439, 461), True, 'import tensorflow.compat.v1 as tf\n'), ((526, 560), 'tensorf... |
roolrd/incubator-datalab | infrastructure-provisioning/src/general/api/install_libs.py | 2045207ecd1b381193f1a1ec143cc968716ad989 | #!/usr/bin/python3
# *****************************************************************************
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ... | [((1086, 1154), 'subprocess.run', 'subprocess.run', (['"""cd /root; fab install-libs"""'], {'shell': '(True)', 'check': '(True)'}), "('cd /root; fab install-libs', shell=True, check=True)\n", (1100, 1154), False, 'import subprocess\n'), ((2173, 2236), 'subprocess.run', 'subprocess.run', (['"""chmod 666 /response/*"""']... |
GuoSuiming/mindspore | model_zoo/official/nlp/bert_thor/src/evaluation_config.py | 48afc4cfa53d970c0b20eedfb46e039db2a133d5 | # Copyright 2020 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to... | [((844, 990), 'easydict.EasyDict', 'edict', (["{'task': 'NER', 'num_labels': 41, 'data_file': '', 'schema_file': None,\n 'finetune_ckpt': '', 'use_crf': False, 'clue_benchmark': False}"], {}), "({'task': 'NER', 'num_labels': 41, 'data_file': '', 'schema_file':\n None, 'finetune_ckpt': '', 'use_crf': False, 'clue_... |
Artis-Physis/utopia-cms | portal/apps/core/management/commands/sync_articleviewedby.py | 5cb8d941d0b2df53fddc566a52e9d3baee4a007e | # -*- coding: utf-8 -*-
# utopia-cms 2020. Aníbal Pacheco.
from django.core.management import BaseCommand
from django.db.utils import IntegrityError
from apps import core_articleviewedby_mdb
from core.models import ArticleViewedBy
class Command(BaseCommand):
help = "Moves article viewed by data from mongodb to ... | [((394, 448), 'apps.core_articleviewedby_mdb.posts.find_one_and_delete', 'core_articleviewedby_mdb.posts.find_one_and_delete', (['{}'], {}), '({})\n', (444, 448), False, 'from apps import core_articleviewedby_mdb\n'), ((997, 1051), 'apps.core_articleviewedby_mdb.posts.find_one_and_delete', 'core_articleviewedby_mdb.pos... |
dendisuhubdy/MinkowskiEngine | examples/minkunet.py | a1cdcba68ef925bfefed2fe161f62e1ec78573b9 | import torch
import torch.nn as nn
from torch.optim import SGD
import MinkowskiEngine as ME
from MinkowskiEngine.modules.resnet_block import BasicBlock, Bottleneck
from examples.common import data_loader
from examples.resnet import ResNetBase
class MinkUNetBase(ResNetBase):
BLOCK = None
PLANES = None
D... | [((6981, 7002), 'torch.nn.CrossEntropyLoss', 'nn.CrossEntropyLoss', ([], {}), '()\n', (7000, 7002), True, 'import torch.nn as nn\n'), ((675, 730), 'examples.resnet.ResNetBase.__init__', 'ResNetBase.__init__', (['self', 'in_channels', 'out_channels', 'D'], {}), '(self, in_channels, out_channels, D)\n', (694, 730), False... |
swfrench/nginx-access-tailer | setup.py | 5e060396ca749935c622e8e9c50b659b39e3675b | """TODO."""
from setuptools import setup
setup(
name='nginx-access-tailer',
version='0.1',
author='swfrench',
url='https://github.com/swfrench/nginx-tailer',
packages=['nginx_access_tailer',],
license='BSD three-clause license',
entry_points={
'console_scripts': ['nginx-access-tail... | [((43, 497), 'setuptools.setup', 'setup', ([], {'name': '"""nginx-access-tailer"""', 'version': '"""0.1"""', 'author': '"""swfrench"""', 'url': '"""https://github.com/swfrench/nginx-tailer"""', 'packages': "['nginx_access_tailer']", 'license': '"""BSD three-clause license"""', 'entry_points': "{'console_scripts': ['ngi... |
oglok/CPU-Manager-for-Kubernetes | tests/integration/test_cmk_describe.py | 503f37dcb20452699ce789b6628fa3ebeb9ffb54 | # Copyright (c) 2017 Intel Corporation
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to i... | [] |
richardARPANET/persistent-celery-beat-scheduler | setup.py | d2cbdd12394eec282ccb97ac5ff894353c2e4ffd | #!/usr/bin/env python
# -*- coding: utf-8 -*
import os
from setuptools import find_packages, setup
# allow setup.py to be run from any path
os.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))
with open('requirements.txt') as f:
install_requires = f.read().splitlines()
setup(
name... | [((396, 436), 'setuptools.find_packages', 'find_packages', (['"""src"""'], {'exclude': "('tests',)"}), "('src', exclude=('tests',))\n", (409, 436), False, 'from setuptools import find_packages, setup\n'), ((182, 207), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (197, 207), False, 'import o... |
mgobec/python-memcached | tests/test_client.py | 8ea5fe5fca3a4f0d1201ca9aa50f9701c1baab01 | import collections
import unittest
import driver
from driver.protocol import *
_server = ('localhost', 11211)
_dead_retry = 30
_socket_timeout = 3
_max_receive_size = 4096
class MockConnection(object):
def __init__(self,
server=_server,
dead_retry=30,
socket_tim... | [((521, 540), 'collections.deque', 'collections.deque', ([], {}), '()\n', (538, 540), False, 'import collections\n'), ((571, 590), 'collections.deque', 'collections.deque', ([], {}), '()\n', (588, 590), False, 'import collections\n'), ((1188, 1210), 'driver.Client', 'driver.Client', (['_server'], {}), '(_server)\n', (1... |
Preethinaidu14/greyatom-python-for-data-science | GREYATOM-PROJECT----DATA--WRANGLING-WITH-PANDAS/code.py | 5b758dd6123d9fc50031c43771b30d69e366c044 | # --------------
# Import packages
import numpy as np
import pandas as pd
from scipy.stats import mode
path
# code starts here
bank = pd.read_csv(path)
categorical_var = bank.select_dtypes(include = 'object')
print(categorical_var)
numerical_var = bank.select_dtypes(include = 'number')
print(numerical_var)
# code... | [((139, 156), 'pandas.read_csv', 'pd.read_csv', (['path'], {}), '(path)\n', (150, 156), True, 'import pandas as pd\n')] |
EkremBayar/bayar | venv/Lib/site-packages/patsy/test_regressions.py | aad1a32044da671d0b4f11908416044753360b39 | # This file is part of Patsy
# Copyright (C) 2013 Nathaniel Smith <[email protected]>
# See file LICENSE.txt for license information.
# Regression tests for fixed bugs (when not otherwise better covered somewhere
# else)
from patsy import (EvalEnvironment, dmatrix, build_design_matrices,
PatsyError, Or... | [((485, 510), 'patsy.EvalEnvironment.capture', 'EvalEnvironment.capture', ([], {}), '()\n', (508, 510), False, 'from patsy import EvalEnvironment, dmatrix, build_design_matrices, PatsyError, Origin\n'), ((655, 677), 'patsy.dmatrix', 'dmatrix', (['formula', 'data'], {}), '(formula, data)\n', (662, 677), False, 'from pat... |
smheidrich/scikit-image | skimage/io/_plugins/pil_plugin.py | e9cf8b850c4c2800cc221be6f1dfff6a2a32a4eb | __all__ = ['imread', 'imsave']
import numpy as np
from PIL import Image
from ...util import img_as_ubyte, img_as_uint
def imread(fname, dtype=None, img_num=None, **kwargs):
"""Load an image from file.
Parameters
----------
fname : str or file
File name or file-like-object.
dtype : numpy ... | [((7329, 7347), 'numpy.asanyarray', 'np.asanyarray', (['arr'], {}), '(arr)\n', (7342, 7347), True, 'import numpy as np\n'), ((1049, 1066), 'PIL.Image.open', 'Image.open', (['fname'], {}), '(fname)\n', (1059, 1066), False, 'from PIL import Image\n'), ((3457, 3473), 'numpy.array', 'np.array', (['frames'], {}), '(frames)\... |
ShaikAsifullah/distributed-tellurium | examples/tellurium-files/linearChain.py | 007e9b3842b614edd34908c001119c6da1d41897 | # -*- coding: utf-8 -*-
"""
Linear chain of reactions.
"""
from __future__ import print_function, division
import tellurium as te
model = '''
model feedback()
// Reactions:
J0: $X0 -> S1; (VM1 * (X0 - S1/Keq1))/(1 + X0 + S1 + S4^h);
J1: S1 -> S2; (10 * S1 - 2 * S2) / (1 + S1 + S2);
J2: S2 -> S3; (10 * S2... | [((622, 637), 'tellurium.loada', 'te.loada', (['model'], {}), '(model)\n', (630, 637), True, 'import tellurium as te\n')] |
kommurisaikumar/savings-manager-server | backend/app/schemas/__init__.py | ed699abddf3cecdd4056aaee0129fbb1ef3762f6 | from .users import User, UserCreate, UserUpdate
from .transactions import Transaction, TransactionCreate, TransactionUpdate
from .accounts import Account, AccountList, AccountSingle, AccountCreate, AccountUpdate
from .categories import Category, CategoryCreate, CategoryUpdate | [] |
ciskoinch8/vimrc | vimfiles/bundle/vim-python/submodules/pylint/tests/functional/s/super/super_with_arguments.py | 5bf77a7e7bc70fac5173ab2e9ea05d7dda3e52b8 | class Foo:
pass
class Bar(Foo):
def __init__(self):
super(Bar, self).__init__() # [super-with-arguments]
class Baz(Foo):
def __init__(self):
super().__init__()
class Qux(Foo):
def __init__(self):
super(Bar, self).__init__()
class NotSuperCall(Foo):
def __init__(self)... | [] |
yukti07/Dell_Hire_hack | machine.py | 9422b7aaa0b96292191b4b880c0a8fb772fd1864 | import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import LabelEncoder
from sklearn.ensemble import RandomForestClassifier
from sklearn.metrics import accuracy_score
from flask import flash
import numpy as np
def ch... | [((402, 413), 'numpy.array', 'np.array', (['X'], {}), '(X)\n', (410, 413), True, 'import numpy as np\n'), ((437, 451), 'sklearn.preprocessing.LabelEncoder', 'LabelEncoder', ([], {}), '()\n', (449, 451), False, 'from sklearn.preprocessing import LabelEncoder\n'), ((529, 543), 'sklearn.preprocessing.LabelEncoder', 'Label... |
OsmanMalik/TM-GCN | TM-GCN-master/experiment_bitcoin_baseline_link_prediction.py | 31b19a538f264f6c30b5503ecefb497ee865b4d7 | # This version of the bitcoin experiment imports data preprocessed in Matlab, and uses the GCN baseline
# The point of this script is to do link prediction
# Imports and aliases
import pickle
import torch as t
import torch.nn as nn
import torch.nn.functional as F
import torchvision
import torchvision.datasets as datas... | [((1080, 1158), 'embedding_help_functions.load_data', 'ehf.load_data', (['data_loc', 'mat_f_name', 'S_train', 'S_val', 'S_test'], {'transformed': '(False)'}), '(data_loc, mat_f_name, S_train, S_val, S_test, transformed=False)\n', (1093, 1158), True, 'import embedding_help_functions as ehf\n'), ((1217, 1291), 'embedding... |
ahara/-blog | elm_mnist/elm_mnist.py | 926ae4808ede6efb1e64381a19a210235a97ac36 | import cPickle
import numpy as np
from elm import ELMClassifier
from sklearn import linear_model
def load_mnist(path='../Data/mnist.pkl'):
with open(path, 'rb') as f:
return cPickle.load(f)
def get_datasets(data):
_train_x, _train_y = data[0][0], np.array(data[0][1]).reshape(len(data[0][1]), 1)
... | [] |
qizhu8/CSCI6230-HW02 | Coding_Part/bob.py | c889c0532db7ff4f25e134937469e5e6181416f0 | # -*- coding: utf-8 -*-
#!/usr/bin/env python3
from PKC_Classes import NetworkUser, KDC
from DES import DES
from RSA_Class import RSA
import socket
import os
import sys
import threading
import time
if sys.version_info[0] < 3:
raise Exception("Must be using Python 3")
def reply_conn(conn, addr):
print('Accep... | [((804, 809), 'DES.DES', 'DES', ([], {}), '()\n', (807, 809), False, 'from DES import DES\n'), ((811, 824), 'RSA_Class.RSA', 'RSA', (['(9973)', '(97)'], {}), '(9973, 97)\n', (814, 824), False, 'from RSA_Class import RSA\n'), ((1007, 1056), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {})... |
zqma/IIC | proj/scripts/cluster/baselines/triplets_greyscale.py | 9d4e30b51535c6ca381389d9c22ce45be4d11883 | from __future__ import print_function
import argparse
import itertools
import os
import pickle
import sys
from datetime import datetime
import matplotlib
import numpy as np
import torch
matplotlib.use('Agg')
import matplotlib.pyplot as plt
import proj.archs as archs
from proj.utils.cluster.general import config_to_... | [((189, 210), 'matplotlib.use', 'matplotlib.use', (['"""Agg"""'], {}), "('Agg')\n", (203, 210), False, 'import matplotlib\n'), ((687, 712), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (710, 712), False, 'import argparse\n'), ((4849, 4875), 'proj.utils.cluster.baselines.triplets.make_triplets... |
cds-snc/notifier-api | migrations/versions/0084_add_job_stats.py | 90b385ec49efbaee7e607516fc7d9f08991af813 | """empty message
Revision ID: 0084_add_job_stats
Revises: 0083_add_perm_types_and_svc_perm
Create Date: 2017-05-12 13:16:14.147368
"""
# revision identifiers, used by Alembic.
revision = "0084_add_job_stats"
down_revision = "0083_add_perm_types_and_svc_perm"
import sqlalchemy as sa
from alembic import op
from sqlal... | [((1575, 1606), 'alembic.op.drop_table', 'op.drop_table', (['"""job_statistics"""'], {}), "('job_statistics')\n", (1588, 1606), False, 'from alembic import op\n'), ((1244, 1292), 'sqlalchemy.ForeignKeyConstraint', 'sa.ForeignKeyConstraint', (["['job_id']", "['jobs.id']"], {}), "(['job_id'], ['jobs.id'])\n", (1267, 1292... |
tsukaeru/RDM-osf.io | addons/twofactor/tests/test_models.py | 2dc3e539322b6110e51772f8bd25ebdeb8e12d0e | import unittest
from future.moves.urllib.parse import urlparse, urljoin, parse_qs
import pytest
from addons.twofactor.tests.utils import _valid_code
from nose.tools import (assert_equal, assert_false, assert_is_none,
assert_is_not_none, assert_true)
from osf_tests.factories import UserFactory
... | [((482, 495), 'osf_tests.factories.UserFactory', 'UserFactory', ([], {}), '()\n', (493, 495), False, 'from osf_tests.factories import UserFactory\n'), ((640, 686), 'nose.tools.assert_equal', 'assert_equal', (['self.user_settings.totp_drift', '(0)'], {}), '(self.user_settings.totp_drift, 0)\n', (652, 686), False, 'from ... |
BinItAI/BetterLoader | betterloader/standard_transforms.py | 29ebcc22b53db6417a4b14d95f0a1e7f5afe7af8 | import numpy as np
from torchvision import transforms
np.random.seed(1)
class TransformWhileSampling(object):
def __init__(self, transform):
self.transform = transform
def __call__(self, sample):
x1 = self.transform(sample)
x2 = self.transform(sample)
return x1, x2 | [((54, 71), 'numpy.random.seed', 'np.random.seed', (['(1)'], {}), '(1)\n', (68, 71), True, 'import numpy as np\n')] |
ztjsw/lanedet | lanedet/runner/utils/net_utils.py | c957e1f70695e39063231612637e22fcad2769f5 | import torch
import os
from torch import nn
import numpy as np
import torch.nn.functional
from termcolor import colored
from .logger import get_logger
def save_model(net, optim, scheduler, recorder, is_best=False):
model_dir = os.path.join(recorder.work_dir, 'ckpt')
os.system('mkdir -p {}'.format(model_dir))
... | [((232, 271), 'os.path.join', 'os.path.join', (['recorder.work_dir', '"""ckpt"""'], {}), "(recorder.work_dir, 'ckpt')\n", (244, 271), False, 'import os\n'), ((1624, 1645), 'torch.load', 'torch.load', (['model_dir'], {}), '(model_dir)\n', (1634, 1645), False, 'import torch\n'), ((1002, 1023), 'torch.load', 'torch.load',... |
zante95/Hexafuel-Oil | hexafuel_oil/hexafuel_oil_app/apps.py | 41dc4c9d855c74d4bb7dd86f3ac3fb1db27b663b | from django.apps import AppConfig #pragma: no cover
class HexafuelOilAppConfig(AppConfig): #pragma: no cover
name = 'hexafuel_oil_app'
| [] |
jonodrew/matchex | main.py | 531e7cd1c328cb9dc34b601a06648bd2c3e709e6 | from __future__ import division
from timeit import default_timer as timer
import csv
import numpy as np
import itertools
from munkres import Munkres, print_matrix, make_cost_matrix
import sys
from classes import *
from functions import *
from math import sqrt
import Tkinter as tk
import tkFileDialog as filedialog
root... | [((323, 330), 'Tkinter.Tk', 'tk.Tk', ([], {}), '()\n', (328, 330), True, 'import Tkinter as tk\n'), ((356, 422), 'tkFileDialog.askopenfilename', 'filedialog.askopenfilename', ([], {'title': '"""Please select the posting file"""'}), "(title='Please select the posting file')\n", (382, 422), True, 'import tkFileDialog as ... |
95616ARG/SyReNN | pip_info/setup.py | 19abf589e84ee67317134573054c648bb25c244d | """Setup script for PySyReNN.
Adapted from:
https://hynek.me/articles/sharing-your-labor-of-love-pypi-quick-and-dirty/
"""
import codecs
import os
import re
from setuptools import setup, find_packages
###################################################################
NAME = "pysyrenn"
PACKAGES = [
"syrenn_prot... | [((1800, 1825), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (1815, 1825), False, 'import os\n'), ((2002, 2028), 'os.path.join', 'os.path.join', (['HERE', '*parts'], {}), '(HERE, *parts)\n', (2014, 2028), False, 'import os\n')] |
jiji-online/neptune-cli | neptune/generated/swagger_client/path_constants.py | 50cf680a80d141497f9331ab7cdaee49fcb90b0c | REST_PATH = u""
WS_PATH = u"/api/notifications/v1"
| [] |
andrewp-as-is/load.py | load/__init__.py | 6ad643d82379a63f9c79d0dd994101ff0b490183 | __all__ = ["load"]
import imp
import importlib
def load(name, path):
"""Load and initialize a module implemented as a Python source file and return its module object"""
if hasattr(importlib, "machinery"):
loader = importlib.machinery.SourceFileLoader(name, path)
return loader.load_module()
... | [((330, 357), 'imp.load_source', 'imp.load_source', (['name', 'path'], {}), '(name, path)\n', (345, 357), False, 'import imp\n'), ((234, 282), 'importlib.machinery.SourceFileLoader', 'importlib.machinery.SourceFileLoader', (['name', 'path'], {}), '(name, path)\n', (270, 282), False, 'import importlib\n')] |
Risto97/pygears | pygears/svgen/modules/sieve.py | 19393e85101a16762cb3bbbf3010946ef69217f2 | import itertools
from pygears.common.sieve import sieve
from pygears.svgen.inst import SVGenInstPlugin
from pygears.svgen.svmod import SVModuleGen
from functools import partial
from pygears.svgen.svgen import SVGenPlugin
from pygears.svgen.util import svgen_visitor
from pygears.core.hier_node import HierVisitorBase
fr... | [((904, 954), 'itertools.chain', 'itertools.chain', (['self.node.pre_sieves', '[self.node]'], {}), '(self.node.pre_sieves, [self.node])\n', (919, 954), False, 'import itertools\n'), ((2299, 2330), 'pygears.rtl.gear.is_gear_instance', 'is_gear_instance', (['p.node', 'sieve'], {}), '(p.node, sieve)\n', (2315, 2330), Fals... |
gzpyy/qlib | examples/my_model_test.py | 56fdd028c8296c75f2a32bdb51869f010dd4f6d1 | #encoding=utf-8
import qlib
import pandas as pd
import pickle
import xgboost as xgb
import numpy as np
import re
from qlib.constant import REG_US
from qlib.utils import exists_qlib_data, init_instance_by_config
from qlib.workflow import R
from qlib.workflow.record_temp import SignalRecord, PortAnaRecord
from qlib.utils... | [((847, 881), 'my_data_handler.MyAlphaHandler.get_custom_config', 'MyAlphaHandler.get_custom_config', ([], {}), '()\n', (879, 881), False, 'from my_data_handler import MyAlphaHandler\n'), ((895, 954), 'pandas.read_csv', 'pd.read_csv', (['"""../stock_data/TSLA.csv"""'], {'parse_dates': "['time']"}), "('../stock_data/TSL... |
wborbajr/RealFastAPI | realfastapi/routes/endpoints/default.py | d97ca994c4c164387632cda814e80c026435a9f7 | from fastapi import APIRouter
router = APIRouter()
@router.get("/")
def working():
return {"Working"}
| [((40, 51), 'fastapi.APIRouter', 'APIRouter', ([], {}), '()\n', (49, 51), False, 'from fastapi import APIRouter\n')] |
junhoher/GraphZoom | graphzoom/embed_methods/dgi/execute.py | 5073b49a34badf7bc6c25bd2a6cc6c78b4ee7d5a | import numpy as np
import scipy.sparse as sp
import torch
import torch.nn as nn
import networkx as nx
import time
from embed_methods.dgi.models import DGI, LogReg
from embed_methods.dgi.utils import process
def dgi(G, features):
batch_size = 1
nb_epochs = 10000
patience = 20
lr = 0.001
l2_coef = 0.0
drop... | [((438, 480), 'networkx.to_scipy_sparse_matrix', 'nx.to_scipy_sparse_matrix', (['G'], {'weight': '"""wgt"""'}), "(G, weight='wgt')\n", (463, 480), True, 'import networkx as nx\n'), ((545, 582), 'embed_methods.dgi.utils.process.preprocess_features', 'process.preprocess_features', (['features'], {}), '(features)\n', (572... |
rodb70/RDMnet | tools/ci/deploy_to_github_release.py | 94d17e1dfda2d1f56b120f6342231c43bf6862b0 | """Deploys binaries to a GitHub release given the specified tag name."""
import argparse
import os
import time
from github import Github
THIS_FILE_DIRECTORY = os.path.dirname(os.path.realpath(__file__))
GH_REPO_IDENT = "ETCLabs/RDMnet"
GH_USERNAME = "svc-etclabs"
GH_API_TOKEN = os.getenv("SVC_ETCLABS_REPO_TOKEN")
de... | [((281, 316), 'os.getenv', 'os.getenv', (['"""SVC_ETCLABS_REPO_TOKEN"""'], {}), "('SVC_ETCLABS_REPO_TOKEN')\n", (290, 316), False, 'import os\n'), ((177, 203), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (193, 203), False, 'import os\n'), ((420, 477), 'github.Github', 'Github', ([], {'lo... |
maximskorik/matchms | matchms/filtering/add_losses.py | 922f5afaef123a793194bdd74391027477cbb844 | import logging
import numpy
from ..Fragments import Fragments
from ..typing import SpectrumType
logger = logging.getLogger("matchms")
def add_losses(spectrum_in: SpectrumType, loss_mz_from=0.0, loss_mz_to=1000.0) -> SpectrumType:
"""Derive losses based on precursor mass.
Parameters
----------
spect... | [((107, 135), 'logging.getLogger', 'logging.getLogger', (['"""matchms"""'], {}), "('matchms')\n", (124, 135), False, 'import logging\n'), ((1156, 1224), 'numpy.where', 'numpy.where', (['((losses_mz >= loss_mz_from) & (losses_mz <= loss_mz_to))'], {}), '((losses_mz >= loss_mz_from) & (losses_mz <= loss_mz_to))\n', (1167... |
baobabsoluciones/cornflow-client | cornflow_client/schema/dictSchema.py | f9996f0b841885d26639cb63c8ba6090387de57f | import re
from .dict_functions import gen_schema, ParameterSchema, sort_dict
from cornflow_client.constants import JSON_TYPES, DATASCHEMA
class DictSchema:
"""
A json-schema to dict-schema parser
"""
def __init__(self, jsonschema):
"""
Class to manage internal dictionary schema
... | [((7995, 8041), 're.search', 're.search', (['"""definitions/(.+)"""', "content['$ref']"], {}), "('definitions/(.+)', content['$ref'])\n", (8004, 8041), False, 'import re\n')] |
EHRI/rspub-core | rspub/util/test/test_resourcefilter.py | 1f6b0c84825037b7df442ae0d258d5d897ff6905 | #! /usr/bin/env python3
# -*- coding: utf-8 -*-
import os
import platform
import unittest
import rspub.util.resourcefilter as rf
def on_windows():
opsys = platform.system()
return opsys == "Windows"
class TestPredicates(unittest.TestCase):
def test_directory_pattern_filter_empty(self):
dpf = r... | [((162, 179), 'platform.system', 'platform.system', ([], {}), '()\n', (177, 179), False, 'import platform\n'), ((319, 351), 'rspub.util.resourcefilter.directory_pattern_predicate', 'rf.directory_pattern_predicate', ([], {}), '()\n', (349, 351), True, 'import rspub.util.resourcefilter as rf\n'), ((711, 748), 'rspub.util... |
fiaasco/grafana | molecule/ubuntu/tests/test_grafana.py | 6a5963e43033d88b5bb4760d47755da1069ec26b | import os
import testinfra.utils.ansible_runner
testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
def test_package(host):
""" check if packages are installed
"""
assert host.package('grafana').is_installed
def test_service(host)... | [] |
sww1235/heatsink-calc | heatsink.py | 3f28ac33b629ab5a12ddea4964f6dbe7dbc3e759 | """Class representations of heatsinks."""
import math
from scipy import constants as const
from materials import Aluminium_6063 as aluminium
class Heatsink:
"""
A Heatsink.
Extended by form factor subclasses
"""
def __init__(self, material, configuration):
"""Init material and configur... | [] |
amin-henteti/airflow-dags | scripts/get_file_name_as_variable.py | eb1e9a1a77d3c868e031cbe7420eae952ce5e767 | import inspect
def foo():
print(inspect.stack()[0][3])
foo() | [((36, 51), 'inspect.stack', 'inspect.stack', ([], {}), '()\n', (49, 51), False, 'import inspect\n')] |
sovrin-foundation/old-sovrin | sovrin/test/did/helper.py | d4e705054b7252c62fea00114060035c6eb314a4 | import base58
from plenum.common.signer_did import DidSigner
from plenum.common.verifier import DidVerifier
from plenum.common.eventually import eventually
from plenum.test.helper import assertEquality
from sovrin.common.identity import Identity
MsgForSigning = {'sender': 'Mario', 'msg': 'Lorem ipsum'}
def signMsg(... | [((433, 454), 'base58.b58decode', 'base58.b58decode', (['sig'], {}), '(sig)\n', (449, 454), False, 'import base58\n'), ((787, 812), 'plenum.common.signer_did.DidSigner', 'DidSigner', ([], {'identifier': 'idr'}), '(identifier=idr)\n', (796, 812), False, 'from plenum.common.signer_did import DidSigner\n'), ((1126, 1166),... |
jancami/edibles | tests/test_EdiblesSpectrum.py | 51263b24c5e8aef786692011289b906a810ad2f7 | import astropy
import datetime
import numpy as np
from edibles.utils.edibles_spectrum import EdiblesSpectrum
def testEdiblesSpectrum(filename="tests/HD170740_w860_redl_20140915_O12.fits"):
# Spectrum information
sp = EdiblesSpectrum(filename=filename, fully_featured=True, noDATADIR=True)
assert isinstanc... | [((228, 299), 'edibles.utils.edibles_spectrum.EdiblesSpectrum', 'EdiblesSpectrum', ([], {'filename': 'filename', 'fully_featured': '(True)', 'noDATADIR': '(True)'}), '(filename=filename, fully_featured=True, noDATADIR=True)\n', (243, 299), False, 'from edibles.utils.edibles_spectrum import EdiblesSpectrum\n'), ((1564, ... |
citrix-openstack-build/swift | swift/common/ondisk.py | 34340ddf49a84f3b3398012c2b60be1215033559 | # Copyright (c) 2010-2013 OpenStack, LLC.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to ... | [((1113, 1127), 'ConfigParser.ConfigParser', 'ConfigParser', ([], {}), '()\n', (1125, 1127), False, 'from ConfigParser import ConfigParser, NoSectionError, NoOptionError\n'), ((4965, 4981), 'swift.common.utils.listdir', 'listdir', (['devices'], {}), '(devices)\n', (4972, 4981), False, 'from swift.common.utils import li... |
KIHestad/WoT-Dossier-Parser-Create-Struct | .OLD_FILES/dossiers2_old1/custom/cache.py | 9eadeeead59b7b6cf78dc6a1e1e89fe2dffb260e | # uncompyle6 version 2.11.3
# Python bytecode 2.7 (62211)
# Decompiled from: Python 2.7.10 (default, May 23 2015, 09:40:32) [MSC v.1500 32 bit (Intel)]
# Embedded file name: scripts/common/dossiers2/custom/cache.py
import nations
from items import vehicles
def getCache():
global _g_cache
return _g_cache
def ... | [((542, 570), 'items.vehicles.getUnlocksSources', 'vehicles.getUnlocksSources', ([], {}), '()\n', (568, 570), False, 'from items import vehicles\n'), ((641, 675), 'items.vehicles.g_list.getList', 'vehicles.g_list.getList', (['nationIdx'], {}), '(nationIdx)\n', (664, 675), False, 'from items import vehicles\n'), ((1675,... |
harkiratbehl/PyGM | src/parser.py | e0a4e0b865afb607dfa0525ca386bfbe77bb6508 | #!/usr/bin/python
from code import TreeNode
from code import ThreeAddressCode
from lexer import tokens
from random import *
from symbol_table import SymbolTable
from symbol_table import SymbolTableNode
import logging
import ply.lex as lex
import ply.yacc as yacc
import sys
from codegen import convert_tac
from code... | [] |
frostburn/branch-cut-mandelbrot | render_video.py | 26c4d2db75a32b9190d40a09ebfb8a67fc4829e8 | import argparse
import imageio
import progressbar
from _routines import ffi, lib
from pylab import *
from random import Random
RESOLUTIONS = {
"2160p": (3840, 2160),
"1440p": (2560, 1440),
"1080p": (1920, 1080),
"720p": (1280, 720),
"480p": (854, 480),
"360p": (640, 360),
"240p": (426, 240)... | [((780, 825), '_routines.ffi.new', 'ffi.new', (['"""double[]"""', '(args.width * args.height)'], {}), "('double[]', args.width * args.height)\n", (787, 825), False, 'from _routines import ffi, lib\n'), ((840, 869), '_routines.ffi.new', 'ffi.new', (['"""double[]"""', 'max_iter'], {}), "('double[]', max_iter)\n", (847, 8... |
smurfix/hbmqtt | tests/mqtt/test_subscribe.py | 914440cd18b43fbe56496a73bb1259132811c539 | # Copyright (c) 2015 Nicolas JOUANIN
#
# See the file license.txt for copying permission.
import anyio
import unittest
from hbmqtt.mqtt.subscribe import SubscribePacket, SubscribePayload
from hbmqtt.mqtt.packet import PacketIdVariableHeader
from hbmqtt.mqtt.constants import QOS_1, QOS_2
from hbmqtt.adapters import Buf... | [((493, 512), 'hbmqtt.adapters.BufferAdapter', 'BufferAdapter', (['data'], {}), '(data)\n', (506, 512), False, 'from hbmqtt.adapters import BufferAdapter\n'), ((531, 577), 'anyio.run', 'anyio.run', (['SubscribePacket.from_stream', 'stream'], {}), '(SubscribePacket.from_stream, stream)\n', (540, 577), False, 'import any... |
fangd123/TextBrewer | examples/cmrc2018_example/main.trainer.py | 866f4363d9bd964f00aa60b0db5e9252a7905448 | import logging
logging.basicConfig(
format='%(asctime)s - %(levelname)s - %(name)s - %(message)s',
datefmt='%Y/%m/%d %H:%M:%S',
level=logging.INFO,
)
logger = logging.getLogger("Main")
import os,random
import numpy as np
import torch
from processing import convert_examples_to_features, read_squad_exam... | [((15, 157), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s - %(levelname)s - %(name)s - %(message)s"""', 'datefmt': '"""%Y/%m/%d %H:%M:%S"""', 'level': 'logging.INFO'}), "(format=\n '%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt=\n '%Y/%m/%d %H:%M:%S', level=logg... |
youngwoon/DnC-RL-Tensorflow | gym/gym/benchmarks/__init__.py | 02dc2750fe301a01e3bd68b1e56fc7fd754c2f3f | # EXPERIMENTAL: all may be removed soon
from gym.benchmarks import scoring
from gym.benchmarks.registration import benchmark_spec, register_benchmark, registry, register_benchmark_view # imports used elsewhere
register_benchmark(
id='Atari200M',
scorer=scoring.TotalReward(),
name='Atari200M',
view_gr... | [((264, 285), 'gym.benchmarks.scoring.TotalReward', 'scoring.TotalReward', ([], {}), '()\n', (283, 285), False, 'from gym.benchmarks import scoring\n'), ((1910, 1931), 'gym.benchmarks.scoring.TotalReward', 'scoring.TotalReward', ([], {}), '()\n', (1929, 1931), False, 'from gym.benchmarks import scoring\n'), ((3565, 358... |
pennfranc/hypnettorch | hypnettorch/data/timeseries/preprocess_audioset.py | 69d4c455028289ebe3d040af0955d909a9fef3ae | #!/usr/bin/env python3
# Copyright 2020 Benjamin Ehret
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law ... | [((2795, 2936), 'warnings.warn', 'warn', (["('The script was created for one time usage and has to be adapted when ' +\n 'reusing it. All paths specified here are absolute.')"], {}), "(\n 'The script was created for one time usage and has to be adapted when ' +\n 'reusing it. All paths specified here are absol... |
CMPUT404-Fa21-Organization/CMPUT404-Project-Social-Distribution | Posts/viewsAPI.py | 63c0ba2a03f0b462e3673ce7a4bf6bae7999440c | from django.conf import settings
from django.core import serializers
from django.utils import timezone
import requests
from Posts.commentModel import Comments
#from Posts.commentView import add_Comment
from rest_framework import status
from rest_framework.decorators import api_view, authentication_classes, permission_c... | [((4040, 4057), 'rest_framework.decorators.api_view', 'api_view', (["['GET']"], {}), "(['GET'])\n", (4048, 4057), False, 'from rest_framework.decorators import api_view, authentication_classes, permission_classes\n'), ((4060, 4106), 'rest_framework.decorators.authentication_classes', 'authentication_classes', (['[Custo... |
Open-EO/openeo-sentinelhub-python-driver | workers/tests/test_array_element.py | 92f990f098065ffb658eba6dca291dd1d5fc70f2 | import pytest
import sys, os
import xarray as xr
import numpy as np
sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
import process
from process._common import ProcessArgumentInvalid, ProcessArgumentRequired
@pytest.fixture
def generate_data():
def _construct(
data = [[[[0... | [((1307, 1469), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""data,return_nodata,index,expected_result"""', "[([9, 8, 7, 6, 5], None, 2, 7), (['A', 'B', 'C'], None, 0, 'A'), ([], True,\n 0, None)]"], {}), "('data,return_nodata,index,expected_result', [([9, 8,\n 7, 6, 5], None, 2, 7), (['A', 'B', 'C'... |
despairblue/esy-skia | gn/gn_to_bp.py | 1c81aac298602f8e872c1079db92868199b6394f | #!/usr/bin/env python
#
# Copyright 2016 Google Inc.
#
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
# Generate Android.bp for Skia from GN configuration.
import json
import os
import pprint
import string
import subprocess
import tempfile
import gn_to_bp_utils
... | [((434, 3856), 'string.Template', 'string.Template', (['"""// This file is autogenerated by gn_to_bp.py.\n\ncc_library_static {\n name: "libskia",\n cflags: [\n $cflags\n ],\n\n cppflags:[\n $cflags_cc\n ],\n\n export_include_dirs: [\n $export_includes\n ],\n\n local_include... |
firebolt55439/ray | python/ray/autoscaler/tags.py | 215300b070628c06f0106906fc6c03bd70ebf140 | """The Ray autoscaler uses tags/labels to associate metadata with instances."""
# Tag for the name of the node
TAG_RAY_NODE_NAME = "ray-node-name"
# Tag for the kind of node (e.g. Head, Worker). For legacy reasons, the tag
# value says 'type' instead of 'kind'.
TAG_RAY_NODE_KIND = "ray-node-type"
NODE_KIND_HEAD = "he... | [] |
stefan-feltmann/lands | tests/generation_test.py | b2f1fc3aab4895763160a135d085a17dceb5f58e | import unittest
from worldengine.plates import Step, center_land, world_gen
from worldengine.world import World
from tests.draw_test import TestBase
class TestGeneration(TestBase):
def setUp(self):
super(TestGeneration, self).setUp()
def test_world_gen_does_not_explode_badly(self):
# FIXME ... | [((1507, 1522), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1520, 1522), False, 'import unittest\n'), ((1143, 1210), 'worldengine.world.World.from_pickle_file', 'World.from_pickle_file', (["('%s/plates_279.world' % self.tests_data_dir)"], {}), "('%s/plates_279.world' % self.tests_data_dir)\n", (1165, 1210), Fa... |
ChenShuwei1001/mmediting | tests/test_models/test_components/test_discriminators/test_light_cnn.py | 285e629fe9da8a13c7538a6bb3347e8870cd7201 | import pytest
import torch
from mmedit.models.builder import build_component
from mmedit.models.components.discriminators.light_cnn import MaxFeature
def test_max_feature():
# cpu
conv2d = MaxFeature(16, 16, filter_type='conv2d')
x1 = torch.rand(3, 16, 16, 16)
y1 = conv2d(x1)
assert y1.shape == (... | [((200, 240), 'mmedit.models.components.discriminators.light_cnn.MaxFeature', 'MaxFeature', (['(16)', '(16)'], {'filter_type': '"""conv2d"""'}), "(16, 16, filter_type='conv2d')\n", (210, 240), False, 'from mmedit.models.components.discriminators.light_cnn import MaxFeature\n'), ((250, 275), 'torch.rand', 'torch.rand', ... |
ssichynskyi/web_metrics_posting | src/consumer.py | 26f104d2fdf31c2d029bac5a4d5337db42df86f5 | import json
import logging
from typing import Iterable
from kafka import KafkaConsumer
log = logging.getLogger(__name__)
log.addHandler(logging.NullHandler())
# I've used this example:
# https://github.com/aiven/aiven-examples/blob/master/kafka/python/consumer_example.py
# as well as Aiven Kafka tutorials
class ... | [((97, 124), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (114, 124), False, 'import logging\n'), ((140, 161), 'logging.NullHandler', 'logging.NullHandler', ([], {}), '()\n', (159, 161), False, 'import logging\n')] |
BetterManlinfeng/hyperclasspptwo | pp2_model.py | 053e9cf8445911e285ac723bdfbceb1cb384ed2e |
from tensorflow.keras import *
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers, Sequential,regularizers
from tensorflow.keras.layers import Dropout
# from tensorflow.keras import *
# 定义一个3x3卷积!kernel_initializer='he_normal','glorot_normal'
from tensorflow.python... | [((4670, 4682), 'tensorflow.keras.Sequential', 'Sequential', ([], {}), '()\n', (4680, 4682), False, 'from tensorflow.keras import layers, Sequential, regularizers\n'), ((831, 862), 'tensorflow.keras.layers.GlobalAveragePooling2D', 'layers.GlobalAveragePooling2D', ([], {}), '()\n', (860, 862), False, 'from tensorflow.ke... |
yuan-xy/medi | medi/inference/utils.py | ffa416b73eb2fbffbae7a27b6eccc267b061ed0f | """ A universal module with functions / classes without dependencies. """
import sys
import contextlib
import functools
import re
import os
from medi._compatibility import reraise
_sep = os.path.sep
if os.path.altsep is not None:
_sep += os.path.altsep
_path_re = re.compile(r'(?:\.[^{0}]+|[{0}]__init__\.py)$'.fo... | [((2299, 2320), 'functools.wraps', 'functools.wraps', (['func'], {}), '(func)\n', (2314, 2320), False, 'import functools\n'), ((325, 340), 're.escape', 're.escape', (['_sep'], {}), '(_sep)\n', (334, 340), False, 'import re\n'), ((2459, 2473), 'sys.exc_info', 'sys.exc_info', ([], {}), '()\n', (2471, 2473), False, 'impor... |
Bioconductor/bioc_git_transition | run.py | 9ca29f9e8058b755163e12bf9324ec1063d0182d | """Bioconductor run git transition code.
This module assembles the classes for the SVN --> Git transition
can be run in a sequential manner.
It runs the following aspects fo the Bioconductor transition.
Note: Update the SVN dump
1. Run Bioconductor Software package transition
2. Run Bioconductor Experiment Data pac... | [((727, 847), 'logging.basicConfig', 'logging.basicConfig', ([], {'filename': '"""transition.log"""', 'format': '"""%(levelname)s %(asctime)s %(message)s"""', 'level': 'logging.DEBUG'}), "(filename='transition.log', format=\n '%(levelname)s %(asctime)s %(message)s', level=logging.DEBUG)\n", (746, 847), False, 'impor... |
tingshao/catapult | third_party/google-endpoints/dogpile/cache/region.py | a8fe19e0c492472a8ed5710be9077e24cc517c5c | from __future__ import with_statement
from .. import Lock, NeedRegenerationException
from ..util import NameRegistry
from . import exception
from ..util import PluginLoader, memoized_property, coerce_string_conf
from .util import function_key_generator, function_multi_key_generator
from .api import NO_VALUE, CachedValu... | [((5166, 5177), 'time.time', 'time.time', ([], {}), '()\n', (5175, 5177), False, 'import time\n'), ((15944, 15960), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (15958, 15960), False, 'import threading\n'), ((22492, 22503), 'time.time', 'time.time', ([], {}), '()\n', (22501, 22503), False, 'import time\n'), ((... |
fury106/ProjectParallelProgrammeren | projectparallelprogrammeren/codesimulatie.py | fd3c198edaca5bcb19d8e665561e8cd14824e894 | # -*- coding: utf-8 -*-
"""
Module projectparallelprogrammeren.codesimulatie
=================================================================
Deze module simuleert alles.
"""
import projectparallelprogrammeren
def simulatie():
"""
Deze functie voert alle versies uit zodat deze vergeleken kunnen worden qua timi... | [((507, 529), 'importlib.import_module', 'import_module', (['version'], {}), '(version)\n', (520, 529), False, 'from importlib import import_module\n')] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.