code stringlengths 4 4.48k | docstring stringlengths 1 6.45k | _id stringlengths 24 24 |
|---|---|---|
def detectBlobs(self, col, row, globalize=False): <NEW_LINE> <INDENT> padded_tile = self.getPaddedTile(col, row) <NEW_LINE> blobs = self.detection_method(padded_tile) <NEW_LINE> pad = self.default_padding <NEW_LINE> outside = [] <NEW_LINE> for i, blob in enumerate(blobs): <NEW_LINE> <INDENT> if min(blob[0:2]) < pad or ... | Run detection and return array of detected blobs for the specified tile
:param col:
:param row:
:param globalize: if true, blobs will be returned with global coordinates
:return: | 625941bf29b78933be1e55e7 |
def get_absolute_url(self): <NEW_LINE> <INDENT> return reverse('resource_detail', args=(self.slug, )) | Get detail url. | 625941bf462c4b4f79d1d607 |
def prepare_batch_input(insts, src_pad_idx, trg_pad_idx, n_head): <NEW_LINE> <INDENT> def __pad_batch_data(insts, pad_idx, is_target=False, return_pos=True, return_attn_bias=True, return_max_len=True): <NEW_LINE> <INDENT> return_list = [] <NEW_LINE> max_len = max(len(inst) for inst in insts) <NEW_LINE> inst_data = np.a... | Pad the instances to the max sequence length in batch, and generate the
corresponding position data and attention bias. Then, convert the numpy
data to tensors and return a dict mapping names to tensors. | 625941bfa934411ee37515ca |
def get_applied_coupons(self, code, **kwargs): <NEW_LINE> <INDENT> kwargs['_return_http_data_only'] = True <NEW_LINE> if kwargs.get('callback'): <NEW_LINE> <INDENT> return self.get_applied_coupons_with_http_info(code, **kwargs) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> (data) = self.get_applied_coupons_with_http_in... | Returns a list of unique coupons which have been applied.
{ "nickname":"Retrieve used unique coupons","response":"getAppliedCodesForCode.html"}
This method makes a synchronous HTTP request by default. To make an
asynchronous HTTP request, please define a `callback` function
to be invoked when receiving the response.
>... | 625941bf283ffb24f3c5583b |
def test_modify_job_pillar(self): <NEW_LINE> <INDENT> schedule = {'foo': 'bar'} <NEW_LINE> ret = copy.deepcopy(self.schedule.opts) <NEW_LINE> ret.update({'pillar': {'schedule': {'name': {'foo': 'bar'}}}}) <NEW_LINE> self.schedule.opts.update({'pillar': {'schedule': {'name': {'foo': 'bar'}}}}) <NEW_LINE> Schedule.modify... | Tests modifying a job in the scheduler in pillar | 625941bf6e29344779a6254b |
def __init__(self): <NEW_LINE> <INDENT> self.tlock = threading.Lock() | Return a new Lock instance. | 625941bf8a349b6b435e80aa |
def new(self, item, count, customer, deadline, destination, comment): <NEW_LINE> <INDENT> time = dt.datetime.now(tz=MSK_TZ).strftime("%d.%m.%Y %H:%M:%S") <NEW_LINE> self.last_id += 1 <NEW_LINE> order_id = "{:05}".format(self.last_id) <NEW_LINE> body = { 'values': [[order_id, item[0], item[1], count, customer, time, '',... | :param item: Item code
:param count: Items count
:param customer: Customer name
:param deadline: Deadline date
:param destination: Destination school
:param comment: Purpose for order
:return: | 625941bfc432627299f04b7b |
def mock_method(name): <NEW_LINE> <INDENT> def decorator(func): <NEW_LINE> <INDENT> setattr(func, mock_method.attr_name, name) <NEW_LINE> return func <NEW_LINE> <DEDENT> setattr(decorator, mock_method.decorator_attr_name, True) <NEW_LINE> return decorator | Decorator to "register" a mock method. `name` is required.
Unlike `remote_method`, and `Mock.method`, this doesn't immediately add
functions to a `Registry` - instead, it sets an attribute on the function,
allowing such functions to be recognised when the test case is instantiated.
Example:
class MyTestCase(ZorpT... | 625941bf379a373c97cfaa7a |
def unload(self): <NEW_LINE> <INDENT> for action in self.actions: <NEW_LINE> <INDENT> self.iface.removePluginWebMenu( self.tr(u'&OSM Data'), action) <NEW_LINE> self.iface.removeToolBarIcon(action) <NEW_LINE> <DEDENT> del self.toolbar | Removes the plugin menu item and icon from QGIS GUI. | 625941bf3eb6a72ae02ec40d |
def OriginalOwnersFiles(self): <NEW_LINE> <INDENT> def owners_file_filter(f): <NEW_LINE> <INDENT> return 'OWNERS' in os.path.split(f.LocalPath())[1] <NEW_LINE> <DEDENT> files = self.AffectedFiles(file_filter=owners_file_filter) <NEW_LINE> return dict([(f.LocalPath(), f.OldContents()) for f in files]) | A map from path names of affected OWNERS files to their old content. | 625941bfd486a94d0b98e07c |
def context_switches(self): <NEW_LINE> <INDENT> return self._css | Returns a dictionary of context switches.
| 625941bf50812a4eaa59c25b |
def set_checkpoint_image(self, ckptin): <NEW_LINE> <INDENT> self.add_input_file(ckptin) <NEW_LINE> self.add_var_opt('_condor_restart', ckptin, short=True) | Adds the argument -_condor_restart for the
cases in which we want to checkpoint and grabs the
checkpoint image name from get_checkpoint_image in
the InspiralAnalysisCode section. | 625941bf6aa9bd52df036cda |
def get_protein_sequence(self, seq_dict): <NEW_LINE> <INDENT> offset = find_offset(self.exon_frames, self.strand) <NEW_LINE> cds = self.get_cds(seq_dict) <NEW_LINE> if len(cds) < 3: <NEW_LINE> <INDENT> return "" <NEW_LINE> <DEDENT> return translate_sequence(cds[offset:].upper()) | Returns the translated protein sequence for this transcript in single
character space. Overrides this function in the Transcript class to make use of frame information. | 625941bf07d97122c41787bd |
def set_background(self, background): <NEW_LINE> <INDENT> self.background = background | Set image to use for background subtraction | 625941bf5fc7496912cc38b5 |
def check_sched_params(*args): <NEW_LINE> <INDENT> hypervisor, dicts, guestname, domobj = args <NEW_LINE> sched_dict = {} <NEW_LINE> if hypervisor == "xen": <NEW_LINE> <INDENT> cmd = "xm sched-credit -d %s" % guestname <NEW_LINE> out = process.system_output(cmd, shell=True, ignore_status=True) <NEW_LINE> sched_dict = e... | Check scheduler parameters validity after setting | 625941bf16aa5153ce3623af |
def letterCombinations(self, digits): <NEW_LINE> <INDENT> if digits == "": <NEW_LINE> <INDENT> return [] <NEW_LINE> <DEDENT> mapping = {"1" : [], "2" : ["a", "b", "c"], "3" : ["d", "e", "f"], "4" : ["g", "h", "i"], "5" : ["j", "k", "l"], "6" : ["m", "n", "o"], "7" : ["p", "q", "r", "s"], "8" : ["t", "u", "v"], "9" : ["... | :type digits: str
:rtype: List[str] | 625941bfdc8b845886cb546b |
def read(self, name): <NEW_LINE> <INDENT> raise NotImplementedError | Read the contents of a file.
If a file with the given `name` does not exist then a
:exc:`storage.exc.FileNotFoundError` exception is raised.
:argument name: the name of the file to retrieve
:returns: a string representing the file contents | 625941bff8510a7c17cf9632 |
@user_blueprint.route("/follow_user", methods=["POST"]) <NEW_LINE> @jwt_required() <NEW_LINE> def follow_user(): <NEW_LINE> <INDENT> current_user = get_jwt_identity() <NEW_LINE> data = request.get_json(force=True) <NEW_LINE> unfollow = False if data["action"] == "follow" else True <NEW_LINE> status = database_client.fo... | User that is signed in will follow/unfollow other user | 625941bf3d592f4c4ed1cfab |
def test_redirect_when_account_created_to_homepage(self): <NEW_LINE> <INDENT> response = self.client.post('/register/', {'username': 'seiph', 'first_name': 'Jean', 'last_name': 'Robert', 'email': '[email protected]', 'password1': 'kevin1234', 'password2': 'kevin1234'}) <NEW_LINE> self.assertEqual(response.status_code, 302) <... | test redirection when the user create his account | 625941bf66656f66f7cbc0e1 |
def snap( self ): <NEW_LINE> <INDENT> if self.fileset == '': <NEW_LINE> <INDENT> snapname = time.strftime("%Y%m%d") + self.snap_name_separator + time.strftime("%H%M") <NEW_LINE> ( rc, cmd_out, cmd_err ) = execute_command( "/usr/lpp/mmfs/bin/mmcrsnapshot {0} {1}".format( self.gpfsdev, snapname ) ) <NEW_LINE> <DEDENT> el... | This code will create a snapshot of the specified filesystem or fileset.
NOTE: You can NOT mix filesystem and fileset snapshots on the same GPFS device.
Filesystem snapshots are named: CCYYMMDD==HHMM for easy sorting / processing.
Filesystem snapshots are named: <Fileset>==CCYYMMDD==HHMM for easy processing again. | 625941bf63d6d428bbe44426 |
def _verify_signal_packet(self, signal_packet): <NEW_LINE> <INDENT> errors.check_type(signal_packet, dict) <NEW_LINE> if len(signal_packet.keys()) > 1: <NEW_LINE> <INDENT> raise ValueError('signal_packet base-level should contain only' + ' the pipe hash identifier as key') <NEW_LINE> <DEDENT> hkey = signal_packet.keys(... | Ensure signal packet is organized properly | 625941bf3cc13d1c6d3c72b2 |
def parse_xml(xml_path): <NEW_LINE> <INDENT> with open(xml_path, 'rb') as f: <NEW_LINE> <INDENT> xml_dict = xmltodict.parse(f) <NEW_LINE> bndboxs = list() <NEW_LINE> objects = xml_dict['annotation']['object'] <NEW_LINE> if isinstance(objects, list): <NEW_LINE> <INDENT> for obj in objects: <NEW_LINE> <INDENT> obj_name =... | 解析xml文件,返回标注边界框坐标 | 625941bf4e4d5625662d4312 |
def setMidiText(ln): <NEW_LINE> <INDENT> if not ln: <NEW_LINE> <INDENT> error("Use: MidiText text") <NEW_LINE> <DEDENT> gbl.mtrks[0].addText(gbl.tickOffset, ' '.join(ln)) | Set midi text into meta track. | 625941bfd8ef3951e3243474 |
def test_generate_tablename_numeric(self): <NEW_LINE> <INDENT> foo = ANumericEnded1(foo="test") <NEW_LINE> self.assertEqual(foo.tablename, "anumericended1") | Check if table name ended by a number is not changed. | 625941bfd99f1b3c44c674cc |
def add_variable(self, name, shape, dtype=None, initializer=None, regularizer=None, trainable=True): <NEW_LINE> <INDENT> if dtype is None: <NEW_LINE> <INDENT> dtype = self.dtype <NEW_LINE> <DEDENT> existing_variables = set(tf_variables.global_variables()) <NEW_LINE> self._set_scope(None) <NEW_LINE> with vs.variable_sco... | Adds a new variable to the layer.
Arguments:
name: variable name.
shape: variable shape.
dtype: The type of the variable. Defaults to `self.dtype`.
initializer: initializer instance (callable).
regularizer: regularizer instance (callable).
trainable: whether the variable should be part of the layer's
"... | 625941bf26238365f5f0eda2 |
def run(file, *args): <NEW_LINE> <INDENT> start_hint = "" <NEW_LINE> if args: <NEW_LINE> <INDENT> start_hint = "I started it as:\n\n" + code( file + " " + " ".join(shlex.quote(a) for a in args) ) <NEW_LINE> args = ["--"] + list(args) <NEW_LINE> <DEDENT> try: <NEW_LINE> <INDENT> proc = subprocess.run( [ "python3", "-m",... | subprocess.run wrapper specialized to run Python with friendly. | 625941bf4d74a7450ccd40fa |
def prime_factors(num: int) -> Dict[Prime, int]: <NEW_LINE> <INDENT> n: Prime = 2 <NEW_LINE> result: Dict[Prime, int] = {} <NEW_LINE> while n <= num: <NEW_LINE> <INDENT> if num % n == 0: <NEW_LINE> <INDENT> num = num // n <NEW_LINE> if n in result: <NEW_LINE> <INDENT> result[n] += 1 <NEW_LINE> <DEDENT> else: <NEW_LINE>... | Creates a dictionary of prime factors for a given number.
params
num: Number to be factored
return:
Dictionary of where the keys are prime factors and the values are their respective powers.
Example:
36 = 2^2 * 3^2
>>> prime_factors(36)
{2: 2, 3: 2,} | 625941bf596a8972360899fa |
def _did_create(self, conn, resp): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> if resp.getType() == "result": <NEW_LINE> <INDENT> log.info("PUBSUB: pubsub node %s has been created" % self.nodename) <NEW_LINE> return self.recover(wait=True) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> log.error("PUBSUB: can't create p... | called after pubsub creation | 625941bf8e05c05ec3eea2a9 |
def getpath_event(): <NEW_LINE> <INDENT> ici = os.getcwd() <NEW_LINE> path_event = check_slash(os.path.realpath(ici)) <NEW_LINE> return path_event | Return the path of the event directory.
:return: path of event directory
:rtype: string | 625941bf4c3428357757c261 |
def init_handle(self): <NEW_LINE> <INDENT> self.c.setopt(pycurl.FOLLOWLOCATION, 1) <NEW_LINE> self.c.setopt(pycurl.MAXREDIRS, 10) <NEW_LINE> self.c.setopt(pycurl.CONNECTTIMEOUT, 30) <NEW_LINE> self.c.setopt(pycurl.NOSIGNAL, 1) <NEW_LINE> self.c.setopt(pycurl.NOPROGRESS, 1) <NEW_LINE> if hasattr(pycurl, "AUTOREFERER"): ... | sets common options to curl handle. | 625941bf23849d37ff7b2fc8 |
def build(): <NEW_LINE> <INDENT> includes = ['handlers', 'resources', 'templates', '*.py'] <NEW_LINE> excludes = ['test', '.*', '*.pyc', '*.pyo'] <NEW_LINE> local('rm -f dist/%s' % _TAR_FILE) <NEW_LINE> with lcd(os.path.join(_current_path(), 'www')): <NEW_LINE> <INDENT> cmd = ['tar', '--dereference', '-czvf', '../dist/... | Build dist package. | 625941bf287bf620b61d399d |
def on_update_folder(self, change): <NEW_LINE> <INDENT> folder_list = list(self.folder_sel.cur_dir.glob("*")) <NEW_LINE> folder_list = [x for x in folder_list if x.is_dir()] <NEW_LINE> folder_list = [x for x in folder_list if not re.search("\..*", x.stem)] <NEW_LINE> experiment = pd.concat( [ pd.DataFrame( pd.read_pick... | Load analysis files for all folders present in current folder | 625941bfcc0a2c11143dcdc8 |
def __init__(self, input_size, hidden_size, output_size, weight_init_std=0.01): <NEW_LINE> <INDENT> self.params = {} <NEW_LINE> self.params['W1'] = weight_init_std * np.random.randn(input_size, hidden_size) <NEW_LINE> self.params['b1'] = np.zeros(hidden_size) <NEW_LINE> self.params['W2'] = weight_init_std * np.random.r... | 初期化を行う
input_size: 入力層のニューロンの数
hidden_size: 隠れ層のニューロンの数
output_size: 出力層のニューロンの数
変数
params: ニューラルネットワークのパラメータを保持するディクショナリ変数
grads: 勾配を保持するディクショナリ変数 | 625941bf0c0af96317bb8120 |
def get_transcripts_from_youtube(youtube_id): <NEW_LINE> <INDENT> utf8_parser = etree.XMLParser(encoding='utf-8') <NEW_LINE> youtube_api = copy.deepcopy(settings.YOUTUBE_API) <NEW_LINE> youtube_api['params']['v'] = youtube_id <NEW_LINE> data = requests.get(youtube_api['url'], params=youtube_api['params']) <NEW_LINE> if... | Gets transcripts from youtube for youtube_id.
Parses only utf-8 encoded transcripts.
Other encodings are not supported at the moment.
Returns (status, transcripts): bool, dict. | 625941bf55399d3f055885eb |
@app.route('/story-form-1') <NEW_LINE> def go_form_1(): <NEW_LINE> <INDENT> return render_template("story-1-form.html") | Navigate to story 1 entry form | 625941bf50812a4eaa59c25c |
def update(self, chart=None): <NEW_LINE> <INDENT> if chart is not None: <NEW_LINE> <INDENT> self._chart = chart <NEW_LINE> self._edgelevels = [] <NEW_LINE> self._marks = {} <NEW_LINE> self._analyze() <NEW_LINE> self._grow() <NEW_LINE> self.draw() <NEW_LINE> self.erase_tree() <NEW_LINE> self._resize() <NEW_LINE> <DEDENT... | Draw any edges that have not been drawn. This is typically
called when a after modifies the canvas that a CanvasView is
displaying. C{update} will cause any edges that have been
added to the chart to be drawn.
If update is given a C{chart} argument, then it will replace
the current chart with the given chart. | 625941bfbde94217f3682d2b |
def flatten(self, root): <NEW_LINE> <INDENT> if not root: <NEW_LINE> <INDENT> return <NEW_LINE> <DEDENT> if root.right: <NEW_LINE> <INDENT> self.flatten(root.right) <NEW_LINE> <DEDENT> if root.left: <NEW_LINE> <INDENT> save = root.right <NEW_LINE> self.flatten(root.left) <NEW_LINE> p = root.left <NEW_LINE> while p.righ... | :type root: TreeNode
:rtype: void Do not return anything, modify root in-place instead. | 625941bf3c8af77a43ae36d6 |
def calculate_precision_recall_all_images( all_prediction_boxes, all_gt_boxes, iou_threshold): <NEW_LINE> <INDENT> num_tp = 0 <NEW_LINE> num_fp = 0 <NEW_LINE> num_fn = 0 <NEW_LINE> all_boxes = zip(all_prediction_boxes, all_gt_boxes) <NEW_LINE> for p, gt in all_boxes: <NEW_LINE> <INDENT> image_result = calculate_individ... | Given a set of prediction boxes and ground truth boxes for all images,
calculates recall and precision over all images.
NB: all_prediction_boxes and all_gt_boxes are not matched!
Args:
all_prediction_boxes: (list of np.array of floats): each element in the list
is a np.array containing all predic... | 625941bf0a366e3fb873e750 |
def plot_chart_ventilation_duration_distribution( values=None, path_directory=None ): <NEW_LINE> <INDENT> fonts = define_font_properties() <NEW_LINE> colors = define_color_properties() <NEW_LINE> path_file = os.path.join( path_directory, "ventilation_duration.svg" ) <NEW_LINE> figure = plot_distribution_histogram( seri... | Plots charts from the analysis process.
arguments:
values (list<float>): values
threshold (float): value of threshold for which to draw line
path_directory (str): path to directory
raises:
returns: | 625941bff548e778e58cd4b4 |
def dot_product(x, kernel): <NEW_LINE> <INDENT> return K.squeeze(K.dot(x, K.expand_dims(kernel)), axis=-1) | Wrapper for dot product operation
Args:
x (): input
kernel (): weights
Returns: | 625941bf656771135c3eb7a4 |
def _ip_2_bin_ip(self, ip=None): <NEW_LINE> <INDENT> if ip is None: <NEW_LINE> <INDENT> ip = self.ip <NEW_LINE> <DEDENT> bin_ip = ip_2_bin_ip(ip) <NEW_LINE> return bin_ip | convert IP from decimal IP to binary string
:param ip: decimal IP in a numpy array
:type ip: np.array
:return: a binary string of the IP address
:rtype: string | 625941bf97e22403b379ced1 |
def _initialize_sensors(client): <NEW_LINE> <INDENT> logging.info("Initializing current sensors") <NEW_LINE> current_sensors = [] <NEW_LINE> for address, name in zip([0x40, 0x41, 0x42, 0x42], ["left_wheel_current", "right_wheel_current", "left_flipper_current", "right_flipper_current"]): <NEW_LINE> <INDENT> try: <NEW_L... | Initialize the sensors. | 625941bfd18da76e2353240b |
def cal_linear_value(color_band, distances): <NEW_LINE> <INDENT> min_value = min(distances, key=distances.get) <NEW_LINE> max_value = max(distances, key=distances.get) <NEW_LINE> closest_values = sorted(distances, key=distances.get) <NEW_LINE> closest_color = color_band[closest_values[0]] <NEW_LINE> cloest_color_distan... | calculate linear value from color band | 625941bfcad5886f8bd26f12 |
def write(self, filename, format_='pickle'): <NEW_LINE> <INDENT> if format_ == 'pickle': <NEW_LINE> <INDENT> io.write(self, filename) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> raise KeyError("Format not supported.") | Exports Simulator object.
Parameters
----------
filename : str
Name of the Simulator object to be created.
format_ : str
Available options are 'pickle' and 'hdf5'. | 625941bff9cc0f698b140536 |
def get_all_active_memories(self, memory_list): <NEW_LINE> <INDENT> available_memories = {} <NEW_LINE> active_memory_counter = 0 <NEW_LINE> active_memory = 'ROM' if any('ROM' in mem_list for mem_list in memory_list) else 'RAM' <NEW_LINE> try: <NEW_LINE> <INDENT> cmsis_part = self._get_cmsis_part() <NEW_LINE> <DEDENT> e... | Get information of all available rom/ram memories in the form of dictionary
{Memory: [start_addr, size]}. Takes in the argument, a list of all available
regions within the ram/rom memory | 625941bf73bcbd0ca4b2bfae |
def test_simple_arith_if_2(self): <NEW_LINE> <INDENT> tra = tgis.TemporalRasterAlgebraParser(run = True, debug = True) <NEW_LINE> expr = 'R = if(A#A == 1, A - A)' <NEW_LINE> ret = tra.setup_common_granularity(expression=expr, lexer = tgis.TemporalRasterAlgebraLexer()) <NEW_LINE> self.assertEqual(ret, True) <NEW_LINE> ... | Simple arithmetic test with if condition | 625941bf63f4b57ef0001057 |
def _attach_zscores(self): <NEW_LINE> <INDENT> aucs = [] <NEW_LINE> for i, submission in enumerate(self.submissions): <NEW_LINE> <INDENT> report = json.loads(submission['substatus']['report']) <NEW_LINE> auc = report['auc'] <NEW_LINE> aucs.append(auc) <NEW_LINE> <DEDENT> ranks = np.argsort(aucs)[::-1] <NEW_LINE> for i,... | attach mean zscore, 32 individual zscores
| 625941bfc4546d3d9de7296a |
def t_BINARY(t): <NEW_LINE> <INDENT> t.value = t.value[2:] <NEW_LINE> t.value = t.value[:-1] <NEW_LINE> try: <NEW_LINE> <INDENT> t.value = int(t.value,2) <NEW_LINE> <DEDENT> except: <NEW_LINE> <INDENT> print("ERROR CONVERSION NUMERO %d", t.value) <NEW_LINE> t.value = 0 <NEW_LINE> <DEDENT> return t | b\'[01]+\' | 625941bf73bcbd0ca4b2bfaf |
def is_number(x): <NEW_LINE> <INDENT> return isinstance(x, numbers.Number) | Check whether the object provided is a number.
:param x: object to check
:return: True if x is a number | 625941bfa8ecb033257d3007 |
def update_vendor(self): <NEW_LINE> <INDENT> vnd_name = self.source.vendor.name if self.source else '' <NEW_LINE> self.vendorLine.setText(vnd_name) | Manually update the vendor name field. It isn't covered by the data mapper. | 625941bfe5267d203edcdbd8 |
def pickle_save(object_, path, **kwargs): <NEW_LINE> <INDENT> logger.info("pickle object %s to %s", object_, path) <NEW_LINE> with gzip.open(path, "wb") as file: <NEW_LINE> <INDENT> cPickle.dump(object_, file, **kwargs) | Pickle serializing with gzip. | 625941bf4a966d76dd550f45 |
def rbridge_id_snmp_server_v3host_udp_port(**kwargs): <NEW_LINE> <INDENT> config = ET.Element("config") <NEW_LINE> rbridge_id = ET.SubElement(config, "rbridge-id", xmlns="urn:brocade.com:mgmt:brocade-rbridge") <NEW_LINE> if kwargs.pop('delete_rbridge_id', False) is True: <NEW_LINE> <INDENT> delete_rbridge_id = config.f... | Auto Generated Code
| 625941bfd268445f265b4da7 |
def update_site_forward(apps, schema_editor): <NEW_LINE> <INDENT> Site = apps.get_model('sites', 'Site') <NEW_LINE> Site.objects.update_or_create( id=settings.SITE_ID, defaults={ 'domain': 'football.com', 'name': 'Football Manager' } ) | Set site domain and name. | 625941bf15baa723493c3eac |
def report_ols(formula, data, fit_regularized=False, L1_wt=1, refit=False, **kwargs): <NEW_LINE> <INDENT> RegressionResultsWrapper = statsmodels.regression.linear_model.RegressionResultsWrapper <NEW_LINE> y, X = patsy.dmatrices(formula, data=data, return_type='dataframe') <NEW_LINE> if fit_regularized: <NEW_LINE> <INDE... | Fit OLS regression, print a report, and return the fit object. | 625941bf8da39b475bd64ea9 |
def test_enableRemote(self): <NEW_LINE> <INDENT> self.assertFalse(self.protocol.enableRemote('\0')) | L{telnet.Telnet.enableRemote} should reject all options, since
L{telnet.Telnet} does not know how to implement any options. | 625941bf5e10d32532c5ee5f |
def _finish_task(self): <NEW_LINE> <INDENT> self.task_interface.Finish() <NEW_LINE> self._check_task_signals() | Finish a task. | 625941bf94891a1f4081b9e0 |
def get_error(deltas, sums, weights): <NEW_LINE> <INDENT> n, nl = sums.shape <NEW_LINE> delt = (deltas.dot(weights))*sigmoid_prime(sums) <NEW_LINE> otvet = delt.mean(axis = 0) <NEW_LINE> otvet1 = otvet.reshape(nl,1) <NEW_LINE> print(otvet1) <NEW_LINE> return otvet1 | compute error on the previous layer of network
deltas - ndarray of shape (n, n_{l+1})
sums - ndarray of shape (n, n_l)
weights - ndarray of shape (n_{l+1}, n_l)
Сигнатура: get_error(deltas, sums, weights), где deltas — ndarray формы (n, nl+1),
содержащий в i-й строке значения ошибок для i-го примера из входных данных,... | 625941bf6fece00bbac2d675 |
def IOWwriteReport(self, wdata, start=True, stop=True, addrScheme=1, msg=""): <NEW_LINE> <INDENT> start = time.time() <NEW_LINE> fncname = " {:12s}: {:15s} ".format("IOWwriteReport", msg) <NEW_LINE> lenwdata = len(wdata) <NEW_LINE> if lenwdata > 6: <NEW_LINE> <INDENT> cdprint(fncname + "Programming ERROR in IOWwrite... | report ID = 2 write a single report | 625941bf5166f23b2e1a5091 |
def parse_test_interactions(test_file): <NEW_LINE> <INDENT> has_action = re.compile(r"^\#\$\s(input|output|verify)\=(.+$)") <NEW_LINE> interactions = {} <NEW_LINE> with open(test_file, 'r') as file: <NEW_LINE> <INDENT> for line_no, line in enumerate(file.readlines(), start=1): <NEW_LINE> <INDENT> check_line = has_actio... | Method to parse a test file, and return the necessary
interaction information.
:param: str test_file: Path to the file to parse.
Test interactions are annotated as comments in the test file.
Interaction markups should be placed one line above the expected
interaction, and must consist of only a single line. The c... | 625941bfa219f33f346288a5 |
def getAllMachines(self, ctx): <NEW_LINE> <INDENT> ret = ComputerManager().getComputersList(ctx) <NEW_LINE> return [Machine(ret[m][1]) for m in ret] | return all declared machines | 625941bfa4f1c619b28aff77 |
def auth(func): <NEW_LINE> <INDENT> @wraps(func) <NEW_LINE> def wapper(who: User, what: Target, check: bool = False): <NEW_LINE> <INDENT> user_roles = who.get_roles() <NEW_LINE> work_roles = [] <NEW_LINE> for role in user_roles: <NEW_LINE> <INDENT> parent_roles = what.get_parent_roles() <NEW_LINE> if OriginalTarget in ... | 验证 | 625941bff7d966606f6a9f3a |
def voltage_sense_configuration(self, auto_range = True, manual_range = 21, NPLCs = 10): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> if 'KEITHLEY' in self.identity and '2400' in self.identity: <NEW_LINE> <INDENT> buffer = ':SENS:FUNC:CONC ON;:SENS:FUNC:ALL;' <NEW_LINE> if auto_range: <NEW_LINE> <INDENT> buffer += ':SE... | This function is used to configure the voltage measurement.
Parameters:
- auto_range = boolean parameter that enables (true) or disables auto voltage ranging (i.e. the instrument is able or not to auto-determine the
most accurate measurement range).
- manual_range = if auto ranging is disabled,the user... | 625941bfec188e330fd5a6dc |
def parse_config(self): <NEW_LINE> <INDENT> self.configraw = [] <NEW_LINE> try: <NEW_LINE> <INDENT> with open(self.configfile) as f: <NEW_LINE> <INDENT> for line in f: <NEW_LINE> <INDENT> linetrim = line.strip() <NEW_LINE> if linetrim and (not linetrim.startswith('#')): <NEW_LINE> <INDENT> self.configraw.append(linetri... | Parse the config file contents, if available.
Logs errors, sets:
self.banpatterns (via self.build_patterns()) | 625941bfb7558d58953c4e51 |
def visualize_gui_smplpose_basic(self, smpl, poseParamList, shapeParamList =None, colorList = None, isRotMat = False, scalingFactor=300, waittime =1): <NEW_LINE> <INDENT> zero_betas = torch.from_numpy(np.zeros( (1,10), dtype=np.float32)) <NEW_LINE> default_color = glViewer.g_colorSet['eft'] <NEW_LINE> meshList =[] <NE... | Visualize SMPL vertices from SMPL pose parameters
This can be used as a quick visualize function if you have a pose parameters
args:
poseParamList: list of pose parameters (numpy array) in angle axis (72,) by default or rot matrix (24,3,3) with isRotMat==True
shapeParamList: (optional) list of shape paramete... | 625941bf82261d6c526ab3d4 |
def checksum1(data, stringlength): <NEW_LINE> <INDENT> value_buffer = 0 <NEW_LINE> for count in range(0, stringlength): <NEW_LINE> <INDENT> value_buffer = value_buffer ^ data[count] <NEW_LINE> <DEDENT> return value_buffer&0xFE | Calculate Checksum 1
Calculate the ckecksum 1 required for the herkulex data packet
Args:
data (list): the data of which checksum is to be calculated
stringlength (int): the length of the data
Returns:
int: The calculated checksum 1 | 625941bfadb09d7d5db6c6ca |
def parse(document): <NEW_LINE> <INDENT> if isinstance(document, str): <NEW_LINE> <INDENT> document = Path(document) <NEW_LINE> <DEDENT> with document.open('r') as fin: <NEW_LINE> <INDENT> lines = [__.strip() for __ in fin.readlines() if __.strip()][1:] <NEW_LINE> content = ''.join(lines) <NEW_LINE> return bs(content, ... | Return a BeautifulSoup XML object or None if no document specified
Parameter
---------
document : str, Path
Path to XML document for parsing | 625941bfa934411ee37515cb |
def __setitem__(self, key: Union[int, np.ndarray], value: Any) -> None: <NEW_LINE> <INDENT> raise NotImplementedError(f"{type(self)} does not implement __setitem__.") | Set one or more values inplace.
This method is not required to satisfy the pandas extension array
interface.
Parameters
----------
key : int, ndarray, or slice
When called from, e.g. ``Series.__setitem__``, ``key`` will be
one of
* scalar int
* ndarray of integers.
* boolean ndarray
* slice o... | 625941bffbf16365ca6f60f7 |
@app.route('/end_session') <NEW_LINE> def end_session(): <NEW_LINE> <INDENT> session.clear() <NEW_LINE> return render_template("home.html") | End session. | 625941bf009cb60464c632ec |
def p_table_wild(p): <NEW_LINE> <INDENT> p[0] = ("*", p[1]) | table_wild : identifier '.' '*' | 625941bf7c178a314d6ef394 |
def set_input(self, deepmag, deepmag_sigma=None, *args, **kwargs): <NEW_LINE> <INDENT> filter_list = DEEPFilterList() <NEW_LINE> maggie, maggie_ivar = deepmag_to_maggie(deepmag, deepmag_sigma) <NEW_LINE> super(DEEPPhotoZ, self).set_input(filter_list, maggie, maggie_ivar, *args, **kwargs) | Set input data
The input DEEP BRI photometry will be converted into AB
maggies via DEEP.deepmag_to_maggie; see the documentation of
the function for detail.
INPUT
redshift -- List of redshifts
deepmag -- List of DEEP BRI magnitudes
deepmag_sigma -- List of 1-sigma uncertainties in DEEP BRI
magnitude... | 625941bf94891a1f4081b9e1 |
def get_extant_genome_by_name(self, name): <NEW_LINE> <INDENT> for taxon in self.taxonomy.leaves: <NEW_LINE> <INDENT> if taxon.name == name: <NEW_LINE> <INDENT> if "genome" in taxon.features: <NEW_LINE> <INDENT> return taxon.genome <NEW_LINE> <DEDENT> <DEDENT> <DEDENT> raise KeyError('No extant genomes match the query ... | Get the :obj:`pyham.genome.ExtantGenome` that match the query name.
Args:
name (:obj:`str`): Name of the :obj:`pyham.genome.ExtantGenome`.
Returns:
:obj:`pyham.genome.ExtantGenome` or raise KeyError | 625941bf2ae34c7f2600d06a |
def lstm_syndir_attention_decoder(inputs, hparams, train, name, initial_state, encoder_outputs, att_scores): <NEW_LINE> <INDENT> def dropout_lstm_cell(): <NEW_LINE> <INDENT> return tf.contrib.rnn.DropoutWrapper( tf.nn.rnn_cell.BasicLSTMCell(hparams.hidden_size), input_keep_prob=1.0 - hparams.dropout * tf.to_float(train... | Run LSTM cell with attention on inputs of shape [batch x time x size]. | 625941bfff9c53063f47c12d |
def getThread(self): <NEW_LINE> <INDENT> return [message.getDetail() for message in self.message_set.order_by('time_sent')] | Returns list of most recent messages with corresponding info | 625941bfb545ff76a8913d4f |
@follows(filtering, peakcalling_tasks, IDR, buildBigWig) <NEW_LINE> def full(): <NEW_LINE> <INDENT> pass | runs entire pipeline | 625941bf01c39578d7e74d74 |
def get_year(self): <NEW_LINE> <INDENT> return self.date_to.year | # Helper method to get the year (normalized between Odoo Versions)
:return: int year of payslip | 625941bf0a50d4780f666dc9 |
def applyAngularConversion(angularCalibrationFile, detectorPosition, channeldata): <NEW_LINE> <INDENT> if angularCalibrationFile is None: <NEW_LINE> <INDENT> raise Exception("Angular calibration file is not available for converting channel numbers to angles.") <NEW_LINE> <DEDENT> if detectorPosition is None: <NEW_LINE>... | convert channel number to actual angle in degree based on angular calibration file provided, return a list of tuples (angle, count, error, channel)
required input parameters:
1. angular calibration file, if None raise exception
2. detector position at which the raw data is taken, if None raise exception
3. ... | 625941bfbe7bc26dc91cd53d |
def predict(self, X): <NEW_LINE> <INDENT> check_is_fitted(self, 'threshold_') <NEW_LINE> X = check_array(X) <NEW_LINE> is_inlier = -np.ones(X.shape[0], dtype=int) <NEW_LINE> if self.contamination is not None: <NEW_LINE> <INDENT> values = self.decision_function(X, raw_values=True) <NEW_LINE> is_inlier[values <= self.thr... | Outlyingness of observations in X according to the fitted model.
Parameters
----------
X : array-like, shape = (n_samples, n_features)
Returns
-------
is_outliers : array, shape = (n_samples, ), dtype = bool
For each observation, tells whether or not it should be considered
as an outlier according to the fitt... | 625941bf67a9b606de4a7df4 |
def swappableDependency(value): <NEW_LINE> <INDENT> return SwappableTuple((value.split(".", 1)[0], "__first__"), value) | Turns a setting value into a dependency. | 625941bf004d5f362079a26e |
def get_file_path(self): <NEW_LINE> <INDENT> file_list = os.listdir(self.apk_path) <NEW_LINE> for file in file_list: <NEW_LINE> <INDENT> if self.style in file: <NEW_LINE> <INDENT> file_path = os.path.join(self.apk_path, file) <NEW_LINE> os.rename(file_path, file_path.replace(' ', '')) <NEW_LINE> time.sleep(1) <NEW_LINE... | 获取安装包路径 | 625941bf85dfad0860c3ad92 |
def is_kind_of_class(obj, a_class): <NEW_LINE> <INDENT> return isinstance(obj, a_class) | is_kind_of_class method checks inheritance of obj | 625941bf442bda511e8be355 |
def decode(bytes): <NEW_LINE> <INDENT> return bytes.decode('utf-8') | Decodes a given list of bytes as UTF-8.
@type bytes: list of bytes
@rtype: string
@return: A decoded version of the input bytes. | 625941bf4f6381625f114975 |
def set_qsmiles(self,qsmilesmat,qcountsmat,querylengths,querymags): <NEW_LINE> <INDENT> tempqlmat = self._padded_array(qsmilesmat.T) <NEW_LINE> if tempqlmat.shape[1] > 65536 or tempqlmat.shape[0] > 32768: <NEW_LINE> <INDENT> raise ValueError("Error: query matrix is not allowed to have more than 65536 rows (molecules) o... | Sets the reference SMILES set to use Lingo matrix *qsmilesmat*, count matrix *qcountsmat*,
and length vector *querylengths*. If *querymags* is provided, it will be used as the magnitude
vector; else, the magnitude vector will be computed (on the GPU) from the count matrix.
Because of hardware limitations, the query ma... | 625941bffff4ab517eb2f373 |
def __init__(self, fnam, histnam='h_cor'): <NEW_LINE> <INDENT> self.file = TFile(fnam) <NEW_LINE> if self.file.IsZombie(): <NEW_LINE> <INDENT> raise ValueError(fnam+' cannot be opened') <NEW_LINE> <DEDENT> self.hist = self.file.Get(histnam) <NEW_LINE> if self.hist==None: <NEW_LINE> <INDENT> raise ValueError('{h} cannot... | fnam is a root file containing a 1D histogram giving
the correction factor as a function of eta. | 625941bf50812a4eaa59c25d |
def clean(self): <NEW_LINE> <INDENT> super(SignedSSHKey, self).clean() <NEW_LINE> if not self.certificate.startswith('[email protected]'): <NEW_LINE> <INDENT> raise BadRequestError("Certificate is not a valid signed RSA key.") | # Checks if certificate is specific ssh-rsa-cert
and ensures that self is a valid RSA keypair. | 625941bfab23a570cc2500b9 |
def test(self, message): <NEW_LINE> <INDENT> return super().test(message.channel) and self.pattern.search(message.content) | Test whether or not the given message should trigger the Event's execution. | 625941bf1f037a2d8b946138 |
def load_asdf(self, filename, mode="a"): <NEW_LINE> <INDENT> if self.mpi_mode: <NEW_LINE> <INDENT> return ASDFDataSet(filename, compression=None, debug=self._debug, mode=mode) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> return ASDFDataSet(filename, mode=mode) | Load asdf file
:param filename:
:param mode:
:return: | 625941bfb5575c28eb68df37 |
def get_ip(): <NEW_LINE> <INDENT> s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) <NEW_LINE> try: <NEW_LINE> <INDENT> s.connect(('10.255.255.255', 1)) <NEW_LINE> myip = s.getsockname()[0] <NEW_LINE> <DEDENT> except BaseException: <NEW_LINE> <INDENT> myip = '127.0.0.1' <NEW_LINE> <DEDENT> finally: <NEW_LINE> <INDEN... | Return IP address of current machine | 625941bf507cdc57c6306c0e |
def _stdout_chunk_read(self, chunk_bytes): <NEW_LINE> <INDENT> self._stdout_bytes += chunk_bytes <NEW_LINE> self._stdout.write(chunk_bytes) | See ExternalCmdJob._stdout_chunk_read(). Overridden to accumulate stdout in a member variable, so
that it can be accessed from memory instead of having to be re-read from the file ExternalCmdJob saves
it to. | 625941bf435de62698dfdb85 |
def _set_callbacks(self): <NEW_LINE> <INDENT> self.callbacks = LoadBalancerCallbacks(self.plugin, self.env, self.pool_scheduler) <NEW_LINE> topic = lbaasv1constants.TOPIC_PROCESS_ON_HOST <NEW_LINE> if self.env: <NEW_LINE> <INDENT> topic = topic + "_" + self.env <NEW_LINE> <DEDENT> if PREJUNO: <NEW_LINE> <INDENT> self.c... | Setup callbacks to receive calls from agent | 625941bf379a373c97cfaa7d |
def exam_cmp(exam1, exam2): <NEW_LINE> <INDENT> date1 = str2DateTime(exam1["date"]) <NEW_LINE> date2 = str2DateTime(exam2["date"]) <NEW_LINE> if date1 < date2: <NEW_LINE> <INDENT> return -1 <NEW_LINE> <DEDENT> if date1 == date2: <NEW_LINE> <INDENT> return 0 <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> return 1 | used for sorting exams by date. | 625941bfa4f1c619b28aff78 |
def user_followers_like_follow_helper(data, user_id, users_count=25, likes_count=10, thresholds=(0.7, 1.2), sleep_time=1): <NEW_LINE> <INDENT> result = {} <NEW_LINE> api = data['api'] <NEW_LINE> followers = api.getTotalFollowers(user_id) <NEW_LINE> users_count = users_count if users_count < len(followers) else ( len(fo... | Helper function. It gets user followers list. Then choice
some users and like and follow them.
:param :data: data dictionary with 'api' and 'follows' keys
:param :user_id: user id. His followers function will like and follow
:param :user_count: how much users function should like/follow
:param :likes_count: how much g... | 625941bf4f88993c3716bfa4 |
def test_get_members_by_member(self): <NEW_LINE> <INDENT> new_user = User(username='another_user', fullname='Another user', passhash='hash') <NEW_LINE> server.db.session.add(new_user) <NEW_LINE> new_user.groups.append(self.group) <NEW_LINE> server.db.session.commit() <NEW_LINE> new_user.get_token() <NEW_LINE> url = '/g... | Non admin user requests the list of group members. | 625941bf23849d37ff7b2fc9 |
def creatTaskCpnDeclaration(self): <NEW_LINE> <INDENT> comments = ('/***********************************************************\n') <NEW_LINE> comments += ('* 组件运行任务\n') <NEW_LINE> comments += ('***********************************************************/\n') <NEW_LINE> comments += ('/*运行任务声明\n') <NEW_LINE> comments +... | 组件运行任务 | 625941bf566aa707497f44a6 |
def find_blobs(img_data, verbose=False, min_area=None): <NEW_LINE> <INDENT> blobs = [] <NEW_LINE> copy_data = img_data.copy() <NEW_LINE> if img_data.dtype == 'uint16': <NEW_LINE> <INDENT> if verbose: print("16 bit image found. Scaling down to 8bit.") <NEW_LINE> copy_data = (copy_data/2.**8).astype(np.uint8) <NEW_LINE> ... | Find second level contours in 16bit images
Here is an example to illustrate the contours of the blob this function can find:
.. image:: _static/blobs-contours.png | 625941bf236d856c2ad44710 |
def __init__(__self__, *, application_name: Optional[pulumi.Input[str]] = None, application_version_id: Optional[pulumi.Input[int]] = None, snapshot_creation_timestamp: Optional[pulumi.Input[str]] = None, snapshot_name: Optional[pulumi.Input[str]] = None): <NEW_LINE> <INDENT> if application_name is not None: <NEW_LINE>... | Input properties used for looking up and filtering ApplicationSnapshot resources.
:param pulumi.Input[str] application_name: The name of an existing Kinesis Analytics v2 Application. Note that the application must be running for a snapshot to be created.
:param pulumi.Input[int] application_version_id: The current app... | 625941bfac7a0e7691ed400a |
def test_get_yearly_with_gains(): <NEW_LINE> <INDENT> totals = { 2014: {"contributions": 1000, "transfers": 0, "value": 5000.0}, 2013: {"contributions": 500, "transfers": 500, "value": 1000.0}, 2015: {"contributions": 0, "transfers": 0, "value": 5000.0}, 2016: {"contributions": 3000.0, "transfers": -1000, "value": 4000... | get_yearly_with_gains should produce a sorted list of Years | 625941bf92d797404e3040c2 |
def main_wrap(args=None): <NEW_LINE> <INDENT> step_info = { 'title': '', 'exit_code': 0 } <NEW_LINE> def step_func(title=None, exit_code=None): <NEW_LINE> <INDENT> if title is not None: <NEW_LINE> <INDENT> step_info['title'] = title <NEW_LINE> <DEDENT> if exit_code is not None: <NEW_LINE> <INDENT> step_info['exit_code'... | The main function that provides exception handling.
Call "main_core" to implement the core functionality.
@param args: Command arguments list.
@return: Exit code. | 625941bf1d351010ab855a56 |
def _run_exit_func(self, additional_args: List) -> bool: <NEW_LINE> <INDENT> def _exit_func(): <NEW_LINE> <INDENT> with TimeIt(self.worker_insights.worker_exit_time, self.worker_id): <NEW_LINE> <INDENT> return self.map_params.worker_exit(*additional_args) <NEW_LINE> <DEDENT> <DEDENT> results, should_return = self._run_... | Runs the exit function when provided and stores its results.
:param additional_args: Additional args to pass to the function (worker ID, shared objects, worker state)
:return: True when the worker needs to shut down, False otherwise | 625941bf5510c4643540f323 |
def restore(destination_ip, source_ip): <NEW_LINE> <INDENT> destination_mac = get_mac(destination_ip) <NEW_LINE> source_mac = get_mac(source_ip) <NEW_LINE> packet = scapy.ARP(op=2, pdst=destination_ip, hwdst=destination_mac, psrc=source_ip, hwsrc=source_mac) <NEW_LINE> scapy.send(packet, count=4, verbose=False) | Restore all MAC addresses on program exit. | 625941bf5fdd1c0f98dc016b |
def geo_lng_qq(self): <NEW_LINE> <INDENT> return self.geo_qq[1] | 返回 QQ 坐标系经度 | 625941bf21bff66bcd68488e |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.