repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1
value | code stringlengths 75 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 | partition stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
CybOXProject/mixbox | mixbox/entities.py | Entity.from_json | def from_json(cls, json_doc):
"""Parse a JSON string and build an entity."""
try:
d = json.load(json_doc)
except AttributeError: # catch the read() error
d = json.loads(json_doc)
return cls.from_dict(d) | python | def from_json(cls, json_doc):
"""Parse a JSON string and build an entity."""
try:
d = json.load(json_doc)
except AttributeError: # catch the read() error
d = json.loads(json_doc)
return cls.from_dict(d) | [
"def",
"from_json",
"(",
"cls",
",",
"json_doc",
")",
":",
"try",
":",
"d",
"=",
"json",
".",
"load",
"(",
"json_doc",
")",
"except",
"AttributeError",
":",
"# catch the read() error",
"d",
"=",
"json",
".",
"loads",
"(",
"json_doc",
")",
"return",
"cls"... | Parse a JSON string and build an entity. | [
"Parse",
"a",
"JSON",
"string",
"and",
"build",
"an",
"entity",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/entities.py#L490-L497 | train |
CybOXProject/mixbox | mixbox/entities.py | EntityList._multiple_field | def _multiple_field(cls):
"""Return the "multiple" TypedField associated with this EntityList.
This also lazily sets the ``_entitylist_multiplefield`` value if it
hasn't been set yet. This is set to a tuple containing one item because
if we set the class attribute to the TypedField, we ... | python | def _multiple_field(cls):
"""Return the "multiple" TypedField associated with this EntityList.
This also lazily sets the ``_entitylist_multiplefield`` value if it
hasn't been set yet. This is set to a tuple containing one item because
if we set the class attribute to the TypedField, we ... | [
"def",
"_multiple_field",
"(",
"cls",
")",
":",
"klassdict",
"=",
"cls",
".",
"__dict__",
"try",
":",
"# Checking for cls.entitylist_multifield would return any inherited",
"# values, so we check the class __dict__ explicitly.",
"return",
"klassdict",
"[",
"\"_entitylist_multifie... | Return the "multiple" TypedField associated with this EntityList.
This also lazily sets the ``_entitylist_multiplefield`` value if it
hasn't been set yet. This is set to a tuple containing one item because
if we set the class attribute to the TypedField, we would effectively
add a Typed... | [
"Return",
"the",
"multiple",
"TypedField",
"associated",
"with",
"this",
"EntityList",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/entities.py#L594-L628 | train |
CybOXProject/mixbox | mixbox/entities.py | NamespaceCollector._finalize_namespaces | def _finalize_namespaces(self, ns_dict=None):
"""Returns a dictionary of namespaces to be exported with an XML
document.
This loops over all the namespaces that were discovered and built
during the execution of ``collect()`` and
``_parse_collected_classes()`` and attempts to mer... | python | def _finalize_namespaces(self, ns_dict=None):
"""Returns a dictionary of namespaces to be exported with an XML
document.
This loops over all the namespaces that were discovered and built
during the execution of ``collect()`` and
``_parse_collected_classes()`` and attempts to mer... | [
"def",
"_finalize_namespaces",
"(",
"self",
",",
"ns_dict",
"=",
"None",
")",
":",
"if",
"ns_dict",
":",
"# Add the user's entries to our set",
"for",
"ns",
",",
"alias",
"in",
"six",
".",
"iteritems",
"(",
"ns_dict",
")",
":",
"self",
".",
"_collected_namespa... | Returns a dictionary of namespaces to be exported with an XML
document.
This loops over all the namespaces that were discovered and built
during the execution of ``collect()`` and
``_parse_collected_classes()`` and attempts to merge them all.
Raises:
.namespaces.Dup... | [
"Returns",
"a",
"dictionary",
"of",
"namespaces",
"to",
"be",
"exported",
"with",
"an",
"XML",
"document",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/entities.py#L796-L856 | train |
inveniosoftware/invenio-communities | invenio_communities/views/api.py | CommunitiesResource.get | def get(self, query, sort, page, size):
"""Get a list of all the communities.
.. http:get:: /communities/(string:id)
Returns a JSON list with all the communities.
**Request**:
.. sourcecode:: http
GET /communities HTTP/1.1
Accept: appl... | python | def get(self, query, sort, page, size):
"""Get a list of all the communities.
.. http:get:: /communities/(string:id)
Returns a JSON list with all the communities.
**Request**:
.. sourcecode:: http
GET /communities HTTP/1.1
Accept: appl... | [
"def",
"get",
"(",
"self",
",",
"query",
",",
"sort",
",",
"page",
",",
"size",
")",
":",
"urlkwargs",
"=",
"{",
"'q'",
":",
"query",
",",
"'sort'",
":",
"sort",
",",
"'size'",
":",
"size",
",",
"}",
"communities",
"=",
"Community",
".",
"filter_co... | Get a list of all the communities.
.. http:get:: /communities/(string:id)
Returns a JSON list with all the communities.
**Request**:
.. sourcecode:: http
GET /communities HTTP/1.1
Accept: application/json
Content-Type: applicat... | [
"Get",
"a",
"list",
"of",
"all",
"the",
"communities",
"."
] | 5c4de6783724d276ae1b6dd13a399a9e22fadc7a | https://github.com/inveniosoftware/invenio-communities/blob/5c4de6783724d276ae1b6dd13a399a9e22fadc7a/invenio_communities/views/api.py#L80-L129 | train |
inveniosoftware/invenio-communities | invenio_communities/views/api.py | CommunityDetailsResource.get | def get(self, community_id):
"""Get the details of the specified community.
.. http:get:: /communities/(string:id)
Returns a JSON dictionary with the details of the specified
community.
**Request**:
.. sourcecode:: http
GET /communities/co... | python | def get(self, community_id):
"""Get the details of the specified community.
.. http:get:: /communities/(string:id)
Returns a JSON dictionary with the details of the specified
community.
**Request**:
.. sourcecode:: http
GET /communities/co... | [
"def",
"get",
"(",
"self",
",",
"community_id",
")",
":",
"community",
"=",
"Community",
".",
"get",
"(",
"community_id",
")",
"if",
"not",
"community",
":",
"abort",
"(",
"404",
")",
"etag",
"=",
"community",
".",
"version_id",
"self",
".",
"check_etag"... | Get the details of the specified community.
.. http:get:: /communities/(string:id)
Returns a JSON dictionary with the details of the specified
community.
**Request**:
.. sourcecode:: http
GET /communities/communities/comm1 HTTP/1.1
... | [
"Get",
"the",
"details",
"of",
"the",
"specified",
"community",
"."
] | 5c4de6783724d276ae1b6dd13a399a9e22fadc7a | https://github.com/inveniosoftware/invenio-communities/blob/5c4de6783724d276ae1b6dd13a399a9e22fadc7a/invenio_communities/views/api.py#L143-L185 | train |
OpenTreeOfLife/peyotl | peyotl/phylesystem/phylesystem_umbrella.py | Phylesystem | def Phylesystem(repos_dict=None,
repos_par=None,
with_caching=True,
repo_nexml2json=None,
git_ssh=None,
pkey=None,
git_action_class=PhylesystemGitAction,
mirror_info=None,
new_study_prefix=Non... | python | def Phylesystem(repos_dict=None,
repos_par=None,
with_caching=True,
repo_nexml2json=None,
git_ssh=None,
pkey=None,
git_action_class=PhylesystemGitAction,
mirror_info=None,
new_study_prefix=Non... | [
"def",
"Phylesystem",
"(",
"repos_dict",
"=",
"None",
",",
"repos_par",
"=",
"None",
",",
"with_caching",
"=",
"True",
",",
"repo_nexml2json",
"=",
"None",
",",
"git_ssh",
"=",
"None",
",",
"pkey",
"=",
"None",
",",
"git_action_class",
"=",
"PhylesystemGitAc... | Factory function for a _Phylesystem object.
A wrapper around the _Phylesystem class instantiation for
the most common use case: a singleton _Phylesystem.
If you need distinct _Phylesystem objects, you'll need to
call that class directly. | [
"Factory",
"function",
"for",
"a",
"_Phylesystem",
"object",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/peyotl/phylesystem/phylesystem_umbrella.py#L273-L304 | train |
SpotlightData/preprocessing | preprocessing/text.py | convert_html_entities | def convert_html_entities(text_string):
'''
Converts HTML5 character references within text_string to their corresponding unicode characters
and returns converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-str... | python | def convert_html_entities(text_string):
'''
Converts HTML5 character references within text_string to their corresponding unicode characters
and returns converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-str... | [
"def",
"convert_html_entities",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"html",
".",
"unescape",
"(",... | Converts HTML5 character references within text_string to their corresponding unicode characters
and returns converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Converts",
"HTML5",
"character",
"references",
"within",
"text_string",
"to",
"their",
"corresponding",
"unicode",
"characters",
"and",
"returns",
"converted",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L33-L51 | train |
SpotlightData/preprocessing | preprocessing/text.py | convert_ligatures | def convert_ligatures(text_string):
'''
Coverts Latin character references within text_string to their corresponding unicode characters
and returns converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or No... | python | def convert_ligatures(text_string):
'''
Coverts Latin character references within text_string to their corresponding unicode characters
and returns converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or No... | [
"def",
"convert_ligatures",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"for",
"i",
"in",
"range",
"(",
"0",
",... | Coverts Latin character references within text_string to their corresponding unicode characters
and returns converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or NoneType not be passed as an argument | [
"Coverts",
"Latin",
"character",
"references",
"within",
"text_string",
"to",
"their",
"corresponding",
"unicode",
"characters",
"and",
"returns",
"converted",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L53-L73 | train |
SpotlightData/preprocessing | preprocessing/text.py | correct_spelling | def correct_spelling(text_string):
'''
Splits string and converts words not found within a pre-built dictionary to their
most likely actual word based on a relative probability dictionary. Returns edited
string as type str.
Keyword argument:
- text_string: string instance
Exceptions raise... | python | def correct_spelling(text_string):
'''
Splits string and converts words not found within a pre-built dictionary to their
most likely actual word based on a relative probability dictionary. Returns edited
string as type str.
Keyword argument:
- text_string: string instance
Exceptions raise... | [
"def",
"correct_spelling",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"word_list",
"=",
"text_string",
".",
"split... | Splits string and converts words not found within a pre-built dictionary to their
most likely actual word based on a relative probability dictionary. Returns edited
string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or... | [
"Splits",
"string",
"and",
"converts",
"words",
"not",
"found",
"within",
"a",
"pre",
"-",
"built",
"dictionary",
"to",
"their",
"most",
"likely",
"actual",
"word",
"based",
"on",
"a",
"relative",
"probability",
"dictionary",
".",
"Returns",
"edited",
"string"... | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L75-L98 | train |
SpotlightData/preprocessing | preprocessing/text.py | create_sentence_list | def create_sentence_list(text_string):
'''
Splits text_string into a list of sentences based on NLTK's english.pickle tokenizer, and
returns said list as type list of str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argumen... | python | def create_sentence_list(text_string):
'''
Splits text_string into a list of sentences based on NLTK's english.pickle tokenizer, and
returns said list as type list of str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argumen... | [
"def",
"create_sentence_list",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"[",
"]",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"SENTENCE_TOKENIZER",
".",
"... | Splits text_string into a list of sentences based on NLTK's english.pickle tokenizer, and
returns said list as type list of str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Splits",
"text_string",
"into",
"a",
"list",
"of",
"sentences",
"based",
"on",
"NLTK",
"s",
"english",
".",
"pickle",
"tokenizer",
"and",
"returns",
"said",
"list",
"as",
"type",
"list",
"of",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L100-L118 | train |
SpotlightData/preprocessing | preprocessing/text.py | keyword_tokenize | def keyword_tokenize(text_string):
'''
Extracts keywords from text_string using NLTK's list of English stopwords, ignoring words of a
length smaller than 3, and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs sh... | python | def keyword_tokenize(text_string):
'''
Extracts keywords from text_string using NLTK's list of English stopwords, ignoring words of a
length smaller than 3, and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs sh... | [
"def",
"keyword_tokenize",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"\" \"",
".",
"join",
"(",
"[",
... | Extracts keywords from text_string using NLTK's list of English stopwords, ignoring words of a
length smaller than 3, and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Extracts",
"keywords",
"from",
"text_string",
"using",
"NLTK",
"s",
"list",
"of",
"English",
"stopwords",
"ignoring",
"words",
"of",
"a",
"length",
"smaller",
"than",
"3",
"and",
"returns",
"the",
"new",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L120-L138 | train |
SpotlightData/preprocessing | preprocessing/text.py | lemmatize | def lemmatize(text_string):
'''
Returns base from of text_string using NLTK's WordNetLemmatizer as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None ... | python | def lemmatize(text_string):
'''
Returns base from of text_string using NLTK's WordNetLemmatizer as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None ... | [
"def",
"lemmatize",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"LEMMATIZER",
".",
"lemmatize",
"(",
"t... | Returns base from of text_string using NLTK's WordNetLemmatizer as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Returns",
"base",
"from",
"of",
"text_string",
"using",
"NLTK",
"s",
"WordNetLemmatizer",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L140-L157 | train |
SpotlightData/preprocessing | preprocessing/text.py | lowercase | def lowercase(text_string):
'''
Converts text_string into lowercase and returns the converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None or text_string ... | python | def lowercase(text_string):
'''
Converts text_string into lowercase and returns the converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None or text_string ... | [
"def",
"lowercase",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"text_string",
".",
"lower",
"(",
")",
... | Converts text_string into lowercase and returns the converted string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Converts",
"text_string",
"into",
"lowercase",
"and",
"returns",
"the",
"converted",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L159-L176 | train |
SpotlightData/preprocessing | preprocessing/text.py | preprocess_text | def preprocess_text(text_string, function_list):
'''
Given each function within function_list, applies the order of functions put forward onto
text_string, returning the processed string as type str.
Keyword argument:
- function_list: list of functions available in preprocessing.text
- text_st... | python | def preprocess_text(text_string, function_list):
'''
Given each function within function_list, applies the order of functions put forward onto
text_string, returning the processed string as type str.
Keyword argument:
- function_list: list of functions available in preprocessing.text
- text_st... | [
"def",
"preprocess_text",
"(",
"text_string",
",",
"function_list",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"if",
"isinstance",
"("... | Given each function within function_list, applies the order of functions put forward onto
text_string, returning the processed string as type str.
Keyword argument:
- function_list: list of functions available in preprocessing.text
- text_string: string instance
Exceptions raised:
- Func... | [
"Given",
"each",
"function",
"within",
"function_list",
"applies",
"the",
"order",
"of",
"functions",
"put",
"forward",
"onto",
"text_string",
"returning",
"the",
"processed",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L178-L208 | train |
SpotlightData/preprocessing | preprocessing/text.py | remove_esc_chars | def remove_esc_chars(text_string):
'''
Removes any escape character within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None or... | python | def remove_esc_chars(text_string):
'''
Removes any escape character within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None or... | [
"def",
"remove_esc_chars",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"\" \"",
".",
"join",
"(",
"re",... | Removes any escape character within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Removes",
"any",
"escape",
"character",
"within",
"text_string",
"and",
"returns",
"the",
"new",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L210-L227 | train |
SpotlightData/preprocessing | preprocessing/text.py | remove_numbers | def remove_numbers(text_string):
'''
Removes any digit value discovered within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is Non... | python | def remove_numbers(text_string):
'''
Removes any digit value discovered within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is Non... | [
"def",
"remove_numbers",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"\" \"",
".",
"join",
"(",
"re",
... | Removes any digit value discovered within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Removes",
"any",
"digit",
"value",
"discovered",
"within",
"text_string",
"and",
"returns",
"the",
"new",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L229-L246 | train |
SpotlightData/preprocessing | preprocessing/text.py | remove_number_words | def remove_number_words(text_string):
'''
Removes any integer represented as a word within text_string and returns the new string as
type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if te... | python | def remove_number_words(text_string):
'''
Removes any integer represented as a word within text_string and returns the new string as
type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if te... | [
"def",
"remove_number_words",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"for",
"word",
"in",
"NUMBER_WORDS",
":",... | Removes any integer represented as a word within text_string and returns the new string as
type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Removes",
"any",
"integer",
"represented",
"as",
"a",
"word",
"within",
"text_string",
"and",
"returns",
"the",
"new",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L248-L268 | train |
SpotlightData/preprocessing | preprocessing/text.py | remove_urls | def remove_urls(text_string):
'''
Removes all URLs within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None or text_string == "... | python | def remove_urls(text_string):
'''
Removes all URLs within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed
'''
if text_string is None or text_string == "... | [
"def",
"remove_urls",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"\" \"",
".",
"join",
"(",
"re",
".... | Removes all URLs within text_string and returns the new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a non-string argument be passed | [
"Removes",
"all",
"URLs",
"within",
"text_string",
"and",
"returns",
"the",
"new",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L313-L330 | train |
SpotlightData/preprocessing | preprocessing/text.py | remove_whitespace | def remove_whitespace(text_string):
'''
Removes all whitespace found within text_string and returns new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or NoneType not be passed as an argument
'''
if text_st... | python | def remove_whitespace(text_string):
'''
Removes all whitespace found within text_string and returns new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or NoneType not be passed as an argument
'''
if text_st... | [
"def",
"remove_whitespace",
"(",
"text_string",
")",
":",
"if",
"text_string",
"is",
"None",
"or",
"text_string",
"==",
"\"\"",
":",
"return",
"\"\"",
"elif",
"isinstance",
"(",
"text_string",
",",
"str",
")",
":",
"return",
"\" \"",
".",
"join",
"(",
"tex... | Removes all whitespace found within text_string and returns new string as type str.
Keyword argument:
- text_string: string instance
Exceptions raised:
- InputError: occurs should a string or NoneType not be passed as an argument | [
"Removes",
"all",
"whitespace",
"found",
"within",
"text_string",
"and",
"returns",
"new",
"string",
"as",
"type",
"str",
"."
] | 180c6472bc2642afbd7a1ece08d0b0d14968a708 | https://github.com/SpotlightData/preprocessing/blob/180c6472bc2642afbd7a1ece08d0b0d14968a708/preprocessing/text.py#L332-L349 | train |
romaryd/python-logging-mixin | loggingmixin/__init__.py | WrappedLogger.log | def log(self, level, message, *args, **kwargs):
"""
This is the primary method to override to ensure logging with extra
options gets correctly specified.
"""
extra = self.extras.copy()
extra.update(kwargs.pop('extra', {}))
kwargs['extra'] = extra
self.log... | python | def log(self, level, message, *args, **kwargs):
"""
This is the primary method to override to ensure logging with extra
options gets correctly specified.
"""
extra = self.extras.copy()
extra.update(kwargs.pop('extra', {}))
kwargs['extra'] = extra
self.log... | [
"def",
"log",
"(",
"self",
",",
"level",
",",
"message",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"extra",
"=",
"self",
".",
"extras",
".",
"copy",
"(",
")",
"extra",
".",
"update",
"(",
"kwargs",
".",
"pop",
"(",
"'extra'",
",",
"{... | This is the primary method to override to ensure logging with extra
options gets correctly specified. | [
"This",
"is",
"the",
"primary",
"method",
"to",
"override",
"to",
"ensure",
"logging",
"with",
"extra",
"options",
"gets",
"correctly",
"specified",
"."
] | 8ac77df5731b607e6ff9ef762e71398cb5a892ea | https://github.com/romaryd/python-logging-mixin/blob/8ac77df5731b607e6ff9ef762e71398cb5a892ea/loggingmixin/__init__.py#L90-L99 | train |
romaryd/python-logging-mixin | loggingmixin/__init__.py | WrappedLogger.warning | def warning(self, message, *args, **kwargs):
"""
Specialized warnings system. If a warning subclass is passed into
the keyword arguments and raise_warnings is True - the warnning will
be passed to the warnings module.
"""
warncls = kwargs.pop('warning', None)
if w... | python | def warning(self, message, *args, **kwargs):
"""
Specialized warnings system. If a warning subclass is passed into
the keyword arguments and raise_warnings is True - the warnning will
be passed to the warnings module.
"""
warncls = kwargs.pop('warning', None)
if w... | [
"def",
"warning",
"(",
"self",
",",
"message",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"warncls",
"=",
"kwargs",
".",
"pop",
"(",
"'warning'",
",",
"None",
")",
"if",
"warncls",
"and",
"self",
".",
"raise_warnings",
":",
"warnings",
".",... | Specialized warnings system. If a warning subclass is passed into
the keyword arguments and raise_warnings is True - the warnning will
be passed to the warnings module. | [
"Specialized",
"warnings",
"system",
".",
"If",
"a",
"warning",
"subclass",
"is",
"passed",
"into",
"the",
"keyword",
"arguments",
"and",
"raise_warnings",
"is",
"True",
"-",
"the",
"warnning",
"will",
"be",
"passed",
"to",
"the",
"warnings",
"module",
"."
] | 8ac77df5731b607e6ff9ef762e71398cb5a892ea | https://github.com/romaryd/python-logging-mixin/blob/8ac77df5731b607e6ff9ef762e71398cb5a892ea/loggingmixin/__init__.py#L107-L117 | train |
romaryd/python-logging-mixin | loggingmixin/__init__.py | ServiceLogger.log | def log(self, level, message, *args, **kwargs):
"""
Provide current user as extra context to the logger
"""
extra = kwargs.pop('extra', {})
extra.update({
'user': self.user
})
kwargs['extra'] = extra
super(ServiceLogger, self).log(level, messa... | python | def log(self, level, message, *args, **kwargs):
"""
Provide current user as extra context to the logger
"""
extra = kwargs.pop('extra', {})
extra.update({
'user': self.user
})
kwargs['extra'] = extra
super(ServiceLogger, self).log(level, messa... | [
"def",
"log",
"(",
"self",
",",
"level",
",",
"message",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"extra",
"=",
"kwargs",
".",
"pop",
"(",
"'extra'",
",",
"{",
"}",
")",
"extra",
".",
"update",
"(",
"{",
"'user'",
":",
"self",
".",
... | Provide current user as extra context to the logger | [
"Provide",
"current",
"user",
"as",
"extra",
"context",
"to",
"the",
"logger"
] | 8ac77df5731b607e6ff9ef762e71398cb5a892ea | https://github.com/romaryd/python-logging-mixin/blob/8ac77df5731b607e6ff9ef762e71398cb5a892ea/loggingmixin/__init__.py#L147-L157 | train |
romaryd/python-logging-mixin | loggingmixin/__init__.py | LoggingMixin.logger | def logger(self):
"""
Instantiates and returns a ServiceLogger instance
"""
if not hasattr(self, '_logger') or not self._logger:
self._logger = ServiceLogger()
return self._logger | python | def logger(self):
"""
Instantiates and returns a ServiceLogger instance
"""
if not hasattr(self, '_logger') or not self._logger:
self._logger = ServiceLogger()
return self._logger | [
"def",
"logger",
"(",
"self",
")",
":",
"if",
"not",
"hasattr",
"(",
"self",
",",
"'_logger'",
")",
"or",
"not",
"self",
".",
"_logger",
":",
"self",
".",
"_logger",
"=",
"ServiceLogger",
"(",
")",
"return",
"self",
".",
"_logger"
] | Instantiates and returns a ServiceLogger instance | [
"Instantiates",
"and",
"returns",
"a",
"ServiceLogger",
"instance"
] | 8ac77df5731b607e6ff9ef762e71398cb5a892ea | https://github.com/romaryd/python-logging-mixin/blob/8ac77df5731b607e6ff9ef762e71398cb5a892ea/loggingmixin/__init__.py#L166-L172 | train |
OpenTreeOfLife/peyotl | tutorials/ot-oti-find-studies.py | ot_find_studies | def ot_find_studies(arg_dict, exact=True, verbose=False, oti_wrapper=None):
"""Uses a peyotl wrapper around an Open Tree web service to get a list of studies
including values `value` for a given property to be searched on `porperty`.
The oti_wrapper can be None (in which case the default wrapper from peyot... | python | def ot_find_studies(arg_dict, exact=True, verbose=False, oti_wrapper=None):
"""Uses a peyotl wrapper around an Open Tree web service to get a list of studies
including values `value` for a given property to be searched on `porperty`.
The oti_wrapper can be None (in which case the default wrapper from peyot... | [
"def",
"ot_find_studies",
"(",
"arg_dict",
",",
"exact",
"=",
"True",
",",
"verbose",
"=",
"False",
",",
"oti_wrapper",
"=",
"None",
")",
":",
"if",
"oti_wrapper",
"is",
"None",
":",
"from",
"peyotl",
".",
"sugar",
"import",
"oti",
"oti_wrapper",
"=",
"o... | Uses a peyotl wrapper around an Open Tree web service to get a list of studies
including values `value` for a given property to be searched on `porperty`.
The oti_wrapper can be None (in which case the default wrapper from peyotl.sugar will be used.
All other arguments correspond to the arguments of the we... | [
"Uses",
"a",
"peyotl",
"wrapper",
"around",
"an",
"Open",
"Tree",
"web",
"service",
"to",
"get",
"a",
"list",
"of",
"studies",
"including",
"values",
"value",
"for",
"a",
"given",
"property",
"to",
"be",
"searched",
"on",
"porperty",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/tutorials/ot-oti-find-studies.py#L12-L25 | train |
OpenTreeOfLife/peyotl | tutorials/ot-oti-find-studies.py | main | def main(argv):
"""This function sets up a command-line option parser and then calls print_matching_trees
to do all of the real work.
"""
import argparse
description = 'Uses Open Tree of Life web services to try to find a tree with the value property pair specified. ' \
'setting --... | python | def main(argv):
"""This function sets up a command-line option parser and then calls print_matching_trees
to do all of the real work.
"""
import argparse
description = 'Uses Open Tree of Life web services to try to find a tree with the value property pair specified. ' \
'setting --... | [
"def",
"main",
"(",
"argv",
")",
":",
"import",
"argparse",
"description",
"=",
"'Uses Open Tree of Life web services to try to find a tree with the value property pair specified. '",
"'setting --fuzzy will allow fuzzy matching'",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"... | This function sets up a command-line option parser and then calls print_matching_trees
to do all of the real work. | [
"This",
"function",
"sets",
"up",
"a",
"command",
"-",
"line",
"option",
"parser",
"and",
"then",
"calls",
"print_matching_trees",
"to",
"do",
"all",
"of",
"the",
"real",
"work",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/tutorials/ot-oti-find-studies.py#L36-L59 | train |
OpenTreeOfLife/peyotl | tutorials/ot-taxo-mrca-to-root.py | main | def main(argv):
"""This function sets up a command-line option parser and then calls
to do all of the real work.
"""
import argparse
import codecs
# have to be ready to deal with utf-8 names
out = codecs.getwriter('utf-8')(sys.stdout)
description = '''Takes a series of at least 2 OTT ids... | python | def main(argv):
"""This function sets up a command-line option parser and then calls
to do all of the real work.
"""
import argparse
import codecs
# have to be ready to deal with utf-8 names
out = codecs.getwriter('utf-8')(sys.stdout)
description = '''Takes a series of at least 2 OTT ids... | [
"def",
"main",
"(",
"argv",
")",
":",
"import",
"argparse",
"import",
"codecs",
"# have to be ready to deal with utf-8 names",
"out",
"=",
"codecs",
".",
"getwriter",
"(",
"'utf-8'",
")",
"(",
"sys",
".",
"stdout",
")",
"description",
"=",
"'''Takes a series of at... | This function sets up a command-line option parser and then calls
to do all of the real work. | [
"This",
"function",
"sets",
"up",
"a",
"command",
"-",
"line",
"option",
"parser",
"and",
"then",
"calls",
"to",
"do",
"all",
"of",
"the",
"real",
"work",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/tutorials/ot-taxo-mrca-to-root.py#L23-L46 | train |
CybOXProject/mixbox | mixbox/datautils.py | is_sequence | def is_sequence(value):
"""Determine if a value is a sequence type.
Returns:
``True`` if `value` is a sequence type (e.g., ``list``, or ``tuple``).
String types will return ``False``.
NOTE: On Python 3, strings have the __iter__ defined, so a simple hasattr
check is insufficient.
"""
... | python | def is_sequence(value):
"""Determine if a value is a sequence type.
Returns:
``True`` if `value` is a sequence type (e.g., ``list``, or ``tuple``).
String types will return ``False``.
NOTE: On Python 3, strings have the __iter__ defined, so a simple hasattr
check is insufficient.
"""
... | [
"def",
"is_sequence",
"(",
"value",
")",
":",
"return",
"(",
"hasattr",
"(",
"value",
",",
"\"__iter__\"",
")",
"and",
"not",
"isinstance",
"(",
"value",
",",
"(",
"six",
".",
"string_types",
",",
"six",
".",
"binary_type",
")",
")",
")"
] | Determine if a value is a sequence type.
Returns:
``True`` if `value` is a sequence type (e.g., ``list``, or ``tuple``).
String types will return ``False``.
NOTE: On Python 3, strings have the __iter__ defined, so a simple hasattr
check is insufficient. | [
"Determine",
"if",
"a",
"value",
"is",
"a",
"sequence",
"type",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/datautils.py#L12-L23 | train |
CybOXProject/mixbox | mixbox/datautils.py | import_class | def import_class(classpath):
"""Import the class referred to by the fully qualified class path.
Args:
classpath: A full "foo.bar.MyClass" path to a class definition.
Returns:
The class referred to by the classpath.
Raises:
ImportError: If an error occurs while importing the mo... | python | def import_class(classpath):
"""Import the class referred to by the fully qualified class path.
Args:
classpath: A full "foo.bar.MyClass" path to a class definition.
Returns:
The class referred to by the classpath.
Raises:
ImportError: If an error occurs while importing the mo... | [
"def",
"import_class",
"(",
"classpath",
")",
":",
"modname",
",",
"classname",
"=",
"classpath",
".",
"rsplit",
"(",
"\".\"",
",",
"1",
")",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"modname",
")",
"klass",
"=",
"getattr",
"(",
"module",
"... | Import the class referred to by the fully qualified class path.
Args:
classpath: A full "foo.bar.MyClass" path to a class definition.
Returns:
The class referred to by the classpath.
Raises:
ImportError: If an error occurs while importing the module.
AttributeError: IF the... | [
"Import",
"the",
"class",
"referred",
"to",
"by",
"the",
"fully",
"qualified",
"class",
"path",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/datautils.py#L26-L42 | train |
CybOXProject/mixbox | mixbox/datautils.py | resolve_class | def resolve_class(classref):
"""Attempt to return a Python class for the input class reference.
If `classref` is a class or None, return it. If `classref` is a
python classpath (e.g., "foo.bar.MyClass") import the class and return
it.
Args:
classref: A fully-qualified Python path to class,... | python | def resolve_class(classref):
"""Attempt to return a Python class for the input class reference.
If `classref` is a class or None, return it. If `classref` is a
python classpath (e.g., "foo.bar.MyClass") import the class and return
it.
Args:
classref: A fully-qualified Python path to class,... | [
"def",
"resolve_class",
"(",
"classref",
")",
":",
"if",
"classref",
"is",
"None",
":",
"return",
"None",
"elif",
"isinstance",
"(",
"classref",
",",
"six",
".",
"class_types",
")",
":",
"return",
"classref",
"elif",
"isinstance",
"(",
"classref",
",",
"si... | Attempt to return a Python class for the input class reference.
If `classref` is a class or None, return it. If `classref` is a
python classpath (e.g., "foo.bar.MyClass") import the class and return
it.
Args:
classref: A fully-qualified Python path to class, or a Python class.
Returns:
... | [
"Attempt",
"to",
"return",
"a",
"Python",
"class",
"for",
"the",
"input",
"class",
"reference",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/datautils.py#L45-L65 | train |
CybOXProject/mixbox | mixbox/datautils.py | needkwargs | def needkwargs(*argnames):
"""Function decorator which checks that the decorated function is called
with a set of required kwargs.
Args:
*argnames: String keyword argument names.
Raises:
ValueError: If a required kwarg is missing in the decorated function
call.
"""
... | python | def needkwargs(*argnames):
"""Function decorator which checks that the decorated function is called
with a set of required kwargs.
Args:
*argnames: String keyword argument names.
Raises:
ValueError: If a required kwarg is missing in the decorated function
call.
"""
... | [
"def",
"needkwargs",
"(",
"*",
"argnames",
")",
":",
"required",
"=",
"set",
"(",
"argnames",
")",
"def",
"decorator",
"(",
"func",
")",
":",
"def",
"inner",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"missing",
"=",
"required",
"-",
"set... | Function decorator which checks that the decorated function is called
with a set of required kwargs.
Args:
*argnames: String keyword argument names.
Raises:
ValueError: If a required kwarg is missing in the decorated function
call. | [
"Function",
"decorator",
"which",
"checks",
"that",
"the",
"decorated",
"function",
"is",
"called",
"with",
"a",
"set",
"of",
"required",
"kwargs",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/datautils.py#L83-L104 | train |
flyte/apcaccess | apcaccess/status.py | get | def get(host="localhost", port=3551, timeout=30):
"""
Connect to the APCUPSd NIS and request its status.
"""
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.settimeout(timeout)
sock.connect((host, port))
sock.send(CMD_STATUS)
buffr = ""
while not buffr.endswith(EOF):
... | python | def get(host="localhost", port=3551, timeout=30):
"""
Connect to the APCUPSd NIS and request its status.
"""
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.settimeout(timeout)
sock.connect((host, port))
sock.send(CMD_STATUS)
buffr = ""
while not buffr.endswith(EOF):
... | [
"def",
"get",
"(",
"host",
"=",
"\"localhost\"",
",",
"port",
"=",
"3551",
",",
"timeout",
"=",
"30",
")",
":",
"sock",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_INET",
",",
"socket",
".",
"SOCK_STREAM",
")",
"sock",
".",
"settimeout",
"(... | Connect to the APCUPSd NIS and request its status. | [
"Connect",
"to",
"the",
"APCUPSd",
"NIS",
"and",
"request",
"its",
"status",
"."
] | 0c8a5d5e4ba1c07110e411b4ffea4ddccef4829a | https://github.com/flyte/apcaccess/blob/0c8a5d5e4ba1c07110e411b4ffea4ddccef4829a/apcaccess/status.py#L31-L43 | train |
flyte/apcaccess | apcaccess/status.py | strip_units_from_lines | def strip_units_from_lines(lines):
"""
Removes all units from the ends of the lines.
"""
for line in lines:
for unit in ALL_UNITS:
if line.endswith(" %s" % unit):
line = line[:-1-len(unit)]
yield line | python | def strip_units_from_lines(lines):
"""
Removes all units from the ends of the lines.
"""
for line in lines:
for unit in ALL_UNITS:
if line.endswith(" %s" % unit):
line = line[:-1-len(unit)]
yield line | [
"def",
"strip_units_from_lines",
"(",
"lines",
")",
":",
"for",
"line",
"in",
"lines",
":",
"for",
"unit",
"in",
"ALL_UNITS",
":",
"if",
"line",
".",
"endswith",
"(",
"\" %s\"",
"%",
"unit",
")",
":",
"line",
"=",
"line",
"[",
":",
"-",
"1",
"-",
"... | Removes all units from the ends of the lines. | [
"Removes",
"all",
"units",
"from",
"the",
"ends",
"of",
"the",
"lines",
"."
] | 0c8a5d5e4ba1c07110e411b4ffea4ddccef4829a | https://github.com/flyte/apcaccess/blob/0c8a5d5e4ba1c07110e411b4ffea4ddccef4829a/apcaccess/status.py#L69-L77 | train |
flyte/apcaccess | apcaccess/status.py | print_status | def print_status(raw_status, strip_units=False):
"""
Print the status to stdout in the same format as the original apcaccess.
"""
lines = split(raw_status)
if strip_units:
lines = strip_units_from_lines(lines)
for line in lines:
print(line) | python | def print_status(raw_status, strip_units=False):
"""
Print the status to stdout in the same format as the original apcaccess.
"""
lines = split(raw_status)
if strip_units:
lines = strip_units_from_lines(lines)
for line in lines:
print(line) | [
"def",
"print_status",
"(",
"raw_status",
",",
"strip_units",
"=",
"False",
")",
":",
"lines",
"=",
"split",
"(",
"raw_status",
")",
"if",
"strip_units",
":",
"lines",
"=",
"strip_units_from_lines",
"(",
"lines",
")",
"for",
"line",
"in",
"lines",
":",
"pr... | Print the status to stdout in the same format as the original apcaccess. | [
"Print",
"the",
"status",
"to",
"stdout",
"in",
"the",
"same",
"format",
"as",
"the",
"original",
"apcaccess",
"."
] | 0c8a5d5e4ba1c07110e411b4ffea4ddccef4829a | https://github.com/flyte/apcaccess/blob/0c8a5d5e4ba1c07110e411b4ffea4ddccef4829a/apcaccess/status.py#L80-L88 | train |
OpenTreeOfLife/peyotl | peyotl/api/taxomachine.py | _TaxomachineAPIWrapper.get_cached_parent_for_taxon | def get_cached_parent_for_taxon(self, child_taxon):
"""If the taxa are being cached, this call will create a the lineage "spike" for taxon child_taxon
Expecting child_taxon to have a non-empty _taxonomic_lineage with response dicts that can create
an ancestral TaxonWrapper.
"""
... | python | def get_cached_parent_for_taxon(self, child_taxon):
"""If the taxa are being cached, this call will create a the lineage "spike" for taxon child_taxon
Expecting child_taxon to have a non-empty _taxonomic_lineage with response dicts that can create
an ancestral TaxonWrapper.
"""
... | [
"def",
"get_cached_parent_for_taxon",
"(",
"self",
",",
"child_taxon",
")",
":",
"if",
"self",
".",
"_ott_id2taxon",
"is",
"None",
":",
"resp",
"=",
"child_taxon",
".",
"_taxonomic_lineage",
"[",
"0",
"]",
"tl",
"=",
"child_taxon",
".",
"_taxonomic_lineage",
"... | If the taxa are being cached, this call will create a the lineage "spike" for taxon child_taxon
Expecting child_taxon to have a non-empty _taxonomic_lineage with response dicts that can create
an ancestral TaxonWrapper. | [
"If",
"the",
"taxa",
"are",
"being",
"cached",
"this",
"call",
"will",
"create",
"a",
"the",
"lineage",
"spike",
"for",
"taxon",
"child_taxon"
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/peyotl/api/taxomachine.py#L373-L405 | train |
OpenTreeOfLife/peyotl | peyotl/api/taxon.py | TaxonWrapper.update_empty_fields | def update_empty_fields(self, **kwargs):
"""Updates the field of info about an OTU that might not be filled in by a match_names or taxon call."""
if self._is_deprecated is None:
self._is_deprecated = kwargs.get('is_deprecated')
if self._is_dubious is None:
self._is_dubiou... | python | def update_empty_fields(self, **kwargs):
"""Updates the field of info about an OTU that might not be filled in by a match_names or taxon call."""
if self._is_deprecated is None:
self._is_deprecated = kwargs.get('is_deprecated')
if self._is_dubious is None:
self._is_dubiou... | [
"def",
"update_empty_fields",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"self",
".",
"_is_deprecated",
"is",
"None",
":",
"self",
".",
"_is_deprecated",
"=",
"kwargs",
".",
"get",
"(",
"'is_deprecated'",
")",
"if",
"self",
".",
"_is_dubious",
... | Updates the field of info about an OTU that might not be filled in by a match_names or taxon call. | [
"Updates",
"the",
"field",
"of",
"info",
"about",
"an",
"OTU",
"that",
"might",
"not",
"be",
"filled",
"in",
"by",
"a",
"match_names",
"or",
"taxon",
"call",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/peyotl/api/taxon.py#L47-L70 | train |
OpenTreeOfLife/peyotl | scripts/nexson/prune_to_clean_mapped.py | _check_rev_dict | def _check_rev_dict(tree, ebt):
"""Verifyies that `ebt` is the inverse of the `edgeBySourceId` data member of `tree`"""
ebs = defaultdict(dict)
for edge in ebt.values():
source_id = edge['@source']
edge_id = edge['@id']
ebs[source_id][edge_id] = edge
assert ebs == tree['edgeBySou... | python | def _check_rev_dict(tree, ebt):
"""Verifyies that `ebt` is the inverse of the `edgeBySourceId` data member of `tree`"""
ebs = defaultdict(dict)
for edge in ebt.values():
source_id = edge['@source']
edge_id = edge['@id']
ebs[source_id][edge_id] = edge
assert ebs == tree['edgeBySou... | [
"def",
"_check_rev_dict",
"(",
"tree",
",",
"ebt",
")",
":",
"ebs",
"=",
"defaultdict",
"(",
"dict",
")",
"for",
"edge",
"in",
"ebt",
".",
"values",
"(",
")",
":",
"source_id",
"=",
"edge",
"[",
"'@source'",
"]",
"edge_id",
"=",
"edge",
"[",
"'@id'",... | Verifyies that `ebt` is the inverse of the `edgeBySourceId` data member of `tree` | [
"Verifyies",
"that",
"ebt",
"is",
"the",
"inverse",
"of",
"the",
"edgeBySourceId",
"data",
"member",
"of",
"tree"
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/scripts/nexson/prune_to_clean_mapped.py#L42-L49 | train |
OpenTreeOfLife/peyotl | scripts/nexson/prune_to_clean_mapped.py | NexsonTreeWrapper._create_edge_by_target | def _create_edge_by_target(self):
"""creates a edge_by_target dict with the same edge objects as the edge_by_source.
Also adds an '@id' field to each edge."""
ebt = {}
for edge_dict in self._edge_by_source.values():
for edge_id, edge in edge_dict.items():
targ... | python | def _create_edge_by_target(self):
"""creates a edge_by_target dict with the same edge objects as the edge_by_source.
Also adds an '@id' field to each edge."""
ebt = {}
for edge_dict in self._edge_by_source.values():
for edge_id, edge in edge_dict.items():
targ... | [
"def",
"_create_edge_by_target",
"(",
"self",
")",
":",
"ebt",
"=",
"{",
"}",
"for",
"edge_dict",
"in",
"self",
".",
"_edge_by_source",
".",
"values",
"(",
")",
":",
"for",
"edge_id",
",",
"edge",
"in",
"edge_dict",
".",
"items",
"(",
")",
":",
"target... | creates a edge_by_target dict with the same edge objects as the edge_by_source.
Also adds an '@id' field to each edge. | [
"creates",
"a",
"edge_by_target",
"dict",
"with",
"the",
"same",
"edge",
"objects",
"as",
"the",
"edge_by_source",
".",
"Also",
"adds",
"an"
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/scripts/nexson/prune_to_clean_mapped.py#L86-L97 | train |
OpenTreeOfLife/peyotl | scripts/nexson/prune_to_clean_mapped.py | NexsonTreeWrapper.prune_to_ingroup | def prune_to_ingroup(self):
"""Remove nodes and edges from tree if they are not the ingroup or a descendant of it."""
# Prune to just the ingroup
if not self._ingroup_node_id:
_LOG.debug('No ingroup node was specified.')
self._ingroup_node_id = self.root_node_id
e... | python | def prune_to_ingroup(self):
"""Remove nodes and edges from tree if they are not the ingroup or a descendant of it."""
# Prune to just the ingroup
if not self._ingroup_node_id:
_LOG.debug('No ingroup node was specified.')
self._ingroup_node_id = self.root_node_id
e... | [
"def",
"prune_to_ingroup",
"(",
"self",
")",
":",
"# Prune to just the ingroup",
"if",
"not",
"self",
".",
"_ingroup_node_id",
":",
"_LOG",
".",
"debug",
"(",
"'No ingroup node was specified.'",
")",
"self",
".",
"_ingroup_node_id",
"=",
"self",
".",
"root_node_id",... | Remove nodes and edges from tree if they are not the ingroup or a descendant of it. | [
"Remove",
"nodes",
"and",
"edges",
"from",
"tree",
"if",
"they",
"are",
"not",
"the",
"ingroup",
"or",
"a",
"descendant",
"of",
"it",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/scripts/nexson/prune_to_clean_mapped.py#L119-L130 | train |
OpenTreeOfLife/peyotl | scripts/nexson/prune_to_clean_mapped.py | NexsonTreeWrapper.prune_clade | def prune_clade(self, node_id):
"""Prune `node_id` and the edges and nodes that are tipward of it.
Caller must delete the edge to node_id."""
to_del_nodes = [node_id]
while bool(to_del_nodes):
node_id = to_del_nodes.pop(0)
self._flag_node_as_del_and_del_in_by_targ... | python | def prune_clade(self, node_id):
"""Prune `node_id` and the edges and nodes that are tipward of it.
Caller must delete the edge to node_id."""
to_del_nodes = [node_id]
while bool(to_del_nodes):
node_id = to_del_nodes.pop(0)
self._flag_node_as_del_and_del_in_by_targ... | [
"def",
"prune_clade",
"(",
"self",
",",
"node_id",
")",
":",
"to_del_nodes",
"=",
"[",
"node_id",
"]",
"while",
"bool",
"(",
"to_del_nodes",
")",
":",
"node_id",
"=",
"to_del_nodes",
".",
"pop",
"(",
"0",
")",
"self",
".",
"_flag_node_as_del_and_del_in_by_ta... | Prune `node_id` and the edges and nodes that are tipward of it.
Caller must delete the edge to node_id. | [
"Prune",
"node_id",
"and",
"the",
"edges",
"and",
"nodes",
"that",
"are",
"tipward",
"of",
"it",
".",
"Caller",
"must",
"delete",
"the",
"edge",
"to",
"node_id",
"."
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/scripts/nexson/prune_to_clean_mapped.py#L147-L159 | train |
OpenTreeOfLife/peyotl | scripts/nexson/prune_to_clean_mapped.py | NexsonTreeWrapper.suppress_deg_one_node | def suppress_deg_one_node(self, to_par_edge, nd_id, to_child_edge):
"""Deletes to_par_edge and nd_id. To be used when nd_id is an out-degree= 1 node"""
# circumvent the node with nd_id
to_child_edge_id = to_child_edge['@id']
par = to_par_edge['@source']
self._edge_by_source[par][... | python | def suppress_deg_one_node(self, to_par_edge, nd_id, to_child_edge):
"""Deletes to_par_edge and nd_id. To be used when nd_id is an out-degree= 1 node"""
# circumvent the node with nd_id
to_child_edge_id = to_child_edge['@id']
par = to_par_edge['@source']
self._edge_by_source[par][... | [
"def",
"suppress_deg_one_node",
"(",
"self",
",",
"to_par_edge",
",",
"nd_id",
",",
"to_child_edge",
")",
":",
"# circumvent the node with nd_id",
"to_child_edge_id",
"=",
"to_child_edge",
"[",
"'@id'",
"]",
"par",
"=",
"to_par_edge",
"[",
"'@source'",
"]",
"self",
... | Deletes to_par_edge and nd_id. To be used when nd_id is an out-degree= 1 node | [
"Deletes",
"to_par_edge",
"and",
"nd_id",
".",
"To",
"be",
"used",
"when",
"nd_id",
"is",
"an",
"out",
"-",
"degree",
"=",
"1",
"node"
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/scripts/nexson/prune_to_clean_mapped.py#L278-L288 | train |
palantir/typedjsonrpc | typedjsonrpc/method_info.py | MethodInfo.describe | def describe(self):
"""Describes the method.
:return: Description
:rtype: dict[str, object]
"""
return {
"name": self.name,
"params": self.params,
"returns": self.returns,
"description": self.description,
} | python | def describe(self):
"""Describes the method.
:return: Description
:rtype: dict[str, object]
"""
return {
"name": self.name,
"params": self.params,
"returns": self.returns,
"description": self.description,
} | [
"def",
"describe",
"(",
"self",
")",
":",
"return",
"{",
"\"name\"",
":",
"self",
".",
"name",
",",
"\"params\"",
":",
"self",
".",
"params",
",",
"\"returns\"",
":",
"self",
".",
"returns",
",",
"\"description\"",
":",
"self",
".",
"description",
",",
... | Describes the method.
:return: Description
:rtype: dict[str, object] | [
"Describes",
"the",
"method",
"."
] | 274218fcd236ff9643506caa629029c9ba25a0fb | https://github.com/palantir/typedjsonrpc/blob/274218fcd236ff9643506caa629029c9ba25a0fb/typedjsonrpc/method_info.py#L36-L47 | train |
palantir/typedjsonrpc | typedjsonrpc/method_info.py | MethodInfo.params | def params(self):
"""The parameters for this method in a JSON-compatible format
:rtype: list[dict[str, str]]
"""
return [{"name": p_name, "type": p_type.__name__}
for (p_name, p_type) in self.signature.parameter_types] | python | def params(self):
"""The parameters for this method in a JSON-compatible format
:rtype: list[dict[str, str]]
"""
return [{"name": p_name, "type": p_type.__name__}
for (p_name, p_type) in self.signature.parameter_types] | [
"def",
"params",
"(",
"self",
")",
":",
"return",
"[",
"{",
"\"name\"",
":",
"p_name",
",",
"\"type\"",
":",
"p_type",
".",
"__name__",
"}",
"for",
"(",
"p_name",
",",
"p_type",
")",
"in",
"self",
".",
"signature",
".",
"parameter_types",
"]"
] | The parameters for this method in a JSON-compatible format
:rtype: list[dict[str, str]] | [
"The",
"parameters",
"for",
"this",
"method",
"in",
"a",
"JSON",
"-",
"compatible",
"format"
] | 274218fcd236ff9643506caa629029c9ba25a0fb | https://github.com/palantir/typedjsonrpc/blob/274218fcd236ff9643506caa629029c9ba25a0fb/typedjsonrpc/method_info.py#L50-L56 | train |
palantir/typedjsonrpc | typedjsonrpc/method_info.py | MethodInfo.returns | def returns(self):
"""The return type for this method in a JSON-compatible format.
This handles the special case of ``None`` which allows ``type(None)`` also.
:rtype: str | None
"""
return_type = self.signature.return_type
none_type = type(None)
if return_type i... | python | def returns(self):
"""The return type for this method in a JSON-compatible format.
This handles the special case of ``None`` which allows ``type(None)`` also.
:rtype: str | None
"""
return_type = self.signature.return_type
none_type = type(None)
if return_type i... | [
"def",
"returns",
"(",
"self",
")",
":",
"return_type",
"=",
"self",
".",
"signature",
".",
"return_type",
"none_type",
"=",
"type",
"(",
"None",
")",
"if",
"return_type",
"is",
"not",
"None",
"and",
"return_type",
"is",
"not",
"none_type",
":",
"return",
... | The return type for this method in a JSON-compatible format.
This handles the special case of ``None`` which allows ``type(None)`` also.
:rtype: str | None | [
"The",
"return",
"type",
"for",
"this",
"method",
"in",
"a",
"JSON",
"-",
"compatible",
"format",
"."
] | 274218fcd236ff9643506caa629029c9ba25a0fb | https://github.com/palantir/typedjsonrpc/blob/274218fcd236ff9643506caa629029c9ba25a0fb/typedjsonrpc/method_info.py#L59-L69 | train |
palantir/typedjsonrpc | typedjsonrpc/method_info.py | MethodSignature.create | def create(parameter_names, parameter_types, return_type):
"""Returns a signature object ensuring order of parameter names and types.
:param parameter_names: A list of ordered parameter names
:type parameter_names: list[str]
:param parameter_types: A dictionary of parameter names to typ... | python | def create(parameter_names, parameter_types, return_type):
"""Returns a signature object ensuring order of parameter names and types.
:param parameter_names: A list of ordered parameter names
:type parameter_names: list[str]
:param parameter_types: A dictionary of parameter names to typ... | [
"def",
"create",
"(",
"parameter_names",
",",
"parameter_types",
",",
"return_type",
")",
":",
"ordered_pairs",
"=",
"[",
"(",
"name",
",",
"parameter_types",
"[",
"name",
"]",
")",
"for",
"name",
"in",
"parameter_names",
"]",
"return",
"MethodSignature",
"(",... | Returns a signature object ensuring order of parameter names and types.
:param parameter_names: A list of ordered parameter names
:type parameter_names: list[str]
:param parameter_types: A dictionary of parameter names to types
:type parameter_types: dict[str, type]
:param retur... | [
"Returns",
"a",
"signature",
"object",
"ensuring",
"order",
"of",
"parameter",
"names",
"and",
"types",
"."
] | 274218fcd236ff9643506caa629029c9ba25a0fb | https://github.com/palantir/typedjsonrpc/blob/274218fcd236ff9643506caa629029c9ba25a0fb/typedjsonrpc/method_info.py#L90-L102 | train |
OpenTreeOfLife/peyotl | peyotl/nexson_syntax/nexml2nexson.py | Nexml2Nexson._hbf_handle_child_elements | def _hbf_handle_child_elements(self, obj, ntl):
"""
Indirect recursion through _gen_hbf_el
"""
# accumulate a list of the children names in ko, and
# the a dictionary of tag to xml elements.
# repetition of a tag means that it will map to a list of
# xml eleme... | python | def _hbf_handle_child_elements(self, obj, ntl):
"""
Indirect recursion through _gen_hbf_el
"""
# accumulate a list of the children names in ko, and
# the a dictionary of tag to xml elements.
# repetition of a tag means that it will map to a list of
# xml eleme... | [
"def",
"_hbf_handle_child_elements",
"(",
"self",
",",
"obj",
",",
"ntl",
")",
":",
"# accumulate a list of the children names in ko, and",
"# the a dictionary of tag to xml elements.",
"# repetition of a tag means that it will map to a list of",
"# xml elements",
"cd",
"=",
"{",... | Indirect recursion through _gen_hbf_el | [
"Indirect",
"recursion",
"through",
"_gen_hbf_el"
] | 5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0 | https://github.com/OpenTreeOfLife/peyotl/blob/5e4e52a0fdbd17f490aa644ad79fda6ea2eda7c0/peyotl/nexson_syntax/nexml2nexson.py#L169-L208 | train |
CybOXProject/mixbox | mixbox/xml.py | get_xml_parser | def get_xml_parser(encoding=None):
"""Returns an ``etree.ETCompatXMLParser`` instance."""
parser = etree.ETCompatXMLParser(
huge_tree=True,
remove_comments=True,
strip_cdata=False,
remove_blank_text=True,
resolve_entities=False,
encoding=encoding
)
return... | python | def get_xml_parser(encoding=None):
"""Returns an ``etree.ETCompatXMLParser`` instance."""
parser = etree.ETCompatXMLParser(
huge_tree=True,
remove_comments=True,
strip_cdata=False,
remove_blank_text=True,
resolve_entities=False,
encoding=encoding
)
return... | [
"def",
"get_xml_parser",
"(",
"encoding",
"=",
"None",
")",
":",
"parser",
"=",
"etree",
".",
"ETCompatXMLParser",
"(",
"huge_tree",
"=",
"True",
",",
"remove_comments",
"=",
"True",
",",
"strip_cdata",
"=",
"False",
",",
"remove_blank_text",
"=",
"True",
",... | Returns an ``etree.ETCompatXMLParser`` instance. | [
"Returns",
"an",
"etree",
".",
"ETCompatXMLParser",
"instance",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/xml.py#L34-L45 | train |
CybOXProject/mixbox | mixbox/xml.py | get_etree_root | def get_etree_root(doc, encoding=None):
"""Returns an instance of lxml.etree._Element for the given `doc` input.
Args:
doc: The input XML document. Can be an instance of
``lxml.etree._Element``, ``lxml.etree._ElementTree``, a file-like
object, or a string filename.
encod... | python | def get_etree_root(doc, encoding=None):
"""Returns an instance of lxml.etree._Element for the given `doc` input.
Args:
doc: The input XML document. Can be an instance of
``lxml.etree._Element``, ``lxml.etree._ElementTree``, a file-like
object, or a string filename.
encod... | [
"def",
"get_etree_root",
"(",
"doc",
",",
"encoding",
"=",
"None",
")",
":",
"tree",
"=",
"get_etree",
"(",
"doc",
",",
"encoding",
")",
"root",
"=",
"tree",
".",
"getroot",
"(",
")",
"return",
"root"
] | Returns an instance of lxml.etree._Element for the given `doc` input.
Args:
doc: The input XML document. Can be an instance of
``lxml.etree._Element``, ``lxml.etree._ElementTree``, a file-like
object, or a string filename.
encoding: The character encoding of `doc`. If ``None... | [
"Returns",
"an",
"instance",
"of",
"lxml",
".",
"etree",
".",
"_Element",
"for",
"the",
"given",
"doc",
"input",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/xml.py#L58-L80 | train |
CybOXProject/mixbox | mixbox/xml.py | strip_cdata | def strip_cdata(text):
"""Removes all CDATA blocks from `text` if it contains them.
Note:
If the function contains escaped XML characters outside of a
CDATA block, they will be unescaped.
Args:
A string containing one or more CDATA blocks.
Returns:
An XML unescaped str... | python | def strip_cdata(text):
"""Removes all CDATA blocks from `text` if it contains them.
Note:
If the function contains escaped XML characters outside of a
CDATA block, they will be unescaped.
Args:
A string containing one or more CDATA blocks.
Returns:
An XML unescaped str... | [
"def",
"strip_cdata",
"(",
"text",
")",
":",
"if",
"not",
"is_cdata",
"(",
"text",
")",
":",
"return",
"text",
"xml",
"=",
"\"<e>{0}</e>\"",
".",
"format",
"(",
"text",
")",
"node",
"=",
"etree",
".",
"fromstring",
"(",
"xml",
")",
"return",
"node",
... | Removes all CDATA blocks from `text` if it contains them.
Note:
If the function contains escaped XML characters outside of a
CDATA block, they will be unescaped.
Args:
A string containing one or more CDATA blocks.
Returns:
An XML unescaped string with CDATA block qualifier... | [
"Removes",
"all",
"CDATA",
"blocks",
"from",
"text",
"if",
"it",
"contains",
"them",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/xml.py#L114-L133 | train |
CybOXProject/mixbox | mixbox/typedlist.py | TypedList._is_valid | def _is_valid(self, value):
"""Return True if the input value is valid for insertion into the
inner list.
Args:
value: An object about to be inserted.
"""
# Entities have an istypeof method that can perform more sophisticated
# type checking.
if hasa... | python | def _is_valid(self, value):
"""Return True if the input value is valid for insertion into the
inner list.
Args:
value: An object about to be inserted.
"""
# Entities have an istypeof method that can perform more sophisticated
# type checking.
if hasa... | [
"def",
"_is_valid",
"(",
"self",
",",
"value",
")",
":",
"# Entities have an istypeof method that can perform more sophisticated",
"# type checking.",
"if",
"hasattr",
"(",
"self",
".",
"_type",
",",
"\"istypeof\"",
")",
":",
"return",
"self",
".",
"_type",
".",
"is... | Return True if the input value is valid for insertion into the
inner list.
Args:
value: An object about to be inserted. | [
"Return",
"True",
"if",
"the",
"input",
"value",
"is",
"valid",
"for",
"insertion",
"into",
"the",
"inner",
"list",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/typedlist.py#L40-L53 | train |
CybOXProject/mixbox | mixbox/typedlist.py | TypedList._fix_value | def _fix_value(self, value):
"""Attempt to coerce value into the correct type.
Subclasses can override this function.
"""
try:
return self._castfunc(value)
except:
error = "Can't put '{0}' ({1}) into a {2}. Expected a {3} object."
error = erro... | python | def _fix_value(self, value):
"""Attempt to coerce value into the correct type.
Subclasses can override this function.
"""
try:
return self._castfunc(value)
except:
error = "Can't put '{0}' ({1}) into a {2}. Expected a {3} object."
error = erro... | [
"def",
"_fix_value",
"(",
"self",
",",
"value",
")",
":",
"try",
":",
"return",
"self",
".",
"_castfunc",
"(",
"value",
")",
"except",
":",
"error",
"=",
"\"Can't put '{0}' ({1}) into a {2}. Expected a {3} object.\"",
"error",
"=",
"error",
".",
"format",
"(",
... | Attempt to coerce value into the correct type.
Subclasses can override this function. | [
"Attempt",
"to",
"coerce",
"value",
"into",
"the",
"correct",
"type",
"."
] | 9097dae7a433f5b98c18171c4a5598f69a7d30af | https://github.com/CybOXProject/mixbox/blob/9097dae7a433f5b98c18171c4a5598f69a7d30af/mixbox/typedlist.py#L55-L70 | train |
hsolbrig/pyjsg | pyjsg/parser_impl/jsg_pairdef_parser.py | JSGPairDef.members_entries | def members_entries(self, all_are_optional: Optional[bool] = False) -> List[Tuple[str, str]]:
""" Generate a list quoted raw name, signature type entries for this pairdef, recursively traversing
reference types
:param all_are_optional: If true, all types are forced optional
:return: raw... | python | def members_entries(self, all_are_optional: Optional[bool] = False) -> List[Tuple[str, str]]:
""" Generate a list quoted raw name, signature type entries for this pairdef, recursively traversing
reference types
:param all_are_optional: If true, all types are forced optional
:return: raw... | [
"def",
"members_entries",
"(",
"self",
",",
"all_are_optional",
":",
"Optional",
"[",
"bool",
"]",
"=",
"False",
")",
"->",
"List",
"[",
"Tuple",
"[",
"str",
",",
"str",
"]",
"]",
":",
"if",
"self",
".",
"_type_reference",
":",
"rval",
":",
"List",
"... | Generate a list quoted raw name, signature type entries for this pairdef, recursively traversing
reference types
:param all_are_optional: If true, all types are forced optional
:return: raw name/ signature type for all elements in this pair | [
"Generate",
"a",
"list",
"quoted",
"raw",
"name",
"signature",
"type",
"entries",
"for",
"this",
"pairdef",
"recursively",
"traversing",
"reference",
"types"
] | 9b2b8fa8e3b8448abe70b09f804a79f0f31b32b7 | https://github.com/hsolbrig/pyjsg/blob/9b2b8fa8e3b8448abe70b09f804a79f0f31b32b7/pyjsg/parser_impl/jsg_pairdef_parser.py#L43-L57 | train |
hsolbrig/pyjsg | pyjsg/parser_impl/jsg_pairdef_parser.py | JSGPairDef._initializer_for | def _initializer_for(self, raw_name: str, cooked_name: str, prefix: Optional[str]) -> List[str]:
"""Create an initializer entry for the entry
:param raw_name: name unadjusted for python compatibility.
:param cooked_name: name that may or may not be python compatible
:param prefix: owne... | python | def _initializer_for(self, raw_name: str, cooked_name: str, prefix: Optional[str]) -> List[str]:
"""Create an initializer entry for the entry
:param raw_name: name unadjusted for python compatibility.
:param cooked_name: name that may or may not be python compatible
:param prefix: owne... | [
"def",
"_initializer_for",
"(",
"self",
",",
"raw_name",
":",
"str",
",",
"cooked_name",
":",
"str",
",",
"prefix",
":",
"Optional",
"[",
"str",
"]",
")",
"->",
"List",
"[",
"str",
"]",
":",
"mt_val",
"=",
"self",
".",
"_ebnf",
".",
"mt_value",
"(",
... | Create an initializer entry for the entry
:param raw_name: name unadjusted for python compatibility.
:param cooked_name: name that may or may not be python compatible
:param prefix: owner of the element - used when objects passed as arguments
:return: Initialization statements | [
"Create",
"an",
"initializer",
"entry",
"for",
"the",
"entry"
] | 9b2b8fa8e3b8448abe70b09f804a79f0f31b32b7 | https://github.com/hsolbrig/pyjsg/blob/9b2b8fa8e3b8448abe70b09f804a79f0f31b32b7/pyjsg/parser_impl/jsg_pairdef_parser.py#L96-L129 | train |
PSPC-SPAC-buyandsell/von_agent | von_agent/agent/holder_prover.py | HolderProver._assert_link_secret | def _assert_link_secret(self, action: str):
"""
Raise AbsentLinkSecret if link secret is not set.
:param action: action requiring link secret
"""
if self._link_secret is None:
LOGGER.debug('HolderProver._assert_link_secret: action %s requires link secret but it is n... | python | def _assert_link_secret(self, action: str):
"""
Raise AbsentLinkSecret if link secret is not set.
:param action: action requiring link secret
"""
if self._link_secret is None:
LOGGER.debug('HolderProver._assert_link_secret: action %s requires link secret but it is n... | [
"def",
"_assert_link_secret",
"(",
"self",
",",
"action",
":",
"str",
")",
":",
"if",
"self",
".",
"_link_secret",
"is",
"None",
":",
"LOGGER",
".",
"debug",
"(",
"'HolderProver._assert_link_secret: action %s requires link secret but it is not set'",
",",
"action",
")... | Raise AbsentLinkSecret if link secret is not set.
:param action: action requiring link secret | [
"Raise",
"AbsentLinkSecret",
"if",
"link",
"secret",
"is",
"not",
"set",
"."
] | 0b1c17cca3bd178b6e6974af84dbac1dfce5cf45 | https://github.com/PSPC-SPAC-buyandsell/von_agent/blob/0b1c17cca3bd178b6e6974af84dbac1dfce5cf45/von_agent/agent/holder_prover.py#L90-L99 | train |
PSPC-SPAC-buyandsell/von_agent | von_agent/agent/holder_prover.py | HolderProver.rev_regs | def rev_regs(self) -> list:
"""
Return list of revocation registry identifiers for which HolderProver has tails files.
:return: list of revocation registry identifiers for which HolderProver has tails files
"""
LOGGER.debug('HolderProver.rev_regs >>>')
rv = [basename(f... | python | def rev_regs(self) -> list:
"""
Return list of revocation registry identifiers for which HolderProver has tails files.
:return: list of revocation registry identifiers for which HolderProver has tails files
"""
LOGGER.debug('HolderProver.rev_regs >>>')
rv = [basename(f... | [
"def",
"rev_regs",
"(",
"self",
")",
"->",
"list",
":",
"LOGGER",
".",
"debug",
"(",
"'HolderProver.rev_regs >>>'",
")",
"rv",
"=",
"[",
"basename",
"(",
"f",
")",
"for",
"f",
"in",
"Tails",
".",
"links",
"(",
"self",
".",
"_dir_tails",
")",
"]",
"LO... | Return list of revocation registry identifiers for which HolderProver has tails files.
:return: list of revocation registry identifiers for which HolderProver has tails files | [
"Return",
"list",
"of",
"revocation",
"registry",
"identifiers",
"for",
"which",
"HolderProver",
"has",
"tails",
"files",
"."
] | 0b1c17cca3bd178b6e6974af84dbac1dfce5cf45 | https://github.com/PSPC-SPAC-buyandsell/von_agent/blob/0b1c17cca3bd178b6e6974af84dbac1dfce5cf45/von_agent/agent/holder_prover.py#L539-L550 | train |
PSPC-SPAC-buyandsell/von_agent | von_agent/agent/holder_prover.py | HolderProver.create_cred_req | async def create_cred_req(self, cred_offer_json: str, cd_id: str) -> (str, str):
"""
Create credential request as HolderProver and store in wallet; return credential json and metadata json.
Raise AbsentLinkSecret if link secret not set.
:param cred_offer_json: credential offer json
... | python | async def create_cred_req(self, cred_offer_json: str, cd_id: str) -> (str, str):
"""
Create credential request as HolderProver and store in wallet; return credential json and metadata json.
Raise AbsentLinkSecret if link secret not set.
:param cred_offer_json: credential offer json
... | [
"async",
"def",
"create_cred_req",
"(",
"self",
",",
"cred_offer_json",
":",
"str",
",",
"cd_id",
":",
"str",
")",
"->",
"(",
"str",
",",
"str",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'HolderProver.create_cred_req >>> cred_offer_json: %s, cd_id: %s'",
",",
"cre... | Create credential request as HolderProver and store in wallet; return credential json and metadata json.
Raise AbsentLinkSecret if link secret not set.
:param cred_offer_json: credential offer json
:param cd_id: credential definition identifier
:return: cred request json and correspond... | [
"Create",
"credential",
"request",
"as",
"HolderProver",
"and",
"store",
"in",
"wallet",
";",
"return",
"credential",
"json",
"and",
"metadata",
"json",
"."
] | 0b1c17cca3bd178b6e6974af84dbac1dfce5cf45 | https://github.com/PSPC-SPAC-buyandsell/von_agent/blob/0b1c17cca3bd178b6e6974af84dbac1dfce5cf45/von_agent/agent/holder_prover.py#L578-L612 | train |
PSPC-SPAC-buyandsell/von_agent | von_agent/agent/holder_prover.py | HolderProver.load_cache | async def load_cache(self, archive: bool = False) -> int:
"""
Load caches and archive enough to go offline and be able to generate proof
on all credentials in wallet.
Return timestamp (epoch seconds) of cache load event, also used as subdirectory
for cache archives.
:re... | python | async def load_cache(self, archive: bool = False) -> int:
"""
Load caches and archive enough to go offline and be able to generate proof
on all credentials in wallet.
Return timestamp (epoch seconds) of cache load event, also used as subdirectory
for cache archives.
:re... | [
"async",
"def",
"load_cache",
"(",
"self",
",",
"archive",
":",
"bool",
"=",
"False",
")",
"->",
"int",
":",
"LOGGER",
".",
"debug",
"(",
"'HolderProver.load_cache >>> archive: %s'",
",",
"archive",
")",
"rv",
"=",
"int",
"(",
"time",
"(",
")",
")",
"box... | Load caches and archive enough to go offline and be able to generate proof
on all credentials in wallet.
Return timestamp (epoch seconds) of cache load event, also used as subdirectory
for cache archives.
:return: cache load event timestamp (epoch seconds) | [
"Load",
"caches",
"and",
"archive",
"enough",
"to",
"go",
"offline",
"and",
"be",
"able",
"to",
"generate",
"proof",
"on",
"all",
"credentials",
"in",
"wallet",
"."
] | 0b1c17cca3bd178b6e6974af84dbac1dfce5cf45 | https://github.com/PSPC-SPAC-buyandsell/von_agent/blob/0b1c17cca3bd178b6e6974af84dbac1dfce5cf45/von_agent/agent/holder_prover.py#L649-L688 | train |
PSPC-SPAC-buyandsell/von_agent | von_agent/agent/holder_prover.py | HolderProver.get_creds | async def get_creds(self, proof_req_json: str, filt: dict = None, filt_dflt_incl: bool = False) -> (Set[str], str):
"""
Get credentials from HolderProver wallet corresponding to proof request and
filter criteria; return credential identifiers from wallet and credentials json.
Return empt... | python | async def get_creds(self, proof_req_json: str, filt: dict = None, filt_dflt_incl: bool = False) -> (Set[str], str):
"""
Get credentials from HolderProver wallet corresponding to proof request and
filter criteria; return credential identifiers from wallet and credentials json.
Return empt... | [
"async",
"def",
"get_creds",
"(",
"self",
",",
"proof_req_json",
":",
"str",
",",
"filt",
":",
"dict",
"=",
"None",
",",
"filt_dflt_incl",
":",
"bool",
"=",
"False",
")",
"->",
"(",
"Set",
"[",
"str",
"]",
",",
"str",
")",
":",
"LOGGER",
".",
"debu... | Get credentials from HolderProver wallet corresponding to proof request and
filter criteria; return credential identifiers from wallet and credentials json.
Return empty set and empty production for no such credentials.
:param proof_req_json: proof request json as Verifier creates; has entries ... | [
"Get",
"credentials",
"from",
"HolderProver",
"wallet",
"corresponding",
"to",
"proof",
"request",
"and",
"filter",
"criteria",
";",
"return",
"credential",
"identifiers",
"from",
"wallet",
"and",
"credentials",
"json",
".",
"Return",
"empty",
"set",
"and",
"empty... | 0b1c17cca3bd178b6e6974af84dbac1dfce5cf45 | https://github.com/PSPC-SPAC-buyandsell/von_agent/blob/0b1c17cca3bd178b6e6974af84dbac1dfce5cf45/von_agent/agent/holder_prover.py#L784-L937 | train |
PSPC-SPAC-buyandsell/von_agent | von_agent/agent/holder_prover.py | HolderProver.get_creds_by_id | async def get_creds_by_id(self, proof_req_json: str, cred_ids: set) -> str:
"""
Get creds structure from HolderProver wallet by credential identifiers.
:param proof_req_json: proof request as per get_creds() above
:param cred_ids: set of credential identifiers of interest
:retur... | python | async def get_creds_by_id(self, proof_req_json: str, cred_ids: set) -> str:
"""
Get creds structure from HolderProver wallet by credential identifiers.
:param proof_req_json: proof request as per get_creds() above
:param cred_ids: set of credential identifiers of interest
:retur... | [
"async",
"def",
"get_creds_by_id",
"(",
"self",
",",
"proof_req_json",
":",
"str",
",",
"cred_ids",
":",
"set",
")",
"->",
"str",
":",
"LOGGER",
".",
"debug",
"(",
"'HolderProver.get_creds_by_id >>> proof_req_json: %s, cred_ids: %s'",
",",
"proof_req_json",
",",
"cr... | Get creds structure from HolderProver wallet by credential identifiers.
:param proof_req_json: proof request as per get_creds() above
:param cred_ids: set of credential identifiers of interest
:return: json with cred(s) for input credential identifier(s) | [
"Get",
"creds",
"structure",
"from",
"HolderProver",
"wallet",
"by",
"credential",
"identifiers",
"."
] | 0b1c17cca3bd178b6e6974af84dbac1dfce5cf45 | https://github.com/PSPC-SPAC-buyandsell/von_agent/blob/0b1c17cca3bd178b6e6974af84dbac1dfce5cf45/von_agent/agent/holder_prover.py#L939-L955 | train |
palantir/typedjsonrpc | contrib/multi-module-example/typedjsonrpc_example/valid.py | histogram | def histogram(data):
"""Returns a histogram of your data.
:param data: The data to histogram
:type data: list[object]
:return: The histogram
:rtype: dict[object, int]
"""
ret = {}
for datum in data:
if datum in ret:
ret[datum] += 1
else:
ret[datum... | python | def histogram(data):
"""Returns a histogram of your data.
:param data: The data to histogram
:type data: list[object]
:return: The histogram
:rtype: dict[object, int]
"""
ret = {}
for datum in data:
if datum in ret:
ret[datum] += 1
else:
ret[datum... | [
"def",
"histogram",
"(",
"data",
")",
":",
"ret",
"=",
"{",
"}",
"for",
"datum",
"in",
"data",
":",
"if",
"datum",
"in",
"ret",
":",
"ret",
"[",
"datum",
"]",
"+=",
"1",
"else",
":",
"ret",
"[",
"datum",
"]",
"=",
"1",
"return",
"ret"
] | Returns a histogram of your data.
:param data: The data to histogram
:type data: list[object]
:return: The histogram
:rtype: dict[object, int] | [
"Returns",
"a",
"histogram",
"of",
"your",
"data",
"."
] | 274218fcd236ff9643506caa629029c9ba25a0fb | https://github.com/palantir/typedjsonrpc/blob/274218fcd236ff9643506caa629029c9ba25a0fb/contrib/multi-module-example/typedjsonrpc_example/valid.py#L22-L36 | train |
palantir/typedjsonrpc | contrib/multi-module-example/typedjsonrpc_example/valid.py | print_data | def print_data(data):
"""Prints object key-value pairs in a custom format
:param data: The dict to print
:type data: dict
:rtype: None
"""
print(", ".join(["{}=>{}".format(key, value) for key, value in data])) | python | def print_data(data):
"""Prints object key-value pairs in a custom format
:param data: The dict to print
:type data: dict
:rtype: None
"""
print(", ".join(["{}=>{}".format(key, value) for key, value in data])) | [
"def",
"print_data",
"(",
"data",
")",
":",
"print",
"(",
"\", \"",
".",
"join",
"(",
"[",
"\"{}=>{}\"",
".",
"format",
"(",
"key",
",",
"value",
")",
"for",
"key",
",",
"value",
"in",
"data",
"]",
")",
")"
] | Prints object key-value pairs in a custom format
:param data: The dict to print
:type data: dict
:rtype: None | [
"Prints",
"object",
"key",
"-",
"value",
"pairs",
"in",
"a",
"custom",
"format"
] | 274218fcd236ff9643506caa629029c9ba25a0fb | https://github.com/palantir/typedjsonrpc/blob/274218fcd236ff9643506caa629029c9ba25a0fb/contrib/multi-module-example/typedjsonrpc_example/valid.py#L65-L72 | train |
yamins81/tabular | setup.py | subdir_findall | def subdir_findall(dir, subdir):
"""
Find all files in a subdirectory and return paths relative to dir
This is similar to (and uses) setuptools.findall
However, the paths returned are in the form needed for package_data
"""
strip_n = len(dir.split('/'))
path = '/'.join((dir, subdir))
re... | python | def subdir_findall(dir, subdir):
"""
Find all files in a subdirectory and return paths relative to dir
This is similar to (and uses) setuptools.findall
However, the paths returned are in the form needed for package_data
"""
strip_n = len(dir.split('/'))
path = '/'.join((dir, subdir))
re... | [
"def",
"subdir_findall",
"(",
"dir",
",",
"subdir",
")",
":",
"strip_n",
"=",
"len",
"(",
"dir",
".",
"split",
"(",
"'/'",
")",
")",
"path",
"=",
"'/'",
".",
"join",
"(",
"(",
"dir",
",",
"subdir",
")",
")",
"return",
"[",
"'/'",
".",
"join",
"... | Find all files in a subdirectory and return paths relative to dir
This is similar to (and uses) setuptools.findall
However, the paths returned are in the form needed for package_data | [
"Find",
"all",
"files",
"in",
"a",
"subdirectory",
"and",
"return",
"paths",
"relative",
"to",
"dir"
] | 1caf091c8c395960a9ad7078f95158b533cc52dd | https://github.com/yamins81/tabular/blob/1caf091c8c395960a9ad7078f95158b533cc52dd/setup.py#L82-L91 | train |
yamins81/tabular | setup.py | find_package_data | def find_package_data(packages):
"""
For a list of packages, find the package_data
This function scans the subdirectories of a package and considers all
non-submodule subdirectories as resources, including them in
the package_data
Returns a dictionary suitable for setup(package_data=<result>)
... | python | def find_package_data(packages):
"""
For a list of packages, find the package_data
This function scans the subdirectories of a package and considers all
non-submodule subdirectories as resources, including them in
the package_data
Returns a dictionary suitable for setup(package_data=<result>)
... | [
"def",
"find_package_data",
"(",
"packages",
")",
":",
"package_data",
"=",
"{",
"}",
"for",
"package",
"in",
"packages",
":",
"package_data",
"[",
"package",
"]",
"=",
"[",
"]",
"for",
"subdir",
"in",
"find_subdirectories",
"(",
"package",
")",
":",
"if",... | For a list of packages, find the package_data
This function scans the subdirectories of a package and considers all
non-submodule subdirectories as resources, including them in
the package_data
Returns a dictionary suitable for setup(package_data=<result>) | [
"For",
"a",
"list",
"of",
"packages",
"find",
"the",
"package_data"
] | 1caf091c8c395960a9ad7078f95158b533cc52dd | https://github.com/yamins81/tabular/blob/1caf091c8c395960a9ad7078f95158b533cc52dd/setup.py#L93-L114 | train |
finklabs/metrics | metrics/metrics_utils.py | process_file_metrics | def process_file_metrics(context, file_processors):
"""Main routine for metrics."""
file_metrics = OrderedDict()
# TODO make available the includes and excludes feature
gitignore = []
if os.path.isfile('.gitignore'):
with open('.gitignore', 'r') as ifile:
gitignore = ifile.read(... | python | def process_file_metrics(context, file_processors):
"""Main routine for metrics."""
file_metrics = OrderedDict()
# TODO make available the includes and excludes feature
gitignore = []
if os.path.isfile('.gitignore'):
with open('.gitignore', 'r') as ifile:
gitignore = ifile.read(... | [
"def",
"process_file_metrics",
"(",
"context",
",",
"file_processors",
")",
":",
"file_metrics",
"=",
"OrderedDict",
"(",
")",
"# TODO make available the includes and excludes feature",
"gitignore",
"=",
"[",
"]",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"'.giti... | Main routine for metrics. | [
"Main",
"routine",
"for",
"metrics",
"."
] | fd9974af498831664b9ae8e8f3834e1ec2e8a699 | https://github.com/finklabs/metrics/blob/fd9974af498831664b9ae8e8f3834e1ec2e8a699/metrics/metrics_utils.py#L117-L150 | train |
finklabs/metrics | metrics/metrics_utils.py | process_build_metrics | def process_build_metrics(context, build_processors):
"""use processors to collect build metrics."""
build_metrics = OrderedDict()
# reset all processors
for p in build_processors:
p.reset()
# collect metrics from all processors
for p in build_processors:
build_metrics.update(p... | python | def process_build_metrics(context, build_processors):
"""use processors to collect build metrics."""
build_metrics = OrderedDict()
# reset all processors
for p in build_processors:
p.reset()
# collect metrics from all processors
for p in build_processors:
build_metrics.update(p... | [
"def",
"process_build_metrics",
"(",
"context",
",",
"build_processors",
")",
":",
"build_metrics",
"=",
"OrderedDict",
"(",
")",
"# reset all processors",
"for",
"p",
"in",
"build_processors",
":",
"p",
".",
"reset",
"(",
")",
"# collect metrics from all processors",... | use processors to collect build metrics. | [
"use",
"processors",
"to",
"collect",
"build",
"metrics",
"."
] | fd9974af498831664b9ae8e8f3834e1ec2e8a699 | https://github.com/finklabs/metrics/blob/fd9974af498831664b9ae8e8f3834e1ec2e8a699/metrics/metrics_utils.py#L153-L165 | train |
finklabs/metrics | metrics/metrics_utils.py | summary | def summary(processors, metrics, context):
"""Print the summary"""
# display aggregated metric values on language level
def display_header(processors, before='', after=''):
"""Display the header for the summary results."""
print(before, end=' ')
for processor in processors:
... | python | def summary(processors, metrics, context):
"""Print the summary"""
# display aggregated metric values on language level
def display_header(processors, before='', after=''):
"""Display the header for the summary results."""
print(before, end=' ')
for processor in processors:
... | [
"def",
"summary",
"(",
"processors",
",",
"metrics",
",",
"context",
")",
":",
"# display aggregated metric values on language level",
"def",
"display_header",
"(",
"processors",
",",
"before",
"=",
"''",
",",
"after",
"=",
"''",
")",
":",
"\"\"\"Display the header ... | Print the summary | [
"Print",
"the",
"summary"
] | fd9974af498831664b9ae8e8f3834e1ec2e8a699 | https://github.com/finklabs/metrics/blob/fd9974af498831664b9ae8e8f3834e1ec2e8a699/metrics/metrics_utils.py#L168-L224 | train |
rwl/pylon | examples/pyreto/thesis/ex6_1.py | get_portfolios3 | def get_portfolios3():
""" Returns portfolios with U12 and U20 generators removed and generators
of the same type at the same bus aggregated.
"""
g1 = [0]
g2 = [1]
g7 = [2]
g13 = [3]
g14 = [4] # sync cond
g15 = [5]
g16 = [6]
g18 = [7]
g21 = [8]
g22 = [9]
g23 = [10... | python | def get_portfolios3():
""" Returns portfolios with U12 and U20 generators removed and generators
of the same type at the same bus aggregated.
"""
g1 = [0]
g2 = [1]
g7 = [2]
g13 = [3]
g14 = [4] # sync cond
g15 = [5]
g16 = [6]
g18 = [7]
g21 = [8]
g22 = [9]
g23 = [10... | [
"def",
"get_portfolios3",
"(",
")",
":",
"g1",
"=",
"[",
"0",
"]",
"g2",
"=",
"[",
"1",
"]",
"g7",
"=",
"[",
"2",
"]",
"g13",
"=",
"[",
"3",
"]",
"g14",
"=",
"[",
"4",
"]",
"# sync cond",
"g15",
"=",
"[",
"5",
"]",
"g16",
"=",
"[",
"6",
... | Returns portfolios with U12 and U20 generators removed and generators
of the same type at the same bus aggregated. | [
"Returns",
"portfolios",
"with",
"U12",
"and",
"U20",
"generators",
"removed",
"and",
"generators",
"of",
"the",
"same",
"type",
"at",
"the",
"same",
"bus",
"aggregated",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/ex6_1.py#L95-L118 | train |
pymoca/pymoca | src/pymoca/backends/xml/parser.py | ModelListener.call | def call(self, tag_name: str, *args, **kwargs):
"""Convenience method for calling methods with walker."""
if hasattr(self, tag_name):
getattr(self, tag_name)(*args, **kwargs) | python | def call(self, tag_name: str, *args, **kwargs):
"""Convenience method for calling methods with walker."""
if hasattr(self, tag_name):
getattr(self, tag_name)(*args, **kwargs) | [
"def",
"call",
"(",
"self",
",",
"tag_name",
":",
"str",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"hasattr",
"(",
"self",
",",
"tag_name",
")",
":",
"getattr",
"(",
"self",
",",
"tag_name",
")",
"(",
"*",
"args",
",",
"*",
"*"... | Convenience method for calling methods with walker. | [
"Convenience",
"method",
"for",
"calling",
"methods",
"with",
"walker",
"."
] | 14b5eb7425e96689de6cc5c10f400895d586a978 | https://github.com/pymoca/pymoca/blob/14b5eb7425e96689de6cc5c10f400895d586a978/src/pymoca/backends/xml/parser.py#L85-L88 | train |
pymoca/pymoca | src/pymoca/backends/xml/parser.py | ModelListener.der | def der(self, x: Sym):
"""Get the derivative of the variable, create it if it doesn't exist."""
name = 'der({:s})'.format(x.name())
if name not in self.scope['dvar'].keys():
self.scope['dvar'][name] = self.sym.sym(name, *x.shape)
self.scope['states'].append(x.name())
... | python | def der(self, x: Sym):
"""Get the derivative of the variable, create it if it doesn't exist."""
name = 'der({:s})'.format(x.name())
if name not in self.scope['dvar'].keys():
self.scope['dvar'][name] = self.sym.sym(name, *x.shape)
self.scope['states'].append(x.name())
... | [
"def",
"der",
"(",
"self",
",",
"x",
":",
"Sym",
")",
":",
"name",
"=",
"'der({:s})'",
".",
"format",
"(",
"x",
".",
"name",
"(",
")",
")",
"if",
"name",
"not",
"in",
"self",
".",
"scope",
"[",
"'dvar'",
"]",
".",
"keys",
"(",
")",
":",
"self... | Get the derivative of the variable, create it if it doesn't exist. | [
"Get",
"the",
"derivative",
"of",
"the",
"variable",
"create",
"it",
"if",
"it",
"doesn",
"t",
"exist",
"."
] | 14b5eb7425e96689de6cc5c10f400895d586a978 | https://github.com/pymoca/pymoca/blob/14b5eb7425e96689de6cc5c10f400895d586a978/src/pymoca/backends/xml/parser.py#L94-L100 | train |
pymoca/pymoca | src/pymoca/backends/xml/parser.py | ModelListener.noise_gaussian | def noise_gaussian(self, mean, std):
"""Create a gaussian noise variable"""
assert std > 0
ng = self.sym.sym('ng_{:d}'.format(len(self.scope['ng'])))
self.scope['ng'].append(ng)
return mean + std*ng | python | def noise_gaussian(self, mean, std):
"""Create a gaussian noise variable"""
assert std > 0
ng = self.sym.sym('ng_{:d}'.format(len(self.scope['ng'])))
self.scope['ng'].append(ng)
return mean + std*ng | [
"def",
"noise_gaussian",
"(",
"self",
",",
"mean",
",",
"std",
")",
":",
"assert",
"std",
">",
"0",
"ng",
"=",
"self",
".",
"sym",
".",
"sym",
"(",
"'ng_{:d}'",
".",
"format",
"(",
"len",
"(",
"self",
".",
"scope",
"[",
"'ng'",
"]",
")",
")",
"... | Create a gaussian noise variable | [
"Create",
"a",
"gaussian",
"noise",
"variable"
] | 14b5eb7425e96689de6cc5c10f400895d586a978 | https://github.com/pymoca/pymoca/blob/14b5eb7425e96689de6cc5c10f400895d586a978/src/pymoca/backends/xml/parser.py#L126-L131 | train |
pymoca/pymoca | src/pymoca/backends/xml/parser.py | ModelListener.noise_uniform | def noise_uniform(self, lower_bound, upper_bound):
"""Create a uniform noise variable"""
assert upper_bound > lower_bound
nu = self.sym.sym('nu_{:d}'.format(len(self.scope['nu'])))
self.scope['nu'].append(nu)
return lower_bound + nu*(upper_bound - lower_bound) | python | def noise_uniform(self, lower_bound, upper_bound):
"""Create a uniform noise variable"""
assert upper_bound > lower_bound
nu = self.sym.sym('nu_{:d}'.format(len(self.scope['nu'])))
self.scope['nu'].append(nu)
return lower_bound + nu*(upper_bound - lower_bound) | [
"def",
"noise_uniform",
"(",
"self",
",",
"lower_bound",
",",
"upper_bound",
")",
":",
"assert",
"upper_bound",
">",
"lower_bound",
"nu",
"=",
"self",
".",
"sym",
".",
"sym",
"(",
"'nu_{:d}'",
".",
"format",
"(",
"len",
"(",
"self",
".",
"scope",
"[",
... | Create a uniform noise variable | [
"Create",
"a",
"uniform",
"noise",
"variable"
] | 14b5eb7425e96689de6cc5c10f400895d586a978 | https://github.com/pymoca/pymoca/blob/14b5eb7425e96689de6cc5c10f400895d586a978/src/pymoca/backends/xml/parser.py#L133-L138 | train |
pymoca/pymoca | src/pymoca/backends/xml/parser.py | ModelListener.log | def log(self, *args, **kwargs):
"""Convenience function for printing indenting debug output."""
if self.verbose:
print(' ' * self.depth, *args, **kwargs) | python | def log(self, *args, **kwargs):
"""Convenience function for printing indenting debug output."""
if self.verbose:
print(' ' * self.depth, *args, **kwargs) | [
"def",
"log",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"self",
".",
"verbose",
":",
"print",
"(",
"' '",
"*",
"self",
".",
"depth",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | Convenience function for printing indenting debug output. | [
"Convenience",
"function",
"for",
"printing",
"indenting",
"debug",
"output",
"."
] | 14b5eb7425e96689de6cc5c10f400895d586a978 | https://github.com/pymoca/pymoca/blob/14b5eb7425e96689de6cc5c10f400895d586a978/src/pymoca/backends/xml/parser.py#L158-L161 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_case6ww | def get_case6ww():
""" Returns the 6 bus case from Wood & Wollenberg PG&C.
"""
path = os.path.dirname(pylon.__file__)
path = os.path.join(path, "test", "data")
path = os.path.join(path, "case6ww", "case6ww.pkl")
case = pylon.Case.load(path)
case.generators[0].p_cost = (0.0, 4.0, 200.0)
... | python | def get_case6ww():
""" Returns the 6 bus case from Wood & Wollenberg PG&C.
"""
path = os.path.dirname(pylon.__file__)
path = os.path.join(path, "test", "data")
path = os.path.join(path, "case6ww", "case6ww.pkl")
case = pylon.Case.load(path)
case.generators[0].p_cost = (0.0, 4.0, 200.0)
... | [
"def",
"get_case6ww",
"(",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"pylon",
".",
"__file__",
")",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"path",
",",
"\"test\"",
",",
"\"data\"",
")",
"path",
"=",
"os",
".",
"pat... | Returns the 6 bus case from Wood & Wollenberg PG&C. | [
"Returns",
"the",
"6",
"bus",
"case",
"from",
"Wood",
"&",
"Wollenberg",
"PG&C",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L36-L70 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_case24_ieee_rts | def get_case24_ieee_rts():
""" Returns the 24 bus IEEE Reliability Test System.
"""
path = os.path.dirname(pylon.__file__)
path = os.path.join(path, "test", "data")
path = os.path.join(path, "case24_ieee_rts", "case24_ieee_rts.pkl")
case = pylon.Case.load(path)
# FIXME: Correct generator n... | python | def get_case24_ieee_rts():
""" Returns the 24 bus IEEE Reliability Test System.
"""
path = os.path.dirname(pylon.__file__)
path = os.path.join(path, "test", "data")
path = os.path.join(path, "case24_ieee_rts", "case24_ieee_rts.pkl")
case = pylon.Case.load(path)
# FIXME: Correct generator n... | [
"def",
"get_case24_ieee_rts",
"(",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"pylon",
".",
"__file__",
")",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"path",
",",
"\"test\"",
",",
"\"data\"",
")",
"path",
"=",
"os",
"."... | Returns the 24 bus IEEE Reliability Test System. | [
"Returns",
"the",
"24",
"bus",
"IEEE",
"Reliability",
"Test",
"System",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L107-L120 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_discrete_task_agent | def get_discrete_task_agent(generators, market, nStates, nOffer, markups,
withholds, maxSteps, learner, Pd0=None, Pd_min=0.0):
""" Returns a tuple of task and agent for the given learner.
"""
env = pyreto.discrete.MarketEnvironment(generators, market,
numS... | python | def get_discrete_task_agent(generators, market, nStates, nOffer, markups,
withholds, maxSteps, learner, Pd0=None, Pd_min=0.0):
""" Returns a tuple of task and agent for the given learner.
"""
env = pyreto.discrete.MarketEnvironment(generators, market,
numS... | [
"def",
"get_discrete_task_agent",
"(",
"generators",
",",
"market",
",",
"nStates",
",",
"nOffer",
",",
"markups",
",",
"withholds",
",",
"maxSteps",
",",
"learner",
",",
"Pd0",
"=",
"None",
",",
"Pd_min",
"=",
"0.0",
")",
":",
"env",
"=",
"pyreto",
".",... | Returns a tuple of task and agent for the given learner. | [
"Returns",
"a",
"tuple",
"of",
"task",
"and",
"agent",
"for",
"the",
"given",
"learner",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L166-L184 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_zero_task_agent | def get_zero_task_agent(generators, market, nOffer, maxSteps):
""" Returns a task-agent tuple whose action is always zero.
"""
env = pyreto.discrete.MarketEnvironment(generators, market, nOffer)
task = pyreto.discrete.ProfitTask(env, maxSteps=maxSteps)
agent = pyreto.util.ZeroAgent(env.outdim, env.i... | python | def get_zero_task_agent(generators, market, nOffer, maxSteps):
""" Returns a task-agent tuple whose action is always zero.
"""
env = pyreto.discrete.MarketEnvironment(generators, market, nOffer)
task = pyreto.discrete.ProfitTask(env, maxSteps=maxSteps)
agent = pyreto.util.ZeroAgent(env.outdim, env.i... | [
"def",
"get_zero_task_agent",
"(",
"generators",
",",
"market",
",",
"nOffer",
",",
"maxSteps",
")",
":",
"env",
"=",
"pyreto",
".",
"discrete",
".",
"MarketEnvironment",
"(",
"generators",
",",
"market",
",",
"nOffer",
")",
"task",
"=",
"pyreto",
".",
"di... | Returns a task-agent tuple whose action is always zero. | [
"Returns",
"a",
"task",
"-",
"agent",
"tuple",
"whose",
"action",
"is",
"always",
"zero",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L211-L217 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_neg_one_task_agent | def get_neg_one_task_agent(generators, market, nOffer, maxSteps):
""" Returns a task-agent tuple whose action is always minus one.
"""
env = pyreto.discrete.MarketEnvironment(generators, market, nOffer)
task = pyreto.discrete.ProfitTask(env, maxSteps=maxSteps)
agent = pyreto.util.NegOneAgent(env.out... | python | def get_neg_one_task_agent(generators, market, nOffer, maxSteps):
""" Returns a task-agent tuple whose action is always minus one.
"""
env = pyreto.discrete.MarketEnvironment(generators, market, nOffer)
task = pyreto.discrete.ProfitTask(env, maxSteps=maxSteps)
agent = pyreto.util.NegOneAgent(env.out... | [
"def",
"get_neg_one_task_agent",
"(",
"generators",
",",
"market",
",",
"nOffer",
",",
"maxSteps",
")",
":",
"env",
"=",
"pyreto",
".",
"discrete",
".",
"MarketEnvironment",
"(",
"generators",
",",
"market",
",",
"nOffer",
")",
"task",
"=",
"pyreto",
".",
... | Returns a task-agent tuple whose action is always minus one. | [
"Returns",
"a",
"task",
"-",
"agent",
"tuple",
"whose",
"action",
"is",
"always",
"minus",
"one",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L220-L226 | train |
rwl/pylon | examples/pyreto/thesis/common.py | run_experiment | def run_experiment(experiment, roleouts, episodes, in_cloud=False,
dynProfile=None):
""" Runs the given experiment and returns the results.
"""
def run():
if dynProfile is None:
maxsteps = len(experiment.profile) # episode length
else:
maxsteps = dy... | python | def run_experiment(experiment, roleouts, episodes, in_cloud=False,
dynProfile=None):
""" Runs the given experiment and returns the results.
"""
def run():
if dynProfile is None:
maxsteps = len(experiment.profile) # episode length
else:
maxsteps = dy... | [
"def",
"run_experiment",
"(",
"experiment",
",",
"roleouts",
",",
"episodes",
",",
"in_cloud",
"=",
"False",
",",
"dynProfile",
"=",
"None",
")",
":",
"def",
"run",
"(",
")",
":",
"if",
"dynProfile",
"is",
"None",
":",
"maxsteps",
"=",
"len",
"(",
"exp... | Runs the given experiment and returns the results. | [
"Runs",
"the",
"given",
"experiment",
"and",
"returns",
"the",
"results",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L241-L314 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_full_year | def get_full_year():
""" Returns percentages of peak load for all hours of the year.
@return:
Numpy array of doubles with length 8736.
"""
weekly = get_weekly()
daily = get_daily()
hourly_winter_wkdy, hourly_winter_wknd = get_winter_hourly()
hourly_summer_wkdy, hourly_summer_wknd = ... | python | def get_full_year():
""" Returns percentages of peak load for all hours of the year.
@return:
Numpy array of doubles with length 8736.
"""
weekly = get_weekly()
daily = get_daily()
hourly_winter_wkdy, hourly_winter_wknd = get_winter_hourly()
hourly_summer_wkdy, hourly_summer_wknd = ... | [
"def",
"get_full_year",
"(",
")",
":",
"weekly",
"=",
"get_weekly",
"(",
")",
"daily",
"=",
"get_daily",
"(",
")",
"hourly_winter_wkdy",
",",
"hourly_winter_wknd",
"=",
"get_winter_hourly",
"(",
")",
"hourly_summer_wkdy",
",",
"hourly_summer_wknd",
"=",
"get_summe... | Returns percentages of peak load for all hours of the year.
@return:
Numpy array of doubles with length 8736. | [
"Returns",
"percentages",
"of",
"peak",
"load",
"for",
"all",
"hours",
"of",
"the",
"year",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L426-L457 | train |
rwl/pylon | examples/pyreto/thesis/common.py | get_all_days | def get_all_days():
""" Returns percentages of peak load for all days of the year.
Data from the IEEE RTS.
"""
weekly = get_weekly()
daily = get_daily()
return [w * (d / 100.0) for w in weekly for d in daily] | python | def get_all_days():
""" Returns percentages of peak load for all days of the year.
Data from the IEEE RTS.
"""
weekly = get_weekly()
daily = get_daily()
return [w * (d / 100.0) for w in weekly for d in daily] | [
"def",
"get_all_days",
"(",
")",
":",
"weekly",
"=",
"get_weekly",
"(",
")",
"daily",
"=",
"get_daily",
"(",
")",
"return",
"[",
"w",
"*",
"(",
"d",
"/",
"100.0",
")",
"for",
"w",
"in",
"weekly",
"for",
"d",
"in",
"daily",
"]"
] | Returns percentages of peak load for all days of the year.
Data from the IEEE RTS. | [
"Returns",
"percentages",
"of",
"peak",
"load",
"for",
"all",
"days",
"of",
"the",
"year",
".",
"Data",
"from",
"the",
"IEEE",
"RTS",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/common.py#L460-L467 | train |
rwl/pylon | examples/pyreto/thesis/ex5_1.py | get_q_experiment | def get_q_experiment(case, minor=1):
""" Returns an experiment that uses Q-learning.
"""
gen = case.generators
profile = array([1.0])
maxSteps = len(profile)
if minor == 1:
alpha = 0.3 # Learning rate.
gamma = 0.99 # Discount factor
# The closer epsilon gets to 0, the m... | python | def get_q_experiment(case, minor=1):
""" Returns an experiment that uses Q-learning.
"""
gen = case.generators
profile = array([1.0])
maxSteps = len(profile)
if minor == 1:
alpha = 0.3 # Learning rate.
gamma = 0.99 # Discount factor
# The closer epsilon gets to 0, the m... | [
"def",
"get_q_experiment",
"(",
"case",
",",
"minor",
"=",
"1",
")",
":",
"gen",
"=",
"case",
".",
"generators",
"profile",
"=",
"array",
"(",
"[",
"1.0",
"]",
")",
"maxSteps",
"=",
"len",
"(",
"profile",
")",
"if",
"minor",
"==",
"1",
":",
"alpha"... | Returns an experiment that uses Q-learning. | [
"Returns",
"an",
"experiment",
"that",
"uses",
"Q",
"-",
"learning",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/examples/pyreto/thesis/ex5_1.py#L79-L133 | train |
rwl/pylon | pylon/generator.py | Generator.q_limited | def q_limited(self):
""" Is the machine at it's limit of reactive power?
"""
if (self.q >= self.q_max) or (self.q <= self.q_min):
return True
else:
return False | python | def q_limited(self):
""" Is the machine at it's limit of reactive power?
"""
if (self.q >= self.q_max) or (self.q <= self.q_min):
return True
else:
return False | [
"def",
"q_limited",
"(",
"self",
")",
":",
"if",
"(",
"self",
".",
"q",
">=",
"self",
".",
"q_max",
")",
"or",
"(",
"self",
".",
"q",
"<=",
"self",
".",
"q_min",
")",
":",
"return",
"True",
"else",
":",
"return",
"False"
] | Is the machine at it's limit of reactive power? | [
"Is",
"the",
"machine",
"at",
"it",
"s",
"limit",
"of",
"reactive",
"power?"
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L156-L162 | train |
rwl/pylon | pylon/generator.py | Generator.total_cost | def total_cost(self, p=None, p_cost=None, pcost_model=None):
""" Computes total cost for the generator at the given output level.
"""
p = self.p if p is None else p
p_cost = self.p_cost if p_cost is None else p_cost
pcost_model = self.pcost_model if pcost_model is None else pcost... | python | def total_cost(self, p=None, p_cost=None, pcost_model=None):
""" Computes total cost for the generator at the given output level.
"""
p = self.p if p is None else p
p_cost = self.p_cost if p_cost is None else p_cost
pcost_model = self.pcost_model if pcost_model is None else pcost... | [
"def",
"total_cost",
"(",
"self",
",",
"p",
"=",
"None",
",",
"p_cost",
"=",
"None",
",",
"pcost_model",
"=",
"None",
")",
":",
"p",
"=",
"self",
".",
"p",
"if",
"p",
"is",
"None",
"else",
"p",
"p_cost",
"=",
"self",
".",
"p_cost",
"if",
"p_cost"... | Computes total cost for the generator at the given output level. | [
"Computes",
"total",
"cost",
"for",
"the",
"generator",
"at",
"the",
"given",
"output",
"level",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L180-L220 | train |
rwl/pylon | pylon/generator.py | Generator.poly_to_pwl | def poly_to_pwl(self, n_points=4):
""" Sets the piece-wise linear cost attribute, converting the
polynomial cost variable by evaluating at zero and then at n_points
evenly spaced points between p_min and p_max.
"""
assert self.pcost_model == POLYNOMIAL
p_min = self.p_min
... | python | def poly_to_pwl(self, n_points=4):
""" Sets the piece-wise linear cost attribute, converting the
polynomial cost variable by evaluating at zero and then at n_points
evenly spaced points between p_min and p_max.
"""
assert self.pcost_model == POLYNOMIAL
p_min = self.p_min
... | [
"def",
"poly_to_pwl",
"(",
"self",
",",
"n_points",
"=",
"4",
")",
":",
"assert",
"self",
".",
"pcost_model",
"==",
"POLYNOMIAL",
"p_min",
"=",
"self",
".",
"p_min",
"p_max",
"=",
"self",
".",
"p_max",
"p_cost",
"=",
"[",
"]",
"if",
"p_min",
">",
"0.... | Sets the piece-wise linear cost attribute, converting the
polynomial cost variable by evaluating at zero and then at n_points
evenly spaced points between p_min and p_max. | [
"Sets",
"the",
"piece",
"-",
"wise",
"linear",
"cost",
"attribute",
"converting",
"the",
"polynomial",
"cost",
"variable",
"by",
"evaluating",
"at",
"zero",
"and",
"then",
"at",
"n_points",
"evenly",
"spaced",
"points",
"between",
"p_min",
"and",
"p_max",
"."
... | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L278-L308 | train |
rwl/pylon | pylon/generator.py | Generator.get_offers | def get_offers(self, n_points=6):
""" Returns quantity and price offers created from the cost function.
"""
from pyreto.smart_market import Offer
qtyprc = self._get_qtyprc(n_points)
return [Offer(self, qty, prc) for qty, prc in qtyprc] | python | def get_offers(self, n_points=6):
""" Returns quantity and price offers created from the cost function.
"""
from pyreto.smart_market import Offer
qtyprc = self._get_qtyprc(n_points)
return [Offer(self, qty, prc) for qty, prc in qtyprc] | [
"def",
"get_offers",
"(",
"self",
",",
"n_points",
"=",
"6",
")",
":",
"from",
"pyreto",
".",
"smart_market",
"import",
"Offer",
"qtyprc",
"=",
"self",
".",
"_get_qtyprc",
"(",
"n_points",
")",
"return",
"[",
"Offer",
"(",
"self",
",",
"qty",
",",
"prc... | Returns quantity and price offers created from the cost function. | [
"Returns",
"quantity",
"and",
"price",
"offers",
"created",
"from",
"the",
"cost",
"function",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L311-L317 | train |
rwl/pylon | pylon/generator.py | Generator.get_bids | def get_bids(self, n_points=6):
""" Returns quantity and price bids created from the cost function.
"""
from pyreto.smart_market import Bid
qtyprc = self._get_qtyprc(n_points)
return [Bid(self, qty, prc) for qty, prc in qtyprc] | python | def get_bids(self, n_points=6):
""" Returns quantity and price bids created from the cost function.
"""
from pyreto.smart_market import Bid
qtyprc = self._get_qtyprc(n_points)
return [Bid(self, qty, prc) for qty, prc in qtyprc] | [
"def",
"get_bids",
"(",
"self",
",",
"n_points",
"=",
"6",
")",
":",
"from",
"pyreto",
".",
"smart_market",
"import",
"Bid",
"qtyprc",
"=",
"self",
".",
"_get_qtyprc",
"(",
"n_points",
")",
"return",
"[",
"Bid",
"(",
"self",
",",
"qty",
",",
"prc",
"... | Returns quantity and price bids created from the cost function. | [
"Returns",
"quantity",
"and",
"price",
"bids",
"created",
"from",
"the",
"cost",
"function",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L320-L326 | train |
rwl/pylon | pylon/generator.py | Generator.offers_to_pwl | def offers_to_pwl(self, offers):
""" Updates the piece-wise linear total cost function using the given
offer blocks.
Based on off2case.m from MATPOWER by Ray Zimmerman, developed at PSERC
Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.
"""
assert no... | python | def offers_to_pwl(self, offers):
""" Updates the piece-wise linear total cost function using the given
offer blocks.
Based on off2case.m from MATPOWER by Ray Zimmerman, developed at PSERC
Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.
"""
assert no... | [
"def",
"offers_to_pwl",
"(",
"self",
",",
"offers",
")",
":",
"assert",
"not",
"self",
".",
"is_load",
"# Only apply offers associated with this generator.",
"g_offers",
"=",
"[",
"offer",
"for",
"offer",
"in",
"offers",
"if",
"offer",
".",
"generator",
"==",
"s... | Updates the piece-wise linear total cost function using the given
offer blocks.
Based on off2case.m from MATPOWER by Ray Zimmerman, developed at PSERC
Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info. | [
"Updates",
"the",
"piece",
"-",
"wise",
"linear",
"total",
"cost",
"function",
"using",
"the",
"given",
"offer",
"blocks",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L420-L466 | train |
rwl/pylon | pylon/generator.py | Generator.bids_to_pwl | def bids_to_pwl(self, bids):
""" Updates the piece-wise linear total cost function using the given
bid blocks.
Based on off2case.m from MATPOWER by Ray Zimmerman, developed at PSERC
Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.
"""
assert self.is_... | python | def bids_to_pwl(self, bids):
""" Updates the piece-wise linear total cost function using the given
bid blocks.
Based on off2case.m from MATPOWER by Ray Zimmerman, developed at PSERC
Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info.
"""
assert self.is_... | [
"def",
"bids_to_pwl",
"(",
"self",
",",
"bids",
")",
":",
"assert",
"self",
".",
"is_load",
"# Apply only those bids associated with this dispatchable load.",
"vl_bids",
"=",
"[",
"bid",
"for",
"bid",
"in",
"bids",
"if",
"bid",
".",
"vLoad",
"==",
"self",
"]",
... | Updates the piece-wise linear total cost function using the given
bid blocks.
Based on off2case.m from MATPOWER by Ray Zimmerman, developed at PSERC
Cornell. See U{http://www.pserc.cornell.edu/matpower/} for more info. | [
"Updates",
"the",
"piece",
"-",
"wise",
"linear",
"total",
"cost",
"function",
"using",
"the",
"given",
"bid",
"blocks",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L469-L508 | train |
rwl/pylon | pylon/generator.py | Generator._adjust_limits | def _adjust_limits(self):
""" Sets the active power limits, 'p_max' and 'p_min', according to
the pwl cost function points.
"""
if not self.is_load:
# self.p_min = min([point[0] for point in self.p_cost])
self.p_max = max([point[0] for point in self.p_cost])
... | python | def _adjust_limits(self):
""" Sets the active power limits, 'p_max' and 'p_min', according to
the pwl cost function points.
"""
if not self.is_load:
# self.p_min = min([point[0] for point in self.p_cost])
self.p_max = max([point[0] for point in self.p_cost])
... | [
"def",
"_adjust_limits",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"is_load",
":",
"# self.p_min = min([point[0] for point in self.p_cost])",
"self",
".",
"p_max",
"=",
"max",
"(",
"[",
"point",
"[",
"0",
"]",
"for",
"point",
"in",
"self",
".... | Sets the active power limits, 'p_max' and 'p_min', according to
the pwl cost function points. | [
"Sets",
"the",
"active",
"power",
"limits",
"p_max",
"and",
"p_min",
"according",
"to",
"the",
"pwl",
"cost",
"function",
"points",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/generator.py#L537-L549 | train |
rwl/pylon | pyreto/continuous/environment.py | MarketEnvironment.indim | def indim(self):
""" The number of action values that the environment accepts.
"""
indim = self.numOffbids * len(self.generators)
if self.maxWithhold is not None:
return indim * 2
else:
return indim | python | def indim(self):
""" The number of action values that the environment accepts.
"""
indim = self.numOffbids * len(self.generators)
if self.maxWithhold is not None:
return indim * 2
else:
return indim | [
"def",
"indim",
"(",
"self",
")",
":",
"indim",
"=",
"self",
".",
"numOffbids",
"*",
"len",
"(",
"self",
".",
"generators",
")",
"if",
"self",
".",
"maxWithhold",
"is",
"not",
"None",
":",
"return",
"indim",
"*",
"2",
"else",
":",
"return",
"indim"
] | The number of action values that the environment accepts. | [
"The",
"number",
"of",
"action",
"values",
"that",
"the",
"environment",
"accepts",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pyreto/continuous/environment.py#L103-L111 | train |
rwl/pylon | pyreto/continuous/environment.py | MarketEnvironment._getBusVoltageLambdaSensor | def _getBusVoltageLambdaSensor(self):
""" Returns an array of length nb where each value is the sum of the
Lagrangian multipliers on the upper and the negative of the Lagrangian
multipliers on the lower voltage limits. """
muVmin = array([b.mu_vmin for b in self.market.case.connected_bus... | python | def _getBusVoltageLambdaSensor(self):
""" Returns an array of length nb where each value is the sum of the
Lagrangian multipliers on the upper and the negative of the Lagrangian
multipliers on the lower voltage limits. """
muVmin = array([b.mu_vmin for b in self.market.case.connected_bus... | [
"def",
"_getBusVoltageLambdaSensor",
"(",
"self",
")",
":",
"muVmin",
"=",
"array",
"(",
"[",
"b",
".",
"mu_vmin",
"for",
"b",
"in",
"self",
".",
"market",
".",
"case",
".",
"connected_buses",
"]",
")",
"muVmax",
"=",
"array",
"(",
"[",
"b",
".",
"mu... | Returns an array of length nb where each value is the sum of the
Lagrangian multipliers on the upper and the negative of the Lagrangian
multipliers on the lower voltage limits. | [
"Returns",
"an",
"array",
"of",
"length",
"nb",
"where",
"each",
"value",
"is",
"the",
"sum",
"of",
"the",
"Lagrangian",
"multipliers",
"on",
"the",
"upper",
"and",
"the",
"negative",
"of",
"the",
"Lagrangian",
"multipliers",
"on",
"the",
"lower",
"voltage",... | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pyreto/continuous/environment.py#L177-L185 | train |
melizalab/libtfr | site_scons/site_tools/doxygen.py | DoxyfileParse | def DoxyfileParse(file_contents):
"""
Parse a Doxygen source file and return a dictionary of all the values.
Values will be strings and lists of strings.
"""
data = {}
import shlex
lex = shlex.shlex(instream = file_contents, posix = True)
lex.wordchars += "*+./-:"
lex.whitespace = lex.whites... | python | def DoxyfileParse(file_contents):
"""
Parse a Doxygen source file and return a dictionary of all the values.
Values will be strings and lists of strings.
"""
data = {}
import shlex
lex = shlex.shlex(instream = file_contents, posix = True)
lex.wordchars += "*+./-:"
lex.whitespace = lex.whites... | [
"def",
"DoxyfileParse",
"(",
"file_contents",
")",
":",
"data",
"=",
"{",
"}",
"import",
"shlex",
"lex",
"=",
"shlex",
".",
"shlex",
"(",
"instream",
"=",
"file_contents",
",",
"posix",
"=",
"True",
")",
"lex",
".",
"wordchars",
"+=",
"\"*+./-:\"",
"lex"... | Parse a Doxygen source file and return a dictionary of all the values.
Values will be strings and lists of strings. | [
"Parse",
"a",
"Doxygen",
"source",
"file",
"and",
"return",
"a",
"dictionary",
"of",
"all",
"the",
"values",
".",
"Values",
"will",
"be",
"strings",
"and",
"lists",
"of",
"strings",
"."
] | 9f7e7705793d258a0b205f185b20e3bbcda473da | https://github.com/melizalab/libtfr/blob/9f7e7705793d258a0b205f185b20e3bbcda473da/site_scons/site_tools/doxygen.py#L28-L97 | train |
melizalab/libtfr | site_scons/site_tools/doxygen.py | DoxySourceScan | def DoxySourceScan(node, env, path):
"""
Doxygen Doxyfile source scanner. This should scan the Doxygen file and add
any files used to generate docs to the list of source files.
"""
default_file_patterns = [
'*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx',
'*.ipp', '*.i++'... | python | def DoxySourceScan(node, env, path):
"""
Doxygen Doxyfile source scanner. This should scan the Doxygen file and add
any files used to generate docs to the list of source files.
"""
default_file_patterns = [
'*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx',
'*.ipp', '*.i++'... | [
"def",
"DoxySourceScan",
"(",
"node",
",",
"env",
",",
"path",
")",
":",
"default_file_patterns",
"=",
"[",
"'*.c'",
",",
"'*.cc'",
",",
"'*.cxx'",
",",
"'*.cpp'",
",",
"'*.c++'",
",",
"'*.java'",
",",
"'*.ii'",
",",
"'*.ixx'",
",",
"'*.ipp'",
",",
"'*.i... | Doxygen Doxyfile source scanner. This should scan the Doxygen file and add
any files used to generate docs to the list of source files. | [
"Doxygen",
"Doxyfile",
"source",
"scanner",
".",
"This",
"should",
"scan",
"the",
"Doxygen",
"file",
"and",
"add",
"any",
"files",
"used",
"to",
"generate",
"docs",
"to",
"the",
"list",
"of",
"source",
"files",
"."
] | 9f7e7705793d258a0b205f185b20e3bbcda473da | https://github.com/melizalab/libtfr/blob/9f7e7705793d258a0b205f185b20e3bbcda473da/site_scons/site_tools/doxygen.py#L99-L174 | train |
melizalab/libtfr | site_scons/site_tools/doxygen.py | DoxyEmitter | def DoxyEmitter(source, target, env):
"""Doxygen Doxyfile emitter"""
# possible output formats and their default values and output locations
output_formats = {
"HTML": ("YES", "html"),
"LATEX": ("YES", "latex"),
"RTF": ("NO", "rtf"),
"MAN": ("YES", "man"),
"XML": ("NO", "xml"),
... | python | def DoxyEmitter(source, target, env):
"""Doxygen Doxyfile emitter"""
# possible output formats and their default values and output locations
output_formats = {
"HTML": ("YES", "html"),
"LATEX": ("YES", "latex"),
"RTF": ("NO", "rtf"),
"MAN": ("YES", "man"),
"XML": ("NO", "xml"),
... | [
"def",
"DoxyEmitter",
"(",
"source",
",",
"target",
",",
"env",
")",
":",
"# possible output formats and their default values and output locations",
"output_formats",
"=",
"{",
"\"HTML\"",
":",
"(",
"\"YES\"",
",",
"\"html\"",
")",
",",
"\"LATEX\"",
":",
"(",
"\"YES... | Doxygen Doxyfile emitter | [
"Doxygen",
"Doxyfile",
"emitter"
] | 9f7e7705793d258a0b205f185b20e3bbcda473da | https://github.com/melizalab/libtfr/blob/9f7e7705793d258a0b205f185b20e3bbcda473da/site_scons/site_tools/doxygen.py#L181-L221 | train |
melizalab/libtfr | site_scons/site_tools/doxygen.py | generate | def generate(env):
"""
Add builders and construction variables for the
Doxygen tool. This is currently for Doxygen 1.4.6.
"""
doxyfile_scanner = env.Scanner(
DoxySourceScan,
"DoxySourceScan",
scan_check = DoxySourceScanCheck,
)
import SCons.Builder
doxyfile_builder = SCons.Bu... | python | def generate(env):
"""
Add builders and construction variables for the
Doxygen tool. This is currently for Doxygen 1.4.6.
"""
doxyfile_scanner = env.Scanner(
DoxySourceScan,
"DoxySourceScan",
scan_check = DoxySourceScanCheck,
)
import SCons.Builder
doxyfile_builder = SCons.Bu... | [
"def",
"generate",
"(",
"env",
")",
":",
"doxyfile_scanner",
"=",
"env",
".",
"Scanner",
"(",
"DoxySourceScan",
",",
"\"DoxySourceScan\"",
",",
"scan_check",
"=",
"DoxySourceScanCheck",
",",
")",
"import",
"SCons",
".",
"Builder",
"doxyfile_builder",
"=",
"SCons... | Add builders and construction variables for the
Doxygen tool. This is currently for Doxygen 1.4.6. | [
"Add",
"builders",
"and",
"construction",
"variables",
"for",
"the",
"Doxygen",
"tool",
".",
"This",
"is",
"currently",
"for",
"Doxygen",
"1",
".",
"4",
".",
"6",
"."
] | 9f7e7705793d258a0b205f185b20e3bbcda473da | https://github.com/melizalab/libtfr/blob/9f7e7705793d258a0b205f185b20e3bbcda473da/site_scons/site_tools/doxygen.py#L223-L249 | train |
finklabs/metrics | metrics/position.py | PosMetric.reset | def reset(self):
"""Reset metric counter."""
self._positions = []
self._line = 1
self._curr = None # current scope we are analyzing
self._scope = 0
self.language = None | python | def reset(self):
"""Reset metric counter."""
self._positions = []
self._line = 1
self._curr = None # current scope we are analyzing
self._scope = 0
self.language = None | [
"def",
"reset",
"(",
"self",
")",
":",
"self",
".",
"_positions",
"=",
"[",
"]",
"self",
".",
"_line",
"=",
"1",
"self",
".",
"_curr",
"=",
"None",
"# current scope we are analyzing",
"self",
".",
"_scope",
"=",
"0",
"self",
".",
"language",
"=",
"None... | Reset metric counter. | [
"Reset",
"metric",
"counter",
"."
] | fd9974af498831664b9ae8e8f3834e1ec2e8a699 | https://github.com/finklabs/metrics/blob/fd9974af498831664b9ae8e8f3834e1ec2e8a699/metrics/position.py#L87-L93 | train |
finklabs/metrics | metrics/position.py | PosMetric.add_scope | def add_scope(self, scope_type, scope_name, scope_start, is_method=False):
"""we identified a scope and add it to positions."""
if self._curr is not None:
self._curr['end'] = scope_start - 1 # close last scope
self._curr = {
'type': scope_type, 'name': scope_name,
... | python | def add_scope(self, scope_type, scope_name, scope_start, is_method=False):
"""we identified a scope and add it to positions."""
if self._curr is not None:
self._curr['end'] = scope_start - 1 # close last scope
self._curr = {
'type': scope_type, 'name': scope_name,
... | [
"def",
"add_scope",
"(",
"self",
",",
"scope_type",
",",
"scope_name",
",",
"scope_start",
",",
"is_method",
"=",
"False",
")",
":",
"if",
"self",
".",
"_curr",
"is",
"not",
"None",
":",
"self",
".",
"_curr",
"[",
"'end'",
"]",
"=",
"scope_start",
"-",... | we identified a scope and add it to positions. | [
"we",
"identified",
"a",
"scope",
"and",
"add",
"it",
"to",
"positions",
"."
] | fd9974af498831664b9ae8e8f3834e1ec2e8a699 | https://github.com/finklabs/metrics/blob/fd9974af498831664b9ae8e8f3834e1ec2e8a699/metrics/position.py#L95-L110 | train |
finklabs/metrics | metrics/position.py | PosMetric.process_token | def process_token(self, tok):
"""count lines and track position of classes and functions"""
if tok[0] == Token.Text:
count = tok[1].count('\n')
if count:
self._line += count # adjust linecount
if self._detector.process(tok):
pass # works bee... | python | def process_token(self, tok):
"""count lines and track position of classes and functions"""
if tok[0] == Token.Text:
count = tok[1].count('\n')
if count:
self._line += count # adjust linecount
if self._detector.process(tok):
pass # works bee... | [
"def",
"process_token",
"(",
"self",
",",
"tok",
")",
":",
"if",
"tok",
"[",
"0",
"]",
"==",
"Token",
".",
"Text",
":",
"count",
"=",
"tok",
"[",
"1",
"]",
".",
"count",
"(",
"'\\n'",
")",
"if",
"count",
":",
"self",
".",
"_line",
"+=",
"count"... | count lines and track position of classes and functions | [
"count",
"lines",
"and",
"track",
"position",
"of",
"classes",
"and",
"functions"
] | fd9974af498831664b9ae8e8f3834e1ec2e8a699 | https://github.com/finklabs/metrics/blob/fd9974af498831664b9ae8e8f3834e1ec2e8a699/metrics/position.py#L112-L132 | train |
rwl/pylon | pylon/solver.py | _Solver._unpack_model | def _unpack_model(self, om):
""" Returns data from the OPF model.
"""
buses = om.case.connected_buses
branches = om.case.online_branches
gens = om.case.online_generators
cp = om.get_cost_params()
# Bf = om._Bf
# Pfinj = om._Pfinj
return buses, bra... | python | def _unpack_model(self, om):
""" Returns data from the OPF model.
"""
buses = om.case.connected_buses
branches = om.case.online_branches
gens = om.case.online_generators
cp = om.get_cost_params()
# Bf = om._Bf
# Pfinj = om._Pfinj
return buses, bra... | [
"def",
"_unpack_model",
"(",
"self",
",",
"om",
")",
":",
"buses",
"=",
"om",
".",
"case",
".",
"connected_buses",
"branches",
"=",
"om",
".",
"case",
".",
"online_branches",
"gens",
"=",
"om",
".",
"case",
".",
"online_generators",
"cp",
"=",
"om",
".... | Returns data from the OPF model. | [
"Returns",
"data",
"from",
"the",
"OPF",
"model",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/solver.py#L78-L90 | train |
rwl/pylon | pylon/solver.py | _Solver._dimension_data | def _dimension_data(self, buses, branches, generators):
""" Returns the problem dimensions.
"""
ipol = [i for i, g in enumerate(generators)
if g.pcost_model == POLYNOMIAL]
ipwl = [i for i, g in enumerate(generators)
if g.pcost_model == PW_LINEAR]
n... | python | def _dimension_data(self, buses, branches, generators):
""" Returns the problem dimensions.
"""
ipol = [i for i, g in enumerate(generators)
if g.pcost_model == POLYNOMIAL]
ipwl = [i for i, g in enumerate(generators)
if g.pcost_model == PW_LINEAR]
n... | [
"def",
"_dimension_data",
"(",
"self",
",",
"buses",
",",
"branches",
",",
"generators",
")",
":",
"ipol",
"=",
"[",
"i",
"for",
"i",
",",
"g",
"in",
"enumerate",
"(",
"generators",
")",
"if",
"g",
".",
"pcost_model",
"==",
"POLYNOMIAL",
"]",
"ipwl",
... | Returns the problem dimensions. | [
"Returns",
"the",
"problem",
"dimensions",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/solver.py#L93-L112 | train |
rwl/pylon | pylon/solver.py | _Solver._linear_constraints | def _linear_constraints(self, om):
""" Returns the linear problem constraints.
"""
A, l, u = om.linear_constraints() # l <= A*x <= u
# Indexes for equality, greater than (unbounded above), less than
# (unbounded below) and doubly-bounded box constraints.
# ieq = flatnonze... | python | def _linear_constraints(self, om):
""" Returns the linear problem constraints.
"""
A, l, u = om.linear_constraints() # l <= A*x <= u
# Indexes for equality, greater than (unbounded above), less than
# (unbounded below) and doubly-bounded box constraints.
# ieq = flatnonze... | [
"def",
"_linear_constraints",
"(",
"self",
",",
"om",
")",
":",
"A",
",",
"l",
",",
"u",
"=",
"om",
".",
"linear_constraints",
"(",
")",
"# l <= A*x <= u",
"# Indexes for equality, greater than (unbounded above), less than",
"# (unbounded below) and doubly-bounded box const... | Returns the linear problem constraints. | [
"Returns",
"the",
"linear",
"problem",
"constraints",
"."
] | 916514255db1ae1661406f0283df756baf960d14 | https://github.com/rwl/pylon/blob/916514255db1ae1661406f0283df756baf960d14/pylon/solver.py#L115-L149 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.