nwo
stringlengths 5
86
| sha
stringlengths 40
40
| path
stringlengths 4
189
| language
stringclasses 1
value | identifier
stringlengths 1
94
| parameters
stringlengths 2
4.03k
| argument_list
stringclasses 1
value | return_statement
stringlengths 0
11.5k
| docstring
stringlengths 1
33.2k
| docstring_summary
stringlengths 0
5.15k
| docstring_tokens
sequence | function
stringlengths 34
151k
| function_tokens
sequence | url
stringlengths 90
278
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
jeog/TDAmeritradeAPI | 91c738afd7d57b54f6231170bd64c2550fafd34d | python/tdma_api/auth.py | python | set_certificate_bundle_path | (path) | Set certificate bundle file(.pem) path for ssl/tls host authentication.
If library is built against default ssl/tls library the default certificate
store should be used. If not(a connection error is returned) you'll have to
provide a certificate bundle to the connection libraries.
def set_certificate_bundle_path(path);
path :: str :: path to the certificate bundle file(.pem)
returns -> None
throws -> LibraryNotLoaded, CLibException | Set certificate bundle file(.pem) path for ssl/tls host authentication.
If library is built against default ssl/tls library the default certificate
store should be used. If not(a connection error is returned) you'll have to
provide a certificate bundle to the connection libraries.
def set_certificate_bundle_path(path);
path :: str :: path to the certificate bundle file(.pem)
returns -> None
throws -> LibraryNotLoaded, CLibException | [
"Set",
"certificate",
"bundle",
"file",
"(",
".",
"pem",
")",
"path",
"for",
"ssl",
"/",
"tls",
"host",
"authentication",
".",
"If",
"library",
"is",
"built",
"against",
"default",
"ssl",
"/",
"tls",
"library",
"the",
"default",
"certificate",
"store",
"should",
"be",
"used",
".",
"If",
"not",
"(",
"a",
"connection",
"error",
"is",
"returned",
")",
"you",
"ll",
"have",
"to",
"provide",
"a",
"certificate",
"bundle",
"to",
"the",
"connection",
"libraries",
".",
"def",
"set_certificate_bundle_path",
"(",
"path",
")",
";",
"path",
"::",
"str",
"::",
"path",
"to",
"the",
"certificate",
"bundle",
"file",
"(",
".",
"pem",
")",
"returns",
"-",
">",
"None",
"throws",
"-",
">",
"LibraryNotLoaded",
"CLibException"
] | def set_certificate_bundle_path(path):
"""Set certificate bundle file(.pem) path for ssl/tls host authentication.
If library is built against default ssl/tls library the default certificate
store should be used. If not(a connection error is returned) you'll have to
provide a certificate bundle to the connection libraries.
def set_certificate_bundle_path(path);
path :: str :: path to the certificate bundle file(.pem)
returns -> None
throws -> LibraryNotLoaded, CLibException
"""
clib.set_str('SetCertificateBundlePath_ABI', path) | [
"def",
"set_certificate_bundle_path",
"(",
"path",
")",
":",
"clib",
".",
"set_str",
"(",
"'SetCertificateBundlePath_ABI'",
",",
"path",
")"
] | https://github.com/jeog/TDAmeritradeAPI/blob/91c738afd7d57b54f6231170bd64c2550fafd34d/python/tdma_api/auth.py#L145-L160 |
||
su2code/SU2 | 72b2fa977b64b9683a388920f05298a40d39e5c5 | SU2_PY/SU2/opt/scipy_tools.py | python | obj_df | (x,project) | return dobj | dobj = obj_df(x,project)
Objective Function Gradients
SU2 Project interface to scipy.fmin_slsqp
su2: df(x), list[nobj x dim]
scipy_slsqp: df(x), ndarray[dim] | dobj = obj_df(x,project)
Objective Function Gradients
SU2 Project interface to scipy.fmin_slsqp
su2: df(x), list[nobj x dim]
scipy_slsqp: df(x), ndarray[dim] | [
"dobj",
"=",
"obj_df",
"(",
"x",
"project",
")",
"Objective",
"Function",
"Gradients",
"SU2",
"Project",
"interface",
"to",
"scipy",
".",
"fmin_slsqp",
"su2",
":",
"df",
"(",
"x",
")",
"list",
"[",
"nobj",
"x",
"dim",
"]",
"scipy_slsqp",
":",
"df",
"(",
"x",
")",
"ndarray",
"[",
"dim",
"]"
] | def obj_df(x,project):
""" dobj = obj_df(x,project)
Objective Function Gradients
SU2 Project interface to scipy.fmin_slsqp
su2: df(x), list[nobj x dim]
scipy_slsqp: df(x), ndarray[dim]
"""
dobj_list = project.obj_df(x)
dobj=[0.0]*len(dobj_list[0])
for this_dobj in dobj_list:
idv=0
for this_dv_dobj in this_dobj:
dobj[idv] = dobj[idv]+this_dv_dobj;
idv+=1
dobj = array( dobj )
return dobj | [
"def",
"obj_df",
"(",
"x",
",",
"project",
")",
":",
"dobj_list",
"=",
"project",
".",
"obj_df",
"(",
"x",
")",
"dobj",
"=",
"[",
"0.0",
"]",
"*",
"len",
"(",
"dobj_list",
"[",
"0",
"]",
")",
"for",
"this_dobj",
"in",
"dobj_list",
":",
"idv",
"=",
"0",
"for",
"this_dv_dobj",
"in",
"this_dobj",
":",
"dobj",
"[",
"idv",
"]",
"=",
"dobj",
"[",
"idv",
"]",
"+",
"this_dv_dobj",
"idv",
"+=",
"1",
"dobj",
"=",
"array",
"(",
"dobj",
")",
"return",
"dobj"
] | https://github.com/su2code/SU2/blob/72b2fa977b64b9683a388920f05298a40d39e5c5/SU2_PY/SU2/opt/scipy_tools.py#L390-L410 |
|
hanpfei/chromium-net | 392cc1fa3a8f92f42e4071ab6e674d8e0482f83f | third_party/catapult/third_party/apiclient/googleapiclient/model.py | python | BaseModel.deserialize | (self, content) | Perform the actual deserialization from response string to Python
object.
Args:
content: string, the body of the HTTP response
Returns:
The body de-serialized as a Python object. | Perform the actual deserialization from response string to Python
object. | [
"Perform",
"the",
"actual",
"deserialization",
"from",
"response",
"string",
"to",
"Python",
"object",
"."
] | def deserialize(self, content):
"""Perform the actual deserialization from response string to Python
object.
Args:
content: string, the body of the HTTP response
Returns:
The body de-serialized as a Python object.
"""
_abstract() | [
"def",
"deserialize",
"(",
"self",
",",
"content",
")",
":",
"_abstract",
"(",
")"
] | https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/apiclient/googleapiclient/model.py#L223-L233 |
||
nucleic/atom | 9f0cb2a8101dd63c354a98ebc7489b2c616dc82a | examples/tutorial/person.py | python | Person.debug_print | (self, change) | Prints out a debug message whenever the person's age changes. | Prints out a debug message whenever the person's age changes. | [
"Prints",
"out",
"a",
"debug",
"message",
"whenever",
"the",
"person",
"s",
"age",
"changes",
"."
] | def debug_print(self, change):
"""Prints out a debug message whenever the person's age changes."""
if self.debug:
templ = "{first} {last} is {age} years old."
s = templ.format(
first=self.first_name,
last=self.last_name,
age=self.age,
)
print(s) | [
"def",
"debug_print",
"(",
"self",
",",
"change",
")",
":",
"if",
"self",
".",
"debug",
":",
"templ",
"=",
"\"{first} {last} is {age} years old.\"",
"s",
"=",
"templ",
".",
"format",
"(",
"first",
"=",
"self",
".",
"first_name",
",",
"last",
"=",
"self",
".",
"last_name",
",",
"age",
"=",
"self",
".",
"age",
",",
")",
"print",
"(",
"s",
")"
] | https://github.com/nucleic/atom/blob/9f0cb2a8101dd63c354a98ebc7489b2c616dc82a/examples/tutorial/person.py#L26-L35 |
||
catboost/catboost | 167f64f237114a4d10b2b4ee42adb4569137debe | contrib/python/jedi/jedi/evaluate/arguments.py | python | try_iter_content | (types, depth=0) | Helper method for static analysis. | Helper method for static analysis. | [
"Helper",
"method",
"for",
"static",
"analysis",
"."
] | def try_iter_content(types, depth=0):
"""Helper method for static analysis."""
if depth > 10:
# It's possible that a loop has references on itself (especially with
# CompiledObject). Therefore don't loop infinitely.
return
for typ in types:
try:
f = typ.py__iter__
except AttributeError:
pass
else:
for lazy_context in f():
try_iter_content(lazy_context.infer(), depth + 1) | [
"def",
"try_iter_content",
"(",
"types",
",",
"depth",
"=",
"0",
")",
":",
"if",
"depth",
">",
"10",
":",
"# It's possible that a loop has references on itself (especially with",
"# CompiledObject). Therefore don't loop infinitely.",
"return",
"for",
"typ",
"in",
"types",
":",
"try",
":",
"f",
"=",
"typ",
".",
"py__iter__",
"except",
"AttributeError",
":",
"pass",
"else",
":",
"for",
"lazy_context",
"in",
"f",
"(",
")",
":",
"try_iter_content",
"(",
"lazy_context",
".",
"infer",
"(",
")",
",",
"depth",
"+",
"1",
")"
] | https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/jedi/jedi/evaluate/arguments.py#L16-L30 |
||
wxWidgets/wxPython-Classic | 19571e1ae65f1ac445f5491474121998c97a1bf0 | src/osx_carbon/_misc.py | python | DateSpan.SetWeeks | (*args, **kwargs) | return _misc_.DateSpan_SetWeeks(*args, **kwargs) | SetWeeks(self, int n) -> DateSpan | SetWeeks(self, int n) -> DateSpan | [
"SetWeeks",
"(",
"self",
"int",
"n",
")",
"-",
">",
"DateSpan"
] | def SetWeeks(*args, **kwargs):
"""SetWeeks(self, int n) -> DateSpan"""
return _misc_.DateSpan_SetWeeks(*args, **kwargs) | [
"def",
"SetWeeks",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"_misc_",
".",
"DateSpan_SetWeeks",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_carbon/_misc.py#L4661-L4663 |
|
hanpfei/chromium-net | 392cc1fa3a8f92f42e4071ab6e674d8e0482f83f | third_party/catapult/third_party/gsutil/third_party/boto/boto/swf/layer1.py | python | Layer1.describe_activity_type | (self, domain, activity_name, activity_version) | return self.json_request('DescribeActivityType', {
'domain': domain,
'activityType': {'name': activity_name,
'version': activity_version}
}) | Returns information about the specified activity type. This
includes configuration settings provided at registration time
as well as other general information about the type.
:type domain: string
:param domain: The name of the domain in which the activity
type is registered.
:type activity_name: string
:param activity_name: The name of this activity.
:type activity_version: string
:param activity_version: The version of this activity.
:raises: UnknownResourceFault, SWFOperationNotPermittedError | Returns information about the specified activity type. This
includes configuration settings provided at registration time
as well as other general information about the type. | [
"Returns",
"information",
"about",
"the",
"specified",
"activity",
"type",
".",
"This",
"includes",
"configuration",
"settings",
"provided",
"at",
"registration",
"time",
"as",
"well",
"as",
"other",
"general",
"information",
"about",
"the",
"type",
"."
] | def describe_activity_type(self, domain, activity_name, activity_version):
"""
Returns information about the specified activity type. This
includes configuration settings provided at registration time
as well as other general information about the type.
:type domain: string
:param domain: The name of the domain in which the activity
type is registered.
:type activity_name: string
:param activity_name: The name of this activity.
:type activity_version: string
:param activity_version: The version of this activity.
:raises: UnknownResourceFault, SWFOperationNotPermittedError
"""
return self.json_request('DescribeActivityType', {
'domain': domain,
'activityType': {'name': activity_name,
'version': activity_version}
}) | [
"def",
"describe_activity_type",
"(",
"self",
",",
"domain",
",",
"activity_name",
",",
"activity_version",
")",
":",
"return",
"self",
".",
"json_request",
"(",
"'DescribeActivityType'",
",",
"{",
"'domain'",
":",
"domain",
",",
"'activityType'",
":",
"{",
"'name'",
":",
"activity_name",
",",
"'version'",
":",
"activity_version",
"}",
"}",
")"
] | https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/gsutil/third_party/boto/boto/swf/layer1.py#L909-L931 |
|
wxWidgets/wxPython-Classic | 19571e1ae65f1ac445f5491474121998c97a1bf0 | src/osx_cocoa/wizard.py | python | PyWizardPage.DoGetSize | (*args, **kwargs) | return _wizard.PyWizardPage_DoGetSize(*args, **kwargs) | DoGetSize() -> (width, height) | DoGetSize() -> (width, height) | [
"DoGetSize",
"()",
"-",
">",
"(",
"width",
"height",
")"
] | def DoGetSize(*args, **kwargs):
"""DoGetSize() -> (width, height)"""
return _wizard.PyWizardPage_DoGetSize(*args, **kwargs) | [
"def",
"DoGetSize",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"_wizard",
".",
"PyWizardPage_DoGetSize",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_cocoa/wizard.py#L167-L169 |
|
hanpfei/chromium-net | 392cc1fa3a8f92f42e4071ab6e674d8e0482f83f | third_party/catapult/third_party/closure_linter/closure_linter/statetracker.py | python | DocComment.GetTargetToken | (self) | Get this comment's target token.
Returns:
The token that is the target of this comment, or None if there isn't one. | Get this comment's target token. | [
"Get",
"this",
"comment",
"s",
"target",
"token",
"."
] | def GetTargetToken(self):
"""Get this comment's target token.
Returns:
The token that is the target of this comment, or None if there isn't one.
"""
# File overviews describe the file, not a token.
if self.HasFlag('fileoverview'):
return
skip_types = frozenset([
Type.WHITESPACE,
Type.BLANK_LINE,
Type.START_PAREN])
target_types = frozenset([
Type.FUNCTION_NAME,
Type.IDENTIFIER,
Type.SIMPLE_LVALUE])
token = self.end_token.next
while token:
if token.type in target_types:
return token
# Handles the case of a comment on "var foo = ...'
if token.IsKeyword('var'):
next_code_token = tokenutil.CustomSearch(
token,
lambda t: t.type not in Type.NON_CODE_TYPES)
if (next_code_token and
next_code_token.IsType(Type.SIMPLE_LVALUE)):
return next_code_token
return
# Handles the case of a comment on "function foo () {}"
if token.type is Type.FUNCTION_DECLARATION:
next_code_token = tokenutil.CustomSearch(
token,
lambda t: t.type not in Type.NON_CODE_TYPES)
if next_code_token.IsType(Type.FUNCTION_NAME):
return next_code_token
return
# Skip types will end the search.
if token.type not in skip_types:
return
token = token.next | [
"def",
"GetTargetToken",
"(",
"self",
")",
":",
"# File overviews describe the file, not a token.",
"if",
"self",
".",
"HasFlag",
"(",
"'fileoverview'",
")",
":",
"return",
"skip_types",
"=",
"frozenset",
"(",
"[",
"Type",
".",
"WHITESPACE",
",",
"Type",
".",
"BLANK_LINE",
",",
"Type",
".",
"START_PAREN",
"]",
")",
"target_types",
"=",
"frozenset",
"(",
"[",
"Type",
".",
"FUNCTION_NAME",
",",
"Type",
".",
"IDENTIFIER",
",",
"Type",
".",
"SIMPLE_LVALUE",
"]",
")",
"token",
"=",
"self",
".",
"end_token",
".",
"next",
"while",
"token",
":",
"if",
"token",
".",
"type",
"in",
"target_types",
":",
"return",
"token",
"# Handles the case of a comment on \"var foo = ...'",
"if",
"token",
".",
"IsKeyword",
"(",
"'var'",
")",
":",
"next_code_token",
"=",
"tokenutil",
".",
"CustomSearch",
"(",
"token",
",",
"lambda",
"t",
":",
"t",
".",
"type",
"not",
"in",
"Type",
".",
"NON_CODE_TYPES",
")",
"if",
"(",
"next_code_token",
"and",
"next_code_token",
".",
"IsType",
"(",
"Type",
".",
"SIMPLE_LVALUE",
")",
")",
":",
"return",
"next_code_token",
"return",
"# Handles the case of a comment on \"function foo () {}\"",
"if",
"token",
".",
"type",
"is",
"Type",
".",
"FUNCTION_DECLARATION",
":",
"next_code_token",
"=",
"tokenutil",
".",
"CustomSearch",
"(",
"token",
",",
"lambda",
"t",
":",
"t",
".",
"type",
"not",
"in",
"Type",
".",
"NON_CODE_TYPES",
")",
"if",
"next_code_token",
".",
"IsType",
"(",
"Type",
".",
"FUNCTION_NAME",
")",
":",
"return",
"next_code_token",
"return",
"# Skip types will end the search.",
"if",
"token",
".",
"type",
"not",
"in",
"skip_types",
":",
"return",
"token",
"=",
"token",
".",
"next"
] | https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/closure_linter/closure_linter/statetracker.py#L452-L505 |
||
nlohmann/json | eb2182414749825be086c825edb5229e5c28503d | third_party/cpplint/cpplint.py | python | IsBlankLine | (line) | return not line or line.isspace() | Returns true if the given line is blank.
We consider a line to be blank if the line is empty or consists of
only white spaces.
Args:
line: A line of a string.
Returns:
True, if the given line is blank. | Returns true if the given line is blank. | [
"Returns",
"true",
"if",
"the",
"given",
"line",
"is",
"blank",
"."
] | def IsBlankLine(line):
"""Returns true if the given line is blank.
We consider a line to be blank if the line is empty or consists of
only white spaces.
Args:
line: A line of a string.
Returns:
True, if the given line is blank.
"""
return not line or line.isspace() | [
"def",
"IsBlankLine",
"(",
"line",
")",
":",
"return",
"not",
"line",
"or",
"line",
".",
"isspace",
"(",
")"
] | https://github.com/nlohmann/json/blob/eb2182414749825be086c825edb5229e5c28503d/third_party/cpplint/cpplint.py#L3513-L3525 |
|
ricardoquesada/Spidermonkey | 4a75ea2543408bd1b2c515aa95901523eeef7858 | dom/bindings/parser/WebIDL.py | python | Parser.p_Ellipsis | (self, p) | Ellipsis : ELLIPSIS | Ellipsis : ELLIPSIS | [
"Ellipsis",
":",
"ELLIPSIS"
] | def p_Ellipsis(self, p):
"""
Ellipsis : ELLIPSIS
"""
p[0] = True | [
"def",
"p_Ellipsis",
"(",
"self",
",",
"p",
")",
":",
"p",
"[",
"0",
"]",
"=",
"True"
] | https://github.com/ricardoquesada/Spidermonkey/blob/4a75ea2543408bd1b2c515aa95901523eeef7858/dom/bindings/parser/WebIDL.py#L5016-L5020 |
||
aws/lumberyard | f85344403c1c2e77ec8c75deb2c116e97b713217 | dev/Gems/CloudGemDefectReporter/v1/AWS/common-code/Lib/requests_toolbelt/multipart/encoder.py | python | Part.from_field | (cls, field, encoding) | return cls(headers, body) | Create a part from a Request Field generated by urllib3. | Create a part from a Request Field generated by urllib3. | [
"Create",
"a",
"part",
"from",
"a",
"Request",
"Field",
"generated",
"by",
"urllib3",
"."
] | def from_field(cls, field, encoding):
"""Create a part from a Request Field generated by urllib3."""
headers = encode_with(field.render_headers(), encoding)
body = coerce_data(field.data, encoding)
return cls(headers, body) | [
"def",
"from_field",
"(",
"cls",
",",
"field",
",",
"encoding",
")",
":",
"headers",
"=",
"encode_with",
"(",
"field",
".",
"render_headers",
"(",
")",
",",
"encoding",
")",
"body",
"=",
"coerce_data",
"(",
"field",
".",
"data",
",",
"encoding",
")",
"return",
"cls",
"(",
"headers",
",",
"body",
")"
] | https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Gems/CloudGemDefectReporter/v1/AWS/common-code/Lib/requests_toolbelt/multipart/encoder.py#L485-L489 |
|
Illumina/strelka | d7377443b62319f7c7bd70c241c4b2df3459e29a | src/python/lib/configureUtil.py | python | validateFixExistingDirArg | (argDir,label) | return _validateFixArgHelper(argDir,label,os.path.isdir) | convert directory arg to absolute path and check that it exists | convert directory arg to absolute path and check that it exists | [
"convert",
"directory",
"arg",
"to",
"absolute",
"path",
"and",
"check",
"that",
"it",
"exists"
] | def validateFixExistingDirArg(argDir,label) :
"""
convert directory arg to absolute path and check that it exists
"""
return _validateFixArgHelper(argDir,label,os.path.isdir) | [
"def",
"validateFixExistingDirArg",
"(",
"argDir",
",",
"label",
")",
":",
"return",
"_validateFixArgHelper",
"(",
"argDir",
",",
"label",
",",
"os",
".",
"path",
".",
"isdir",
")"
] | https://github.com/Illumina/strelka/blob/d7377443b62319f7c7bd70c241c4b2df3459e29a/src/python/lib/configureUtil.py#L186-L190 |
|
ApolloAuto/apollo | 463fb82f9e979d02dcb25044e60931293ab2dba0 | scripts/record_map_data.py | python | ArgManager.args | (self) | return self._args | Get parsed args. | Get parsed args. | [
"Get",
"parsed",
"args",
"."
] | def args(self):
"""Get parsed args."""
if self._args is None:
self._args = self.parser.parse_args()
return self._args | [
"def",
"args",
"(",
"self",
")",
":",
"if",
"self",
".",
"_args",
"is",
"None",
":",
"self",
".",
"_args",
"=",
"self",
".",
"parser",
".",
"parse_args",
"(",
")",
"return",
"self",
".",
"_args"
] | https://github.com/ApolloAuto/apollo/blob/463fb82f9e979d02dcb25044e60931293ab2dba0/scripts/record_map_data.py#L94-L98 |
|
OSGeo/gdal | 3748fc4ba4fba727492774b2b908a2130c864a83 | swig/python/osgeo/ogr.py | python | MajorObject.SetMetadata | (self, *args) | return _ogr.MajorObject_SetMetadata(self, *args) | r"""
SetMetadata(MajorObject self, char ** papszMetadata, char const * pszDomain="") -> CPLErr
SetMetadata(MajorObject self, char * pszMetadataString, char const * pszDomain="") -> CPLErr | r"""
SetMetadata(MajorObject self, char ** papszMetadata, char const * pszDomain="") -> CPLErr
SetMetadata(MajorObject self, char * pszMetadataString, char const * pszDomain="") -> CPLErr | [
"r",
"SetMetadata",
"(",
"MajorObject",
"self",
"char",
"**",
"papszMetadata",
"char",
"const",
"*",
"pszDomain",
"=",
")",
"-",
">",
"CPLErr",
"SetMetadata",
"(",
"MajorObject",
"self",
"char",
"*",
"pszMetadataString",
"char",
"const",
"*",
"pszDomain",
"=",
")",
"-",
">",
"CPLErr"
] | def SetMetadata(self, *args):
r"""
SetMetadata(MajorObject self, char ** papszMetadata, char const * pszDomain="") -> CPLErr
SetMetadata(MajorObject self, char * pszMetadataString, char const * pszDomain="") -> CPLErr
"""
return _ogr.MajorObject_SetMetadata(self, *args) | [
"def",
"SetMetadata",
"(",
"self",
",",
"*",
"args",
")",
":",
"return",
"_ogr",
".",
"MajorObject_SetMetadata",
"(",
"self",
",",
"*",
"args",
")"
] | https://github.com/OSGeo/gdal/blob/3748fc4ba4fba727492774b2b908a2130c864a83/swig/python/osgeo/ogr.py#L436-L441 |
|
panda3d/panda3d | 833ad89ebad58395d0af0b7ec08538e5e4308265 | samples/networking/03-distributed-node/AIDGameObject.py | python | AIDGameObject.announceGenerate | (self) | The AI has created this object, so we send it's distributed object ID
over to the client. That way the client can actually grab the object
and use it to communicate with the AI. Alternatively store it in the
Client Repository in self.cr | The AI has created this object, so we send it's distributed object ID
over to the client. That way the client can actually grab the object
and use it to communicate with the AI. Alternatively store it in the
Client Repository in self.cr | [
"The",
"AI",
"has",
"created",
"this",
"object",
"so",
"we",
"send",
"it",
"s",
"distributed",
"object",
"ID",
"over",
"to",
"the",
"client",
".",
"That",
"way",
"the",
"client",
"can",
"actually",
"grab",
"the",
"object",
"and",
"use",
"it",
"to",
"communicate",
"with",
"the",
"AI",
".",
"Alternatively",
"store",
"it",
"in",
"the",
"Client",
"Repository",
"in",
"self",
".",
"cr"
] | def announceGenerate(self):
""" The AI has created this object, so we send it's distributed object ID
over to the client. That way the client can actually grab the object
and use it to communicate with the AI. Alternatively store it in the
Client Repository in self.cr """
base.messenger.send(self.cr.uniqueName('AIDGameObjectGenerated'), [self.doId])
# call the base class method
DistributedObject.announceGenerate(self) | [
"def",
"announceGenerate",
"(",
"self",
")",
":",
"base",
".",
"messenger",
".",
"send",
"(",
"self",
".",
"cr",
".",
"uniqueName",
"(",
"'AIDGameObjectGenerated'",
")",
",",
"[",
"self",
".",
"doId",
"]",
")",
"# call the base class method",
"DistributedObject",
".",
"announceGenerate",
"(",
"self",
")"
] | https://github.com/panda3d/panda3d/blob/833ad89ebad58395d0af0b7ec08538e5e4308265/samples/networking/03-distributed-node/AIDGameObject.py#L10-L17 |
||
ChromiumWebApps/chromium | c7361d39be8abd1574e6ce8957c8dbddd4c6ccf7 | tools/clang/scripts/run_tool.py | python | _CompilerDispatcher.__ProcessResult | (self, result) | Handles result processing.
Args:
result: The result dictionary returned by _ExecuteTool. | Handles result processing. | [
"Handles",
"result",
"processing",
"."
] | def __ProcessResult(self, result):
"""Handles result processing.
Args:
result: The result dictionary returned by _ExecuteTool.
"""
if result['status']:
self.__success_count += 1
for k, v in result['edits'].iteritems():
self.__edits[k].extend(v)
else:
self.__failed_count += 1
sys.stdout.write('\nFailed to process %s\n' % result['filename'])
sys.stdout.write(result['stderr'])
sys.stdout.write('\n')
percentage = (
float(self.__success_count + self.__failed_count) /
len(self.__filenames)) * 100
sys.stdout.write('Succeeded: %d, Failed: %d [%.2f%%]\r' % (
self.__success_count, self.__failed_count, percentage))
sys.stdout.flush() | [
"def",
"__ProcessResult",
"(",
"self",
",",
"result",
")",
":",
"if",
"result",
"[",
"'status'",
"]",
":",
"self",
".",
"__success_count",
"+=",
"1",
"for",
"k",
",",
"v",
"in",
"result",
"[",
"'edits'",
"]",
".",
"iteritems",
"(",
")",
":",
"self",
".",
"__edits",
"[",
"k",
"]",
".",
"extend",
"(",
"v",
")",
"else",
":",
"self",
".",
"__failed_count",
"+=",
"1",
"sys",
".",
"stdout",
".",
"write",
"(",
"'\\nFailed to process %s\\n'",
"%",
"result",
"[",
"'filename'",
"]",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"result",
"[",
"'stderr'",
"]",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"'\\n'",
")",
"percentage",
"=",
"(",
"float",
"(",
"self",
".",
"__success_count",
"+",
"self",
".",
"__failed_count",
")",
"/",
"len",
"(",
"self",
".",
"__filenames",
")",
")",
"*",
"100",
"sys",
".",
"stdout",
".",
"write",
"(",
"'Succeeded: %d, Failed: %d [%.2f%%]\\r'",
"%",
"(",
"self",
".",
"__success_count",
",",
"self",
".",
"__failed_count",
",",
"percentage",
")",
")",
"sys",
".",
"stdout",
".",
"flush",
"(",
")"
] | https://github.com/ChromiumWebApps/chromium/blob/c7361d39be8abd1574e6ce8957c8dbddd4c6ccf7/tools/clang/scripts/run_tool.py#L163-L183 |
||
BlzFans/wke | b0fa21158312e40c5fbd84682d643022b6c34a93 | cygwin/lib/python2.6/distutils/command/build_py.py | python | build_py.find_data_files | (self, package, src_dir) | return files | Return filenames for package's data files in 'src_dir | Return filenames for package's data files in 'src_dir | [
"Return",
"filenames",
"for",
"package",
"s",
"data",
"files",
"in",
"src_dir"
] | def find_data_files (self, package, src_dir):
"""Return filenames for package's data files in 'src_dir'"""
globs = (self.package_data.get('', [])
+ self.package_data.get(package, []))
files = []
for pattern in globs:
# Each pattern has to be converted to a platform-specific path
filelist = glob(os.path.join(src_dir, convert_path(pattern)))
# Files that match more than one pattern are only added once
files.extend([fn for fn in filelist if fn not in files])
return files | [
"def",
"find_data_files",
"(",
"self",
",",
"package",
",",
"src_dir",
")",
":",
"globs",
"=",
"(",
"self",
".",
"package_data",
".",
"get",
"(",
"''",
",",
"[",
"]",
")",
"+",
"self",
".",
"package_data",
".",
"get",
"(",
"package",
",",
"[",
"]",
")",
")",
"files",
"=",
"[",
"]",
"for",
"pattern",
"in",
"globs",
":",
"# Each pattern has to be converted to a platform-specific path",
"filelist",
"=",
"glob",
"(",
"os",
".",
"path",
".",
"join",
"(",
"src_dir",
",",
"convert_path",
"(",
"pattern",
")",
")",
")",
"# Files that match more than one pattern are only added once",
"files",
".",
"extend",
"(",
"[",
"fn",
"for",
"fn",
"in",
"filelist",
"if",
"fn",
"not",
"in",
"files",
"]",
")",
"return",
"files"
] | https://github.com/BlzFans/wke/blob/b0fa21158312e40c5fbd84682d643022b6c34a93/cygwin/lib/python2.6/distutils/command/build_py.py#L129-L139 |
|
kichik/nsis | e39fe70400b823ac3d00321e338cf3410634b10a | SCons/Tools/mstoolkit.py | python | pch_emitter | (target, source, env) | return (target, source) | Sets up the PDB dependencies for a pch file, and adds the object
file target. | Sets up the PDB dependencies for a pch file, and adds the object
file target. | [
"Sets",
"up",
"the",
"PDB",
"dependencies",
"for",
"a",
"pch",
"file",
"and",
"adds",
"the",
"object",
"file",
"target",
"."
] | def pch_emitter(target, source, env):
"""Sets up the PDB dependencies for a pch file, and adds the object
file target."""
validate_vars(env)
pch = None
obj = None
for t in target:
if SCons.Util.splitext(str(t))[1] == '.pch':
pch = t
if SCons.Util.splitext(str(t))[1] == '.obj':
obj = t
if not obj:
obj = SCons.Util.splitext(str(pch))[0]+'.obj'
target = [pch, obj] # pch must be first, and obj second for the PCHCOM to work
if 'PDB' in env and env['PDB']:
env.SideEffect(env['PDB'], target)
env.Precious(env['PDB'])
return (target, source) | [
"def",
"pch_emitter",
"(",
"target",
",",
"source",
",",
"env",
")",
":",
"validate_vars",
"(",
"env",
")",
"pch",
"=",
"None",
"obj",
"=",
"None",
"for",
"t",
"in",
"target",
":",
"if",
"SCons",
".",
"Util",
".",
"splitext",
"(",
"str",
"(",
"t",
")",
")",
"[",
"1",
"]",
"==",
"'.pch'",
":",
"pch",
"=",
"t",
"if",
"SCons",
".",
"Util",
".",
"splitext",
"(",
"str",
"(",
"t",
")",
")",
"[",
"1",
"]",
"==",
"'.obj'",
":",
"obj",
"=",
"t",
"if",
"not",
"obj",
":",
"obj",
"=",
"SCons",
".",
"Util",
".",
"splitext",
"(",
"str",
"(",
"pch",
")",
")",
"[",
"0",
"]",
"+",
"'.obj'",
"target",
"=",
"[",
"pch",
",",
"obj",
"]",
"# pch must be first, and obj second for the PCHCOM to work",
"if",
"'PDB'",
"in",
"env",
"and",
"env",
"[",
"'PDB'",
"]",
":",
"env",
".",
"SideEffect",
"(",
"env",
"[",
"'PDB'",
"]",
",",
"target",
")",
"env",
".",
"Precious",
"(",
"env",
"[",
"'PDB'",
"]",
")",
"return",
"(",
"target",
",",
"source",
")"
] | https://github.com/kichik/nsis/blob/e39fe70400b823ac3d00321e338cf3410634b10a/SCons/Tools/mstoolkit.py#L104-L128 |
|
wxWidgets/wxPython-Classic | 19571e1ae65f1ac445f5491474121998c97a1bf0 | src/osx_carbon/_core.py | python | WindowCreateEvent.__init__ | (self, *args, **kwargs) | __init__(self, Window win=None) -> WindowCreateEvent
The EVT_WINDOW_CREATE event is sent as soon as the window object (the
underlying GUI object) exists. | __init__(self, Window win=None) -> WindowCreateEvent | [
"__init__",
"(",
"self",
"Window",
"win",
"=",
"None",
")",
"-",
">",
"WindowCreateEvent"
] | def __init__(self, *args, **kwargs):
"""
__init__(self, Window win=None) -> WindowCreateEvent
The EVT_WINDOW_CREATE event is sent as soon as the window object (the
underlying GUI object) exists.
"""
_core_.WindowCreateEvent_swiginit(self,_core_.new_WindowCreateEvent(*args, **kwargs)) | [
"def",
"__init__",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"_core_",
".",
"WindowCreateEvent_swiginit",
"(",
"self",
",",
"_core_",
".",
"new_WindowCreateEvent",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
")"
] | https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_carbon/_core.py#L7332-L7339 |
||
Polidea/SiriusObfuscator | b0e590d8130e97856afe578869b83a209e2b19be | SymbolExtractorAndRenamer/lldb/scripts/Python/static-binding/lldb.py | python | SBValue.__eol_test__ | (val) | Default function for end of list test takes an SBValue object.
Return True if val is invalid or it corresponds to a null pointer.
Otherwise, return False. | Default function for end of list test takes an SBValue object. | [
"Default",
"function",
"for",
"end",
"of",
"list",
"test",
"takes",
"an",
"SBValue",
"object",
"."
] | def __eol_test__(val):
"""Default function for end of list test takes an SBValue object.
Return True if val is invalid or it corresponds to a null pointer.
Otherwise, return False.
"""
if not val or val.GetValueAsUnsigned() == 0:
return True
else:
return False | [
"def",
"__eol_test__",
"(",
"val",
")",
":",
"if",
"not",
"val",
"or",
"val",
".",
"GetValueAsUnsigned",
"(",
")",
"==",
"0",
":",
"return",
"True",
"else",
":",
"return",
"False"
] | https://github.com/Polidea/SiriusObfuscator/blob/b0e590d8130e97856afe578869b83a209e2b19be/SymbolExtractorAndRenamer/lldb/scripts/Python/static-binding/lldb.py#L11724-L11733 |
||
apple/swift-lldb | d74be846ef3e62de946df343e8c234bde93a8912 | scripts/Python/static-binding/lldb.py | python | SBDebugger.SetErrorFileHandle | (self, f, transfer_ownership) | return _lldb.SBDebugger_SetErrorFileHandle(self, f, transfer_ownership) | SetErrorFileHandle(SBDebugger self, FILE * f, bool transfer_ownership) | SetErrorFileHandle(SBDebugger self, FILE * f, bool transfer_ownership) | [
"SetErrorFileHandle",
"(",
"SBDebugger",
"self",
"FILE",
"*",
"f",
"bool",
"transfer_ownership",
")"
] | def SetErrorFileHandle(self, f, transfer_ownership):
"""SetErrorFileHandle(SBDebugger self, FILE * f, bool transfer_ownership)"""
return _lldb.SBDebugger_SetErrorFileHandle(self, f, transfer_ownership) | [
"def",
"SetErrorFileHandle",
"(",
"self",
",",
"f",
",",
"transfer_ownership",
")",
":",
"return",
"_lldb",
".",
"SBDebugger_SetErrorFileHandle",
"(",
"self",
",",
"f",
",",
"transfer_ownership",
")"
] | https://github.com/apple/swift-lldb/blob/d74be846ef3e62de946df343e8c234bde93a8912/scripts/Python/static-binding/lldb.py#L3887-L3889 |
|
miyosuda/TensorFlowAndroidMNIST | 7b5a4603d2780a8a2834575706e9001977524007 | jni-build/jni/include/tensorflow/contrib/learn/python/learn/dataframe/transforms/reader_source.py | python | ReaderSource.__init__ | (self,
reader_cls,
work_units,
reader_kwargs=None,
enqueue_size=None,
batch_size=1,
queue_capacity=None,
shuffle=False,
min_after_dequeue=None,
num_threads=1,
seed=None) | Initializes a ReaderSource.
Args:
reader_cls: A subclass of `tesorflow.ReaderBase` that will be used to read
from `work_units`.
work_units: A list that describes the source(s) of data to read.
Typically, this is a list of filenames.
reader_kwargs: A dictionary of kwargs to be passed to `reader_cls` when it
is constructed.
enqueue_size: block size for each read operation.
batch_size: The desired batch size of output. Defaults to 1.
queue_capacity: Capacity of the queue. Defaults to 10 * `batch_size`.
shuffle: Whether records will be shuffled before returning. Defaults to
false.
min_after_dequeue: Minimum number of elements in the queue to allow a
dequeue operation. Only used when `shuffle` is true. Defaults to
`queue_capacity` / 4.
num_threads: Number of threads that will be used for reading. Each thread
has its own instance of `reader_cls`.
seed: A seed used for shuffling. Only used if `shuffle` is true. | Initializes a ReaderSource. | [
"Initializes",
"a",
"ReaderSource",
"."
] | def __init__(self,
reader_cls,
work_units,
reader_kwargs=None,
enqueue_size=None,
batch_size=1,
queue_capacity=None,
shuffle=False,
min_after_dequeue=None,
num_threads=1,
seed=None):
"""Initializes a ReaderSource.
Args:
reader_cls: A subclass of `tesorflow.ReaderBase` that will be used to read
from `work_units`.
work_units: A list that describes the source(s) of data to read.
Typically, this is a list of filenames.
reader_kwargs: A dictionary of kwargs to be passed to `reader_cls` when it
is constructed.
enqueue_size: block size for each read operation.
batch_size: The desired batch size of output. Defaults to 1.
queue_capacity: Capacity of the queue. Defaults to 10 * `batch_size`.
shuffle: Whether records will be shuffled before returning. Defaults to
false.
min_after_dequeue: Minimum number of elements in the queue to allow a
dequeue operation. Only used when `shuffle` is true. Defaults to
`queue_capacity` / 4.
num_threads: Number of threads that will be used for reading. Each thread
has its own instance of `reader_cls`.
seed: A seed used for shuffling. Only used if `shuffle` is true.
"""
super(ReaderSource, self).__init__()
self._reader_cls = reader_cls
self._reader_kwargs = reader_kwargs
self._work_units = work_units
self._reader_kwargs = {} if reader_kwargs is None else reader_kwargs
if enqueue_size is None:
enqueue_size = max(1, int(batch_size / num_threads))
self._enqueue_size = enqueue_size
self._batch_size = batch_size
self._queue_capacity = (batch_size * 10 if queue_capacity is None else
queue_capacity)
self._shuffle = shuffle
self._min_after_dequeue = int(self.queue_capacity / 4 if min_after_dequeue
is None else min_after_dequeue)
self._num_threads = num_threads
self._seed = seed | [
"def",
"__init__",
"(",
"self",
",",
"reader_cls",
",",
"work_units",
",",
"reader_kwargs",
"=",
"None",
",",
"enqueue_size",
"=",
"None",
",",
"batch_size",
"=",
"1",
",",
"queue_capacity",
"=",
"None",
",",
"shuffle",
"=",
"False",
",",
"min_after_dequeue",
"=",
"None",
",",
"num_threads",
"=",
"1",
",",
"seed",
"=",
"None",
")",
":",
"super",
"(",
"ReaderSource",
",",
"self",
")",
".",
"__init__",
"(",
")",
"self",
".",
"_reader_cls",
"=",
"reader_cls",
"self",
".",
"_reader_kwargs",
"=",
"reader_kwargs",
"self",
".",
"_work_units",
"=",
"work_units",
"self",
".",
"_reader_kwargs",
"=",
"{",
"}",
"if",
"reader_kwargs",
"is",
"None",
"else",
"reader_kwargs",
"if",
"enqueue_size",
"is",
"None",
":",
"enqueue_size",
"=",
"max",
"(",
"1",
",",
"int",
"(",
"batch_size",
"/",
"num_threads",
")",
")",
"self",
".",
"_enqueue_size",
"=",
"enqueue_size",
"self",
".",
"_batch_size",
"=",
"batch_size",
"self",
".",
"_queue_capacity",
"=",
"(",
"batch_size",
"*",
"10",
"if",
"queue_capacity",
"is",
"None",
"else",
"queue_capacity",
")",
"self",
".",
"_shuffle",
"=",
"shuffle",
"self",
".",
"_min_after_dequeue",
"=",
"int",
"(",
"self",
".",
"queue_capacity",
"/",
"4",
"if",
"min_after_dequeue",
"is",
"None",
"else",
"min_after_dequeue",
")",
"self",
".",
"_num_threads",
"=",
"num_threads",
"self",
".",
"_seed",
"=",
"seed"
] | https://github.com/miyosuda/TensorFlowAndroidMNIST/blob/7b5a4603d2780a8a2834575706e9001977524007/jni-build/jni/include/tensorflow/contrib/learn/python/learn/dataframe/transforms/reader_source.py#L29-L76 |
||
aws/lumberyard | f85344403c1c2e77ec8c75deb2c116e97b713217 | dev/Tools/Python/3.7.10/windows/Lib/email/_policybase.py | python | Policy.header_fetch_parse | (self, name, value) | Given the header name and the value from the model, return the value
to be returned to the application program that is requesting that
header. The value passed in by the email package may contain
surrogateescaped binary data if the lines were parsed by a BytesParser.
The returned value should not contain any surrogateescaped data. | Given the header name and the value from the model, return the value
to be returned to the application program that is requesting that
header. The value passed in by the email package may contain
surrogateescaped binary data if the lines were parsed by a BytesParser.
The returned value should not contain any surrogateescaped data. | [
"Given",
"the",
"header",
"name",
"and",
"the",
"value",
"from",
"the",
"model",
"return",
"the",
"value",
"to",
"be",
"returned",
"to",
"the",
"application",
"program",
"that",
"is",
"requesting",
"that",
"header",
".",
"The",
"value",
"passed",
"in",
"by",
"the",
"email",
"package",
"may",
"contain",
"surrogateescaped",
"binary",
"data",
"if",
"the",
"lines",
"were",
"parsed",
"by",
"a",
"BytesParser",
".",
"The",
"returned",
"value",
"should",
"not",
"contain",
"any",
"surrogateescaped",
"data",
"."
] | def header_fetch_parse(self, name, value):
"""Given the header name and the value from the model, return the value
to be returned to the application program that is requesting that
header. The value passed in by the email package may contain
surrogateescaped binary data if the lines were parsed by a BytesParser.
The returned value should not contain any surrogateescaped data.
"""
raise NotImplementedError | [
"def",
"header_fetch_parse",
"(",
"self",
",",
"name",
",",
"value",
")",
":",
"raise",
"NotImplementedError"
] | https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/windows/Lib/email/_policybase.py#L238-L246 |
||
zju3dv/clean-pvnet | 5870c509e3cc205e1bb28910a7b1a9a3c8add9a8 | lib/utils/meshrenderer/pysixd/transform.py | python | is_same_transform | (matrix0, matrix1) | return numpy.allclose(matrix0, matrix1) | Return True if two matrices perform same transformation.
>>> is_same_transform(numpy.identity(4), numpy.identity(4))
True
>>> is_same_transform(numpy.identity(4), random_rotation_matrix())
False | Return True if two matrices perform same transformation. | [
"Return",
"True",
"if",
"two",
"matrices",
"perform",
"same",
"transformation",
"."
] | def is_same_transform(matrix0, matrix1):
"""Return True if two matrices perform same transformation.
>>> is_same_transform(numpy.identity(4), numpy.identity(4))
True
>>> is_same_transform(numpy.identity(4), random_rotation_matrix())
False
"""
matrix0 = numpy.array(matrix0, dtype=numpy.float64, copy=True)
matrix0 /= matrix0[3, 3]
matrix1 = numpy.array(matrix1, dtype=numpy.float64, copy=True)
matrix1 /= matrix1[3, 3]
return numpy.allclose(matrix0, matrix1) | [
"def",
"is_same_transform",
"(",
"matrix0",
",",
"matrix1",
")",
":",
"matrix0",
"=",
"numpy",
".",
"array",
"(",
"matrix0",
",",
"dtype",
"=",
"numpy",
".",
"float64",
",",
"copy",
"=",
"True",
")",
"matrix0",
"/=",
"matrix0",
"[",
"3",
",",
"3",
"]",
"matrix1",
"=",
"numpy",
".",
"array",
"(",
"matrix1",
",",
"dtype",
"=",
"numpy",
".",
"float64",
",",
"copy",
"=",
"True",
")",
"matrix1",
"/=",
"matrix1",
"[",
"3",
",",
"3",
"]",
"return",
"numpy",
".",
"allclose",
"(",
"matrix0",
",",
"matrix1",
")"
] | https://github.com/zju3dv/clean-pvnet/blob/5870c509e3cc205e1bb28910a7b1a9a3c8add9a8/lib/utils/meshrenderer/pysixd/transform.py#L1861-L1874 |
|
flink-extended/dl-on-flink | 60646aa9520f49619b64e9ff03ce73959e8a3858 | flink-ml-tensorflow/python/flink_ml_tensorflow/gpu_info.py | python | _get_free_gpu | (max_gpu_utilization=40, min_free_memory=0.5, num_gpu=1) | return gpus_to_use, free_memory | Get available GPUs according to utilization thresholds.
Args:
:max_gpu_utilization: percent utilization threshold to consider a GPU "free"
:min_free_memory: percent free memory to consider a GPU "free"
:num_gpu: number of requested GPUs
Returns:
A tuple of (available_gpus, minimum_free_memory), where available_gpus is a comma-delimited string of GPU ids, and minimum_free_memory
is the lowest amount of free memory available on the available_gpus. | Get available GPUs according to utilization thresholds. | [
"Get",
"available",
"GPUs",
"according",
"to",
"utilization",
"thresholds",
"."
] | def _get_free_gpu(max_gpu_utilization=40, min_free_memory=0.5, num_gpu=1):
"""Get available GPUs according to utilization thresholds.
Args:
:max_gpu_utilization: percent utilization threshold to consider a GPU "free"
:min_free_memory: percent free memory to consider a GPU "free"
:num_gpu: number of requested GPUs
Returns:
A tuple of (available_gpus, minimum_free_memory), where available_gpus is a comma-delimited string of GPU ids, and minimum_free_memory
is the lowest amount of free memory available on the available_gpus.
"""
def get_gpu_info():
# Get the gpu information
gpu_info = subprocess.check_output(["nvidia-smi", "--format=csv,noheader,nounits",
"--query-gpu=index,memory.total,memory.free,memory.used,utilization.gpu"]).decode()
gpu_info = gpu_info.split('\n')
gpu_info_array = []
# Check each gpu
for line in gpu_info:
if len(line) > 0:
gpu_id, total_memory, free_memory, used_memory, gpu_util = line.split(',')
gpu_memory_util = float(used_memory) / float(total_memory)
gpu_info_array.append((float(gpu_util), gpu_memory_util, gpu_id))
return (gpu_info_array)
# Read the gpu information multiple times
num_times_to_average = 5
current_array = []
for ind in range(num_times_to_average):
current_array.append(get_gpu_info())
time.sleep(1)
# Get number of gpus
num_gpus = len(current_array[0])
# Average the gpu information
avg_array = [(0, 0, str(x)) for x in range(num_gpus)]
for ind in range(num_times_to_average):
for gpu_ind in range(num_gpus):
avg_array[gpu_ind] = (avg_array[gpu_ind][0] + current_array[ind][gpu_ind][0],
avg_array[gpu_ind][1] + current_array[ind][gpu_ind][1], avg_array[gpu_ind][2])
for gpu_ind in range(num_gpus):
avg_array[gpu_ind] = (
float(avg_array[gpu_ind][0]) / num_times_to_average, float(avg_array[gpu_ind][1]) / num_times_to_average,
avg_array[gpu_ind][2])
avg_array.sort()
gpus_found = 0
gpus_to_use = ""
free_memory = 1.0
# Return the least utilized GPUs if it's utilized less than max_gpu_utilization and amount of free memory is at least min_free_memory
# Otherwise, run in cpu only mode
for current_gpu in avg_array:
if current_gpu[0] < max_gpu_utilization and (1 - current_gpu[1]) > min_free_memory:
if gpus_found == 0:
gpus_to_use = current_gpu[2]
free_memory = 1 - current_gpu[1]
else:
gpus_to_use = gpus_to_use + "," + current_gpu[2]
free_memory = min(free_memory, 1 - current_gpu[1])
gpus_found = gpus_found + 1
if gpus_found == num_gpu:
break
return gpus_to_use, free_memory | [
"def",
"_get_free_gpu",
"(",
"max_gpu_utilization",
"=",
"40",
",",
"min_free_memory",
"=",
"0.5",
",",
"num_gpu",
"=",
"1",
")",
":",
"def",
"get_gpu_info",
"(",
")",
":",
"# Get the gpu information",
"gpu_info",
"=",
"subprocess",
".",
"check_output",
"(",
"[",
"\"nvidia-smi\"",
",",
"\"--format=csv,noheader,nounits\"",
",",
"\"--query-gpu=index,memory.total,memory.free,memory.used,utilization.gpu\"",
"]",
")",
".",
"decode",
"(",
")",
"gpu_info",
"=",
"gpu_info",
".",
"split",
"(",
"'\\n'",
")",
"gpu_info_array",
"=",
"[",
"]",
"# Check each gpu",
"for",
"line",
"in",
"gpu_info",
":",
"if",
"len",
"(",
"line",
")",
">",
"0",
":",
"gpu_id",
",",
"total_memory",
",",
"free_memory",
",",
"used_memory",
",",
"gpu_util",
"=",
"line",
".",
"split",
"(",
"','",
")",
"gpu_memory_util",
"=",
"float",
"(",
"used_memory",
")",
"/",
"float",
"(",
"total_memory",
")",
"gpu_info_array",
".",
"append",
"(",
"(",
"float",
"(",
"gpu_util",
")",
",",
"gpu_memory_util",
",",
"gpu_id",
")",
")",
"return",
"(",
"gpu_info_array",
")",
"# Read the gpu information multiple times",
"num_times_to_average",
"=",
"5",
"current_array",
"=",
"[",
"]",
"for",
"ind",
"in",
"range",
"(",
"num_times_to_average",
")",
":",
"current_array",
".",
"append",
"(",
"get_gpu_info",
"(",
")",
")",
"time",
".",
"sleep",
"(",
"1",
")",
"# Get number of gpus",
"num_gpus",
"=",
"len",
"(",
"current_array",
"[",
"0",
"]",
")",
"# Average the gpu information",
"avg_array",
"=",
"[",
"(",
"0",
",",
"0",
",",
"str",
"(",
"x",
")",
")",
"for",
"x",
"in",
"range",
"(",
"num_gpus",
")",
"]",
"for",
"ind",
"in",
"range",
"(",
"num_times_to_average",
")",
":",
"for",
"gpu_ind",
"in",
"range",
"(",
"num_gpus",
")",
":",
"avg_array",
"[",
"gpu_ind",
"]",
"=",
"(",
"avg_array",
"[",
"gpu_ind",
"]",
"[",
"0",
"]",
"+",
"current_array",
"[",
"ind",
"]",
"[",
"gpu_ind",
"]",
"[",
"0",
"]",
",",
"avg_array",
"[",
"gpu_ind",
"]",
"[",
"1",
"]",
"+",
"current_array",
"[",
"ind",
"]",
"[",
"gpu_ind",
"]",
"[",
"1",
"]",
",",
"avg_array",
"[",
"gpu_ind",
"]",
"[",
"2",
"]",
")",
"for",
"gpu_ind",
"in",
"range",
"(",
"num_gpus",
")",
":",
"avg_array",
"[",
"gpu_ind",
"]",
"=",
"(",
"float",
"(",
"avg_array",
"[",
"gpu_ind",
"]",
"[",
"0",
"]",
")",
"/",
"num_times_to_average",
",",
"float",
"(",
"avg_array",
"[",
"gpu_ind",
"]",
"[",
"1",
"]",
")",
"/",
"num_times_to_average",
",",
"avg_array",
"[",
"gpu_ind",
"]",
"[",
"2",
"]",
")",
"avg_array",
".",
"sort",
"(",
")",
"gpus_found",
"=",
"0",
"gpus_to_use",
"=",
"\"\"",
"free_memory",
"=",
"1.0",
"# Return the least utilized GPUs if it's utilized less than max_gpu_utilization and amount of free memory is at least min_free_memory",
"# Otherwise, run in cpu only mode",
"for",
"current_gpu",
"in",
"avg_array",
":",
"if",
"current_gpu",
"[",
"0",
"]",
"<",
"max_gpu_utilization",
"and",
"(",
"1",
"-",
"current_gpu",
"[",
"1",
"]",
")",
">",
"min_free_memory",
":",
"if",
"gpus_found",
"==",
"0",
":",
"gpus_to_use",
"=",
"current_gpu",
"[",
"2",
"]",
"free_memory",
"=",
"1",
"-",
"current_gpu",
"[",
"1",
"]",
"else",
":",
"gpus_to_use",
"=",
"gpus_to_use",
"+",
"\",\"",
"+",
"current_gpu",
"[",
"2",
"]",
"free_memory",
"=",
"min",
"(",
"free_memory",
",",
"1",
"-",
"current_gpu",
"[",
"1",
"]",
")",
"gpus_found",
"=",
"gpus_found",
"+",
"1",
"if",
"gpus_found",
"==",
"num_gpu",
":",
"break",
"return",
"gpus_to_use",
",",
"free_memory"
] | https://github.com/flink-extended/dl-on-flink/blob/60646aa9520f49619b64e9ff03ce73959e8a3858/flink-ml-tensorflow/python/flink_ml_tensorflow/gpu_info.py#L128-L202 |
|
aws/lumberyard | f85344403c1c2e77ec8c75deb2c116e97b713217 | dev/Gems/CloudGemMetric/v1/AWS/python/windows/Lib/pandas/compat/numpy/function.py | python | validate_argsort_with_ascending | (ascending, args, kwargs) | return ascending | If 'Categorical.argsort' is called via the 'numpy' library, the
first parameter in its signature is 'axis', which takes either
an integer or 'None', so check if the 'ascending' parameter has
either integer type or is None, since 'ascending' itself should
be a boolean | If 'Categorical.argsort' is called via the 'numpy' library, the
first parameter in its signature is 'axis', which takes either
an integer or 'None', so check if the 'ascending' parameter has
either integer type or is None, since 'ascending' itself should
be a boolean | [
"If",
"Categorical",
".",
"argsort",
"is",
"called",
"via",
"the",
"numpy",
"library",
"the",
"first",
"parameter",
"in",
"its",
"signature",
"is",
"axis",
"which",
"takes",
"either",
"an",
"integer",
"or",
"None",
"so",
"check",
"if",
"the",
"ascending",
"parameter",
"has",
"either",
"integer",
"type",
"or",
"is",
"None",
"since",
"ascending",
"itself",
"should",
"be",
"a",
"boolean"
] | def validate_argsort_with_ascending(ascending, args, kwargs):
"""
If 'Categorical.argsort' is called via the 'numpy' library, the
first parameter in its signature is 'axis', which takes either
an integer or 'None', so check if the 'ascending' parameter has
either integer type or is None, since 'ascending' itself should
be a boolean
"""
if is_integer(ascending) or ascending is None:
args = (ascending,) + args
ascending = True
validate_argsort_kind(args, kwargs, max_fname_arg_count=3)
return ascending | [
"def",
"validate_argsort_with_ascending",
"(",
"ascending",
",",
"args",
",",
"kwargs",
")",
":",
"if",
"is_integer",
"(",
"ascending",
")",
"or",
"ascending",
"is",
"None",
":",
"args",
"=",
"(",
"ascending",
",",
")",
"+",
"args",
"ascending",
"=",
"True",
"validate_argsort_kind",
"(",
"args",
",",
"kwargs",
",",
"max_fname_arg_count",
"=",
"3",
")",
"return",
"ascending"
] | https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Gems/CloudGemMetric/v1/AWS/python/windows/Lib/pandas/compat/numpy/function.py#L133-L147 |
|
miyosuda/TensorFlowAndroidDemo | 35903e0221aa5f109ea2dbef27f20b52e317f42d | jni-build/jni/include/tensorflow/contrib/graph_editor/util.py | python | get_tensors | (graph) | return ts | get all the tensors which are input or output of an op in the graph.
Args:
graph: a tf.Graph.
Returns:
A list of tf.Tensor.
Raises:
TypeError: if graph is not a tf.Graph. | get all the tensors which are input or output of an op in the graph. | [
"get",
"all",
"the",
"tensors",
"which",
"are",
"input",
"or",
"output",
"of",
"an",
"op",
"in",
"the",
"graph",
"."
] | def get_tensors(graph):
"""get all the tensors which are input or output of an op in the graph.
Args:
graph: a tf.Graph.
Returns:
A list of tf.Tensor.
Raises:
TypeError: if graph is not a tf.Graph.
"""
if not isinstance(graph, tf_ops.Graph):
raise TypeError("Expected a graph, got: {}".format(type(graph)))
ts = []
for op in graph.get_operations():
concatenate_unique(ts, op.inputs)
concatenate_unique(ts, op.outputs)
return ts | [
"def",
"get_tensors",
"(",
"graph",
")",
":",
"if",
"not",
"isinstance",
"(",
"graph",
",",
"tf_ops",
".",
"Graph",
")",
":",
"raise",
"TypeError",
"(",
"\"Expected a graph, got: {}\"",
".",
"format",
"(",
"type",
"(",
"graph",
")",
")",
")",
"ts",
"=",
"[",
"]",
"for",
"op",
"in",
"graph",
".",
"get_operations",
"(",
")",
":",
"concatenate_unique",
"(",
"ts",
",",
"op",
".",
"inputs",
")",
"concatenate_unique",
"(",
"ts",
",",
"op",
".",
"outputs",
")",
"return",
"ts"
] | https://github.com/miyosuda/TensorFlowAndroidDemo/blob/35903e0221aa5f109ea2dbef27f20b52e317f42d/jni-build/jni/include/tensorflow/contrib/graph_editor/util.py#L152-L168 |
|
pytorch/pytorch | 7176c92687d3cc847cc046bf002269c6949a21c2 | torch/ao/ns/fx/graph_passes.py | python | create_a_shadows_b | (
name_a: str,
gm_a: GraphModule,
name_b: str,
gm_b: GraphModule,
matched_subgraph_pairs: Dict[str, Tuple[NSSubgraph, NSSubgraph]],
logger_cls: Callable,
should_log_inputs: bool,
node_type_to_io_type_map: Optional[Dict[str, Set[NSNodeTargetType]]] = None,
) | return gm_c | Creates a new GraphModule consisting of the graph of C, with the meaningful
nodes of A shadowing the corresponding nodes of B. For example,
Graph A:
a0 -> op0_fp32 -> a1 -> op1_fp32 -> a2
Graph B:
b0 -> op0_int8 -> b1 -> op1_int8 -> b2
matched_node_pairs: {'op0': (op0_fp32, op0_int8), 'op1': (op1_fp32, op1_int8)}
Graph C (A shadows B):
/ dequant0 -> op0_fp32 -> logger_a_0 / dequant_1 -> op1_fp32 -> logger_a_1
/ /
b0 -------------> op0_int8 -> logger_b_0 --------------> op1_int8 -> logger_b_1
In a nutshell, this function does the following for each node pair:
* copies the necessary attributes and modules from gm_a to gm_b,
keeping names unique
* adds a dtype cast op (dequant, quant, etc)
* adds a copy of node_a in gm_b's graph
* adds loggers to the outputs of node_a and node_b | Creates a new GraphModule consisting of the graph of C, with the meaningful
nodes of A shadowing the corresponding nodes of B. For example, | [
"Creates",
"a",
"new",
"GraphModule",
"consisting",
"of",
"the",
"graph",
"of",
"C",
"with",
"the",
"meaningful",
"nodes",
"of",
"A",
"shadowing",
"the",
"corresponding",
"nodes",
"of",
"B",
".",
"For",
"example"
] | def create_a_shadows_b(
name_a: str,
gm_a: GraphModule,
name_b: str,
gm_b: GraphModule,
matched_subgraph_pairs: Dict[str, Tuple[NSSubgraph, NSSubgraph]],
logger_cls: Callable,
should_log_inputs: bool,
node_type_to_io_type_map: Optional[Dict[str, Set[NSNodeTargetType]]] = None,
) -> GraphModule:
"""
Creates a new GraphModule consisting of the graph of C, with the meaningful
nodes of A shadowing the corresponding nodes of B. For example,
Graph A:
a0 -> op0_fp32 -> a1 -> op1_fp32 -> a2
Graph B:
b0 -> op0_int8 -> b1 -> op1_int8 -> b2
matched_node_pairs: {'op0': (op0_fp32, op0_int8), 'op1': (op1_fp32, op1_int8)}
Graph C (A shadows B):
/ dequant0 -> op0_fp32 -> logger_a_0 / dequant_1 -> op1_fp32 -> logger_a_1
/ /
b0 -------------> op0_int8 -> logger_b_0 --------------> op1_int8 -> logger_b_1
In a nutshell, this function does the following for each node pair:
* copies the necessary attributes and modules from gm_a to gm_b,
keeping names unique
* adds a dtype cast op (dequant, quant, etc)
* adds a copy of node_a in gm_b's graph
* adds loggers to the outputs of node_a and node_b
"""
if node_type_to_io_type_map is None:
node_type_to_io_type_map = get_node_type_to_io_type_map()
# graph_c is the graph created from copying the nodes of graph_b and inserting
# the shadows with the nodes copied from graph_a
graph_c = Graph()
env_c: Dict[str, Any] = {}
modules = dict(gm_b.named_modules())
def load_arg(a):
return map_arg(a, lambda node: env_c[node.name])
start_node_b_to_matched_subgraph_a_and_name = {}
end_node_b_to_matched_subgraph_a_and_name = {}
for match_name, match in matched_subgraph_pairs.items():
subgraph_a, subgraph_b = match
ref_node_type_a = get_target_type_str(subgraph_a.base_op_node, gm_a)
ref_node_type_b = get_target_type_str(subgraph_b.base_op_node, gm_b)
start_node_b_to_matched_subgraph_a_and_name[subgraph_b.start_node] = \
(subgraph_a, match_name, ref_node_type_a, ref_node_type_b)
end_node_b_to_matched_subgraph_a_and_name[subgraph_b.end_node] = \
(subgraph_a, match_name, ref_node_type_a, ref_node_type_b)
for node_b in gm_b.graph.nodes:
if node_b.op == 'output':
graph_c.output(map_arg(node_b.args[0], load_arg))
continue
# calculate the flags to determine what to do with this node
node_b_is_start_node = node_b in start_node_b_to_matched_subgraph_a_and_name
node_b_is_end_node = node_b in end_node_b_to_matched_subgraph_a_and_name
if (node_b_is_start_node or node_b_is_end_node):
if node_b_is_start_node:
subgraph_a, ref_name, ref_node_type_a, ref_node_type_b = \
start_node_b_to_matched_subgraph_a_and_name[node_b]
else:
assert node_b_is_end_node
subgraph_a, ref_name, ref_node_type_a, ref_node_type_b = \
end_node_b_to_matched_subgraph_a_and_name[node_b]
# For both start_node and end_node verify that we know how to do
# the dtype cast. If we do not, skip.
node_input_type_a, node_output_type_a = \
get_node_first_input_and_output_type(
subgraph_a.start_node, gm_a, logger_cls,
node_type_to_io_type_map)
node_input_type_b, node_output_type_b = \
get_node_first_input_and_output_type(
node_b, gm_b, logger_cls,
node_type_to_io_type_map)
node_io_types_known_a_and_b = (
node_input_type_a != NodeInputOrOutputType.UNKNOWN and
node_output_type_a != NodeInputOrOutputType.UNKNOWN and
node_input_type_b != NodeInputOrOutputType.UNKNOWN and
node_output_type_b != NodeInputOrOutputType.UNKNOWN
)
if not node_io_types_known_a_and_b:
print(
f'skipping shadow loggers for node_b: {get_target_type_str(node_b, gm_b)}' +
f', start_node_a: {get_target_type_str(subgraph_a.start_node, gm_a)}' +
', unknown dtype cast')
env_c[node_b.name] = graph_c.node_copy(node_b, load_arg)
continue
# If we are shadowing from fp32 to int8, we need to insert
# quantize_per_tensor call with qparams from the previous node.
# Only do this if we are able to infer these qparams from the graph.
if (
node_input_type_a == NodeInputOrOutputType.INT8 and
node_input_type_b == NodeInputOrOutputType.FP32
):
node_a_input_qparams = get_node_input_qparams(
subgraph_a.start_node, gm_a, node_type_to_io_type_map)
if not node_a_input_qparams:
print(
f'skipping shadow loggers for node_b: {get_target_type_str(node_b, gm_b)}' +
f', start_node_a: {get_target_type_str(subgraph_a.start_node, gm_a)}' +
', unknown input qparams')
env_c[node_b.name] = graph_c.node_copy(node_b, load_arg)
continue
fqn_base_a = _maybe_get_fqn(subgraph_a.base_op_node, gm_a)
fqn_base_b = _maybe_get_fqn(subgraph_b.base_op_node, gm_b)
if node_b_is_start_node:
# if necessary, log the input of node_c
if should_log_inputs:
if isinstance(node_b.args[0], Node):
prev_node_c = env_c[node_b.args[0].name]
env_c[prev_node_c.name] = _insert_logger_after_node(
prev_node_c, gm_b, logger_cls, '_ns_logger_b_inp_',
node_b.name, name_b, ref_name, ref_node_type_b,
NSSingleResultValuesType.NODE_INPUT.value,
index_within_arg=0, index_of_arg=0,
fqn=fqn_base_b)
elif isinstance(node_b.args[0], list):
# first, save the prev_node instances, because they
# will be overwritten in the env after the first logger
# is added
prev_node_c_list = [env_c[arg.name] for arg in node_b.args[0]]
for arg_idx, arg in enumerate(node_b.args[0]):
prev_node_c = prev_node_c_list[arg_idx]
env_c[prev_node_c.name] = _insert_logger_after_node(
prev_node_c, gm_b, logger_cls, '_ns_logger_b_inp_',
node_b.name, name_b, ref_name, ref_node_type_b,
NSSingleResultValuesType.NODE_INPUT.value,
index_within_arg=arg_idx, index_of_arg=0,
fqn=fqn_base_b)
else:
# logging of inputs which are not lists is not supported yet
raise AssertionError(f"type {type(node_b.args[0])} is not handled yet")
# subgraph so far:
#
# (prev_node_c)+ -> (logger_c_input)?
# Note: this if statement is always True, spelling it out to clarify code
# intent.
if node_b_is_start_node or node_b_is_end_node:
# ensure env_c is populated with base node
env_c[node_b.name] = graph_c.node_copy(node_b, load_arg)
node_c = env_c[node_b.name]
# after this point,
#
# node_a is the original node from graph_a, with parent module gm_a
# node_b is the original node from graph_b, with parent module gm_b
# node_c is the copy of node_b in graph_c
#
# subgraph so far:
#
# (prev_node_c)+ -> (logger_c_input)? -> node_start_c
if node_b_is_start_node:
# cast dtype from the dtype of node_c's input to the dtype of
# node_a's input (dequant, etc)
prev_node_c = node_c.args[0]
if should_log_inputs:
# skip the input logger when inserting a dtype cast
if isinstance(prev_node_c, Node):
prev_node_c = prev_node_c.args[0]
elif isinstance(prev_node_c, list):
prev_node_c = [arg.args[0] for arg in prev_node_c]
dtype_cast_node = _insert_dtype_cast_after_node(
subgraph_a.start_node, node_c, prev_node_c, gm_a, gm_b, graph_c,
node_b.name + '_dtype_cast_', logger_cls,
node_type_to_io_type_map)
# note: not inserting to env_c because all nodes which use the dtype
# casts are copied from graph_a
#
# subgraph so far:
#
# (dtype_cast_node)+
# /
# (prev_node_c)+ -> (logger_c_input)? -> node_start_c
# if input logging is enabled, log the input to the subgraph
if should_log_inputs:
# TODO: explain this
ref_node_name = ''
if isinstance(dtype_cast_node, Node):
dtype_cast_node = _insert_logger_after_node(
dtype_cast_node, gm_b, logger_cls, '_ns_logger_a_inp_',
ref_node_name, name_a, ref_name, ref_node_type_a,
NSSingleResultValuesType.NODE_INPUT.value,
index_within_arg=0, index_of_arg=0,
fqn=fqn_base_a)
input_logger: Union[Node, List[Node]] = dtype_cast_node
else:
assert isinstance(dtype_cast_node, list)
new_loggers = []
for dtype_cast_idx, dtype_cast_node_inner in enumerate(dtype_cast_node):
dtype_cast_logger = _insert_logger_after_node(
dtype_cast_node_inner, gm_b, logger_cls, '_ns_logger_a_inp_',
ref_node_name, name_a, ref_name, ref_node_type_a,
NSSingleResultValuesType.NODE_INPUT.value,
index_within_arg=dtype_cast_idx,
index_of_arg=0,
fqn=fqn_base_a)
new_loggers.append(dtype_cast_logger)
dtype_cast_node = new_loggers
input_logger = dtype_cast_node
# subgraph so far:
#
# (dtype_cast_node)+ -> (logger_a_input)?
# /
# prev_node_c -> (logger_c_input)? -> node_start_c
# hook up the new mod_a copy to be in the graph, receiving the
# same inputs as mod_b does, with dtype cast to match a
# Some ops, such as LSTMs, have two non-param inputs. If we have
# such an op, pass the second param as well. Note: dtype casting
# for the second param is not implemented yet, it can be added
# later if there is a use case.
node_c_second_non_param_arg = None
num_non_param_args_node_a = get_number_of_non_param_args(subgraph_a.start_node, gm_a)
if num_non_param_args_node_a == 2:
node_c_second_non_param_arg = node_c.args[1]
node_a_shadows_c = _insert_copy_of_subgraph_a_after_input_node_c(
dtype_cast_node, node_c_second_non_param_arg,
subgraph_a, gm_a, gm_b, node_c.name + '_shadow_copy_')
env_c[node_a_shadows_c.name] = node_a_shadows_c
# subgraph so far:
#
# dtype_cast_node -> (logger_a_input)? -> subgraph_a_copy(args/kwargs not shown)
# /
# (prev_node_c)+ -> (logger_c_input)? -> node_start_c
if should_log_inputs:
# When we created the input logger, we left the ref_node_name
# as an empty string, because the subgraph copy did not exist
# yet. Now that the subgraph copy exists, we modify this name
# to its true value.
# Note: the alternative to this is to create the input logger
# after creating the subgraph, which is slightly more
# complicated. This is the lesser of two evils.
# input_logger = env_c[dtype_cast_node.name]
# Find the first node in the subgraph
cur_node = node_a_shadows_c
while cur_node.args[0] != input_logger:
cur_node = cur_node.args[0] # type: ignore[assignment]
if isinstance(input_logger, Node):
input_logger_mod = getattr(gm_b, input_logger.name)
input_logger_mod.ref_node_name = cur_node.name
else:
assert isinstance(input_logger, list)
for input_logger_inner in input_logger:
input_logger_mod = getattr(gm_b, input_logger_inner.name)
input_logger_mod.ref_node_name = cur_node.name
# hook up a logger to the mod_a copy
env_c[node_a_shadows_c.name] = _insert_logger_after_node(
env_c[node_a_shadows_c.name], gm_b, logger_cls, '_ns_logger_a_',
node_a_shadows_c.name, name_a, ref_name, ref_node_type_a,
NSSingleResultValuesType.NODE_OUTPUT.value,
index_within_arg=0, index_of_arg=0,
fqn=fqn_base_a)
# subgraph so far:
#
# dtype_cast_node -> (logger_a_input)? -> subgraph_a_copy -> logger_a
# /
# (prev_node_c)+ -> (logger_c_input)? -> node_start_c
if node_b_is_end_node:
# hook up a logger to the mod_b copy
env_c[node_b.name] = _insert_logger_after_node(
env_c[node_b.name], gm_b, logger_cls, '_ns_logger_b_',
node_b.name, name_b, ref_name, ref_node_type_b,
NSSingleResultValuesType.NODE_OUTPUT.value,
index_within_arg=0, index_of_arg=0,
fqn=fqn_base_b)
# subgraph so far:
#
# dtype_cast_node -> (logger_a_input)? -> subgraph_a_copy -> logger_a
# /
# (prev_node_c+) -> (logger_c_input)? -> node_start_c -> ... -> node_end_c -> logger_c
#
# Note: node_start_c may be the same node as node_end_c, or they
# may have nodes inbetween.
else:
env_c[node_b.name] = graph_c.node_copy(node_b, load_arg)
gm_c = GraphModule(gm_b, graph_c)
return gm_c | [
"def",
"create_a_shadows_b",
"(",
"name_a",
":",
"str",
",",
"gm_a",
":",
"GraphModule",
",",
"name_b",
":",
"str",
",",
"gm_b",
":",
"GraphModule",
",",
"matched_subgraph_pairs",
":",
"Dict",
"[",
"str",
",",
"Tuple",
"[",
"NSSubgraph",
",",
"NSSubgraph",
"]",
"]",
",",
"logger_cls",
":",
"Callable",
",",
"should_log_inputs",
":",
"bool",
",",
"node_type_to_io_type_map",
":",
"Optional",
"[",
"Dict",
"[",
"str",
",",
"Set",
"[",
"NSNodeTargetType",
"]",
"]",
"]",
"=",
"None",
",",
")",
"->",
"GraphModule",
":",
"if",
"node_type_to_io_type_map",
"is",
"None",
":",
"node_type_to_io_type_map",
"=",
"get_node_type_to_io_type_map",
"(",
")",
"# graph_c is the graph created from copying the nodes of graph_b and inserting",
"# the shadows with the nodes copied from graph_a",
"graph_c",
"=",
"Graph",
"(",
")",
"env_c",
":",
"Dict",
"[",
"str",
",",
"Any",
"]",
"=",
"{",
"}",
"modules",
"=",
"dict",
"(",
"gm_b",
".",
"named_modules",
"(",
")",
")",
"def",
"load_arg",
"(",
"a",
")",
":",
"return",
"map_arg",
"(",
"a",
",",
"lambda",
"node",
":",
"env_c",
"[",
"node",
".",
"name",
"]",
")",
"start_node_b_to_matched_subgraph_a_and_name",
"=",
"{",
"}",
"end_node_b_to_matched_subgraph_a_and_name",
"=",
"{",
"}",
"for",
"match_name",
",",
"match",
"in",
"matched_subgraph_pairs",
".",
"items",
"(",
")",
":",
"subgraph_a",
",",
"subgraph_b",
"=",
"match",
"ref_node_type_a",
"=",
"get_target_type_str",
"(",
"subgraph_a",
".",
"base_op_node",
",",
"gm_a",
")",
"ref_node_type_b",
"=",
"get_target_type_str",
"(",
"subgraph_b",
".",
"base_op_node",
",",
"gm_b",
")",
"start_node_b_to_matched_subgraph_a_and_name",
"[",
"subgraph_b",
".",
"start_node",
"]",
"=",
"(",
"subgraph_a",
",",
"match_name",
",",
"ref_node_type_a",
",",
"ref_node_type_b",
")",
"end_node_b_to_matched_subgraph_a_and_name",
"[",
"subgraph_b",
".",
"end_node",
"]",
"=",
"(",
"subgraph_a",
",",
"match_name",
",",
"ref_node_type_a",
",",
"ref_node_type_b",
")",
"for",
"node_b",
"in",
"gm_b",
".",
"graph",
".",
"nodes",
":",
"if",
"node_b",
".",
"op",
"==",
"'output'",
":",
"graph_c",
".",
"output",
"(",
"map_arg",
"(",
"node_b",
".",
"args",
"[",
"0",
"]",
",",
"load_arg",
")",
")",
"continue",
"# calculate the flags to determine what to do with this node",
"node_b_is_start_node",
"=",
"node_b",
"in",
"start_node_b_to_matched_subgraph_a_and_name",
"node_b_is_end_node",
"=",
"node_b",
"in",
"end_node_b_to_matched_subgraph_a_and_name",
"if",
"(",
"node_b_is_start_node",
"or",
"node_b_is_end_node",
")",
":",
"if",
"node_b_is_start_node",
":",
"subgraph_a",
",",
"ref_name",
",",
"ref_node_type_a",
",",
"ref_node_type_b",
"=",
"start_node_b_to_matched_subgraph_a_and_name",
"[",
"node_b",
"]",
"else",
":",
"assert",
"node_b_is_end_node",
"subgraph_a",
",",
"ref_name",
",",
"ref_node_type_a",
",",
"ref_node_type_b",
"=",
"end_node_b_to_matched_subgraph_a_and_name",
"[",
"node_b",
"]",
"# For both start_node and end_node verify that we know how to do",
"# the dtype cast. If we do not, skip.",
"node_input_type_a",
",",
"node_output_type_a",
"=",
"get_node_first_input_and_output_type",
"(",
"subgraph_a",
".",
"start_node",
",",
"gm_a",
",",
"logger_cls",
",",
"node_type_to_io_type_map",
")",
"node_input_type_b",
",",
"node_output_type_b",
"=",
"get_node_first_input_and_output_type",
"(",
"node_b",
",",
"gm_b",
",",
"logger_cls",
",",
"node_type_to_io_type_map",
")",
"node_io_types_known_a_and_b",
"=",
"(",
"node_input_type_a",
"!=",
"NodeInputOrOutputType",
".",
"UNKNOWN",
"and",
"node_output_type_a",
"!=",
"NodeInputOrOutputType",
".",
"UNKNOWN",
"and",
"node_input_type_b",
"!=",
"NodeInputOrOutputType",
".",
"UNKNOWN",
"and",
"node_output_type_b",
"!=",
"NodeInputOrOutputType",
".",
"UNKNOWN",
")",
"if",
"not",
"node_io_types_known_a_and_b",
":",
"print",
"(",
"f'skipping shadow loggers for node_b: {get_target_type_str(node_b, gm_b)}'",
"+",
"f', start_node_a: {get_target_type_str(subgraph_a.start_node, gm_a)}'",
"+",
"', unknown dtype cast'",
")",
"env_c",
"[",
"node_b",
".",
"name",
"]",
"=",
"graph_c",
".",
"node_copy",
"(",
"node_b",
",",
"load_arg",
")",
"continue",
"# If we are shadowing from fp32 to int8, we need to insert",
"# quantize_per_tensor call with qparams from the previous node.",
"# Only do this if we are able to infer these qparams from the graph.",
"if",
"(",
"node_input_type_a",
"==",
"NodeInputOrOutputType",
".",
"INT8",
"and",
"node_input_type_b",
"==",
"NodeInputOrOutputType",
".",
"FP32",
")",
":",
"node_a_input_qparams",
"=",
"get_node_input_qparams",
"(",
"subgraph_a",
".",
"start_node",
",",
"gm_a",
",",
"node_type_to_io_type_map",
")",
"if",
"not",
"node_a_input_qparams",
":",
"print",
"(",
"f'skipping shadow loggers for node_b: {get_target_type_str(node_b, gm_b)}'",
"+",
"f', start_node_a: {get_target_type_str(subgraph_a.start_node, gm_a)}'",
"+",
"', unknown input qparams'",
")",
"env_c",
"[",
"node_b",
".",
"name",
"]",
"=",
"graph_c",
".",
"node_copy",
"(",
"node_b",
",",
"load_arg",
")",
"continue",
"fqn_base_a",
"=",
"_maybe_get_fqn",
"(",
"subgraph_a",
".",
"base_op_node",
",",
"gm_a",
")",
"fqn_base_b",
"=",
"_maybe_get_fqn",
"(",
"subgraph_b",
".",
"base_op_node",
",",
"gm_b",
")",
"if",
"node_b_is_start_node",
":",
"# if necessary, log the input of node_c",
"if",
"should_log_inputs",
":",
"if",
"isinstance",
"(",
"node_b",
".",
"args",
"[",
"0",
"]",
",",
"Node",
")",
":",
"prev_node_c",
"=",
"env_c",
"[",
"node_b",
".",
"args",
"[",
"0",
"]",
".",
"name",
"]",
"env_c",
"[",
"prev_node_c",
".",
"name",
"]",
"=",
"_insert_logger_after_node",
"(",
"prev_node_c",
",",
"gm_b",
",",
"logger_cls",
",",
"'_ns_logger_b_inp_'",
",",
"node_b",
".",
"name",
",",
"name_b",
",",
"ref_name",
",",
"ref_node_type_b",
",",
"NSSingleResultValuesType",
".",
"NODE_INPUT",
".",
"value",
",",
"index_within_arg",
"=",
"0",
",",
"index_of_arg",
"=",
"0",
",",
"fqn",
"=",
"fqn_base_b",
")",
"elif",
"isinstance",
"(",
"node_b",
".",
"args",
"[",
"0",
"]",
",",
"list",
")",
":",
"# first, save the prev_node instances, because they",
"# will be overwritten in the env after the first logger",
"# is added",
"prev_node_c_list",
"=",
"[",
"env_c",
"[",
"arg",
".",
"name",
"]",
"for",
"arg",
"in",
"node_b",
".",
"args",
"[",
"0",
"]",
"]",
"for",
"arg_idx",
",",
"arg",
"in",
"enumerate",
"(",
"node_b",
".",
"args",
"[",
"0",
"]",
")",
":",
"prev_node_c",
"=",
"prev_node_c_list",
"[",
"arg_idx",
"]",
"env_c",
"[",
"prev_node_c",
".",
"name",
"]",
"=",
"_insert_logger_after_node",
"(",
"prev_node_c",
",",
"gm_b",
",",
"logger_cls",
",",
"'_ns_logger_b_inp_'",
",",
"node_b",
".",
"name",
",",
"name_b",
",",
"ref_name",
",",
"ref_node_type_b",
",",
"NSSingleResultValuesType",
".",
"NODE_INPUT",
".",
"value",
",",
"index_within_arg",
"=",
"arg_idx",
",",
"index_of_arg",
"=",
"0",
",",
"fqn",
"=",
"fqn_base_b",
")",
"else",
":",
"# logging of inputs which are not lists is not supported yet",
"raise",
"AssertionError",
"(",
"f\"type {type(node_b.args[0])} is not handled yet\"",
")",
"# subgraph so far:",
"#",
"# (prev_node_c)+ -> (logger_c_input)?",
"# Note: this if statement is always True, spelling it out to clarify code",
"# intent.",
"if",
"node_b_is_start_node",
"or",
"node_b_is_end_node",
":",
"# ensure env_c is populated with base node",
"env_c",
"[",
"node_b",
".",
"name",
"]",
"=",
"graph_c",
".",
"node_copy",
"(",
"node_b",
",",
"load_arg",
")",
"node_c",
"=",
"env_c",
"[",
"node_b",
".",
"name",
"]",
"# after this point,",
"#",
"# node_a is the original node from graph_a, with parent module gm_a",
"# node_b is the original node from graph_b, with parent module gm_b",
"# node_c is the copy of node_b in graph_c",
"#",
"# subgraph so far:",
"#",
"# (prev_node_c)+ -> (logger_c_input)? -> node_start_c",
"if",
"node_b_is_start_node",
":",
"# cast dtype from the dtype of node_c's input to the dtype of",
"# node_a's input (dequant, etc)",
"prev_node_c",
"=",
"node_c",
".",
"args",
"[",
"0",
"]",
"if",
"should_log_inputs",
":",
"# skip the input logger when inserting a dtype cast",
"if",
"isinstance",
"(",
"prev_node_c",
",",
"Node",
")",
":",
"prev_node_c",
"=",
"prev_node_c",
".",
"args",
"[",
"0",
"]",
"elif",
"isinstance",
"(",
"prev_node_c",
",",
"list",
")",
":",
"prev_node_c",
"=",
"[",
"arg",
".",
"args",
"[",
"0",
"]",
"for",
"arg",
"in",
"prev_node_c",
"]",
"dtype_cast_node",
"=",
"_insert_dtype_cast_after_node",
"(",
"subgraph_a",
".",
"start_node",
",",
"node_c",
",",
"prev_node_c",
",",
"gm_a",
",",
"gm_b",
",",
"graph_c",
",",
"node_b",
".",
"name",
"+",
"'_dtype_cast_'",
",",
"logger_cls",
",",
"node_type_to_io_type_map",
")",
"# note: not inserting to env_c because all nodes which use the dtype",
"# casts are copied from graph_a",
"#",
"# subgraph so far:",
"#",
"# (dtype_cast_node)+",
"# /",
"# (prev_node_c)+ -> (logger_c_input)? -> node_start_c",
"# if input logging is enabled, log the input to the subgraph",
"if",
"should_log_inputs",
":",
"# TODO: explain this",
"ref_node_name",
"=",
"''",
"if",
"isinstance",
"(",
"dtype_cast_node",
",",
"Node",
")",
":",
"dtype_cast_node",
"=",
"_insert_logger_after_node",
"(",
"dtype_cast_node",
",",
"gm_b",
",",
"logger_cls",
",",
"'_ns_logger_a_inp_'",
",",
"ref_node_name",
",",
"name_a",
",",
"ref_name",
",",
"ref_node_type_a",
",",
"NSSingleResultValuesType",
".",
"NODE_INPUT",
".",
"value",
",",
"index_within_arg",
"=",
"0",
",",
"index_of_arg",
"=",
"0",
",",
"fqn",
"=",
"fqn_base_a",
")",
"input_logger",
":",
"Union",
"[",
"Node",
",",
"List",
"[",
"Node",
"]",
"]",
"=",
"dtype_cast_node",
"else",
":",
"assert",
"isinstance",
"(",
"dtype_cast_node",
",",
"list",
")",
"new_loggers",
"=",
"[",
"]",
"for",
"dtype_cast_idx",
",",
"dtype_cast_node_inner",
"in",
"enumerate",
"(",
"dtype_cast_node",
")",
":",
"dtype_cast_logger",
"=",
"_insert_logger_after_node",
"(",
"dtype_cast_node_inner",
",",
"gm_b",
",",
"logger_cls",
",",
"'_ns_logger_a_inp_'",
",",
"ref_node_name",
",",
"name_a",
",",
"ref_name",
",",
"ref_node_type_a",
",",
"NSSingleResultValuesType",
".",
"NODE_INPUT",
".",
"value",
",",
"index_within_arg",
"=",
"dtype_cast_idx",
",",
"index_of_arg",
"=",
"0",
",",
"fqn",
"=",
"fqn_base_a",
")",
"new_loggers",
".",
"append",
"(",
"dtype_cast_logger",
")",
"dtype_cast_node",
"=",
"new_loggers",
"input_logger",
"=",
"dtype_cast_node",
"# subgraph so far:",
"#",
"# (dtype_cast_node)+ -> (logger_a_input)?",
"# /",
"# prev_node_c -> (logger_c_input)? -> node_start_c",
"# hook up the new mod_a copy to be in the graph, receiving the",
"# same inputs as mod_b does, with dtype cast to match a",
"# Some ops, such as LSTMs, have two non-param inputs. If we have",
"# such an op, pass the second param as well. Note: dtype casting",
"# for the second param is not implemented yet, it can be added",
"# later if there is a use case.",
"node_c_second_non_param_arg",
"=",
"None",
"num_non_param_args_node_a",
"=",
"get_number_of_non_param_args",
"(",
"subgraph_a",
".",
"start_node",
",",
"gm_a",
")",
"if",
"num_non_param_args_node_a",
"==",
"2",
":",
"node_c_second_non_param_arg",
"=",
"node_c",
".",
"args",
"[",
"1",
"]",
"node_a_shadows_c",
"=",
"_insert_copy_of_subgraph_a_after_input_node_c",
"(",
"dtype_cast_node",
",",
"node_c_second_non_param_arg",
",",
"subgraph_a",
",",
"gm_a",
",",
"gm_b",
",",
"node_c",
".",
"name",
"+",
"'_shadow_copy_'",
")",
"env_c",
"[",
"node_a_shadows_c",
".",
"name",
"]",
"=",
"node_a_shadows_c",
"# subgraph so far:",
"#",
"# dtype_cast_node -> (logger_a_input)? -> subgraph_a_copy(args/kwargs not shown)",
"# /",
"# (prev_node_c)+ -> (logger_c_input)? -> node_start_c",
"if",
"should_log_inputs",
":",
"# When we created the input logger, we left the ref_node_name",
"# as an empty string, because the subgraph copy did not exist",
"# yet. Now that the subgraph copy exists, we modify this name",
"# to its true value.",
"# Note: the alternative to this is to create the input logger",
"# after creating the subgraph, which is slightly more",
"# complicated. This is the lesser of two evils.",
"# input_logger = env_c[dtype_cast_node.name]",
"# Find the first node in the subgraph",
"cur_node",
"=",
"node_a_shadows_c",
"while",
"cur_node",
".",
"args",
"[",
"0",
"]",
"!=",
"input_logger",
":",
"cur_node",
"=",
"cur_node",
".",
"args",
"[",
"0",
"]",
"# type: ignore[assignment]",
"if",
"isinstance",
"(",
"input_logger",
",",
"Node",
")",
":",
"input_logger_mod",
"=",
"getattr",
"(",
"gm_b",
",",
"input_logger",
".",
"name",
")",
"input_logger_mod",
".",
"ref_node_name",
"=",
"cur_node",
".",
"name",
"else",
":",
"assert",
"isinstance",
"(",
"input_logger",
",",
"list",
")",
"for",
"input_logger_inner",
"in",
"input_logger",
":",
"input_logger_mod",
"=",
"getattr",
"(",
"gm_b",
",",
"input_logger_inner",
".",
"name",
")",
"input_logger_mod",
".",
"ref_node_name",
"=",
"cur_node",
".",
"name",
"# hook up a logger to the mod_a copy",
"env_c",
"[",
"node_a_shadows_c",
".",
"name",
"]",
"=",
"_insert_logger_after_node",
"(",
"env_c",
"[",
"node_a_shadows_c",
".",
"name",
"]",
",",
"gm_b",
",",
"logger_cls",
",",
"'_ns_logger_a_'",
",",
"node_a_shadows_c",
".",
"name",
",",
"name_a",
",",
"ref_name",
",",
"ref_node_type_a",
",",
"NSSingleResultValuesType",
".",
"NODE_OUTPUT",
".",
"value",
",",
"index_within_arg",
"=",
"0",
",",
"index_of_arg",
"=",
"0",
",",
"fqn",
"=",
"fqn_base_a",
")",
"# subgraph so far:",
"#",
"# dtype_cast_node -> (logger_a_input)? -> subgraph_a_copy -> logger_a",
"# /",
"# (prev_node_c)+ -> (logger_c_input)? -> node_start_c",
"if",
"node_b_is_end_node",
":",
"# hook up a logger to the mod_b copy",
"env_c",
"[",
"node_b",
".",
"name",
"]",
"=",
"_insert_logger_after_node",
"(",
"env_c",
"[",
"node_b",
".",
"name",
"]",
",",
"gm_b",
",",
"logger_cls",
",",
"'_ns_logger_b_'",
",",
"node_b",
".",
"name",
",",
"name_b",
",",
"ref_name",
",",
"ref_node_type_b",
",",
"NSSingleResultValuesType",
".",
"NODE_OUTPUT",
".",
"value",
",",
"index_within_arg",
"=",
"0",
",",
"index_of_arg",
"=",
"0",
",",
"fqn",
"=",
"fqn_base_b",
")",
"# subgraph so far:",
"#",
"# dtype_cast_node -> (logger_a_input)? -> subgraph_a_copy -> logger_a",
"# /",
"# (prev_node_c+) -> (logger_c_input)? -> node_start_c -> ... -> node_end_c -> logger_c",
"#",
"# Note: node_start_c may be the same node as node_end_c, or they",
"# may have nodes inbetween.",
"else",
":",
"env_c",
"[",
"node_b",
".",
"name",
"]",
"=",
"graph_c",
".",
"node_copy",
"(",
"node_b",
",",
"load_arg",
")",
"gm_c",
"=",
"GraphModule",
"(",
"gm_b",
",",
"graph_c",
")",
"return",
"gm_c"
] | https://github.com/pytorch/pytorch/blob/7176c92687d3cc847cc046bf002269c6949a21c2/torch/ao/ns/fx/graph_passes.py#L514-L819 |
|
aws/lumberyard | f85344403c1c2e77ec8c75deb2c116e97b713217 | dev/Gems/CloudGemMetric/v1/AWS/common-code/Lib/pandas/core/indexes/multi.py | python | MultiIndex._hashed_values | (self) | return hash_tuples(self) | return a uint64 ndarray of my hashed values | return a uint64 ndarray of my hashed values | [
"return",
"a",
"uint64",
"ndarray",
"of",
"my",
"hashed",
"values"
] | def _hashed_values(self):
""" return a uint64 ndarray of my hashed values """
return hash_tuples(self) | [
"def",
"_hashed_values",
"(",
"self",
")",
":",
"return",
"hash_tuples",
"(",
"self",
")"
] | https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Gems/CloudGemMetric/v1/AWS/common-code/Lib/pandas/core/indexes/multi.py#L1408-L1410 |
|
bigartm/bigartm | 47e37f982de87aa67bfd475ff1f39da696b181b3 | utils/cpplint.py | python | FileInfo.NoExtension | (self) | return '/'.join(self.Split()[0:2]) | File has no source file extension. | File has no source file extension. | [
"File",
"has",
"no",
"source",
"file",
"extension",
"."
] | def NoExtension(self):
"""File has no source file extension."""
return '/'.join(self.Split()[0:2]) | [
"def",
"NoExtension",
"(",
"self",
")",
":",
"return",
"'/'",
".",
"join",
"(",
"self",
".",
"Split",
"(",
")",
"[",
"0",
":",
"2",
"]",
")"
] | https://github.com/bigartm/bigartm/blob/47e37f982de87aa67bfd475ff1f39da696b181b3/utils/cpplint.py#L1055-L1057 |
|
neoml-lib/neoml | a0d370fba05269a1b2258cef126f77bbd2054a3e | NeoML/Python/neoml/Dnn/Dnn.py | python | Dnn.initializer | (self, new_initializer) | Sets the initializer that will fill in the weight values
before training starts. | Sets the initializer that will fill in the weight values
before training starts. | [
"Sets",
"the",
"initializer",
"that",
"will",
"fill",
"in",
"the",
"weight",
"values",
"before",
"training",
"starts",
"."
] | def initializer(self, new_initializer):
"""Sets the initializer that will fill in the weight values
before training starts.
"""
self.set_initializer(new_initializer._internal) | [
"def",
"initializer",
"(",
"self",
",",
"new_initializer",
")",
":",
"self",
".",
"set_initializer",
"(",
"new_initializer",
".",
"_internal",
")"
] | https://github.com/neoml-lib/neoml/blob/a0d370fba05269a1b2258cef126f77bbd2054a3e/NeoML/Python/neoml/Dnn/Dnn.py#L106-L110 |
||
llvm/llvm-project | ffa6262cb4e2a335d26416fad39a581b4f98c5f4 | llvm/utils/lit/lit/LitConfig.py | python | LitConfig.maxIndividualTestTimeIsSupported | (self) | return lit.util.killProcessAndChildrenIsSupported() | Returns a tuple (<supported> , <error message>)
where
`<supported>` is True if setting maxIndividualTestTime is supported
on the current host, returns False otherwise.
`<error message>` is an empty string if `<supported>` is True,
otherwise is contains a string describing why setting
maxIndividualTestTime is not supported. | Returns a tuple (<supported> , <error message>)
where
`<supported>` is True if setting maxIndividualTestTime is supported
on the current host, returns False otherwise.
`<error message>` is an empty string if `<supported>` is True,
otherwise is contains a string describing why setting
maxIndividualTestTime is not supported. | [
"Returns",
"a",
"tuple",
"(",
"<supported",
">",
"<error",
"message",
">",
")",
"where",
"<supported",
">",
"is",
"True",
"if",
"setting",
"maxIndividualTestTime",
"is",
"supported",
"on",
"the",
"current",
"host",
"returns",
"False",
"otherwise",
".",
"<error",
"message",
">",
"is",
"an",
"empty",
"string",
"if",
"<supported",
">",
"is",
"True",
"otherwise",
"is",
"contains",
"a",
"string",
"describing",
"why",
"setting",
"maxIndividualTestTime",
"is",
"not",
"supported",
"."
] | def maxIndividualTestTimeIsSupported(self):
"""
Returns a tuple (<supported> , <error message>)
where
`<supported>` is True if setting maxIndividualTestTime is supported
on the current host, returns False otherwise.
`<error message>` is an empty string if `<supported>` is True,
otherwise is contains a string describing why setting
maxIndividualTestTime is not supported.
"""
return lit.util.killProcessAndChildrenIsSupported() | [
"def",
"maxIndividualTestTimeIsSupported",
"(",
"self",
")",
":",
"return",
"lit",
".",
"util",
".",
"killProcessAndChildrenIsSupported",
"(",
")"
] | https://github.com/llvm/llvm-project/blob/ffa6262cb4e2a335d26416fad39a581b4f98c5f4/llvm/utils/lit/lit/LitConfig.py#L79-L89 |
|
miyosuda/TensorFlowAndroidMNIST | 7b5a4603d2780a8a2834575706e9001977524007 | jni-build/jni/include/tensorflow/python/ops/control_flow_ops.py | python | CondContext.BuildCondBranch | (self, fn) | return original_r, result | Add the subgraph defined by fn() to the graph. | Add the subgraph defined by fn() to the graph. | [
"Add",
"the",
"subgraph",
"defined",
"by",
"fn",
"()",
"to",
"the",
"graph",
"."
] | def BuildCondBranch(self, fn):
"""Add the subgraph defined by fn() to the graph."""
r = fn()
original_r = r
result = []
if r is not None:
if not isinstance(r, list) and not isinstance(r, _basetuple):
r = [r]
original_r = [original_r]
r = _convert_tensorarrays_to_flows(r)
for v in r:
real_v = v
if isinstance(v, ops.Operation):
# Use pivot as the proxy for this op.
real_v = with_dependencies([v], self._pivot)
elif v.name not in self._values:
# Handle the special case of lambda: x
self._values.add(v.name)
if self._outer_context:
real_v = self._outer_context.AddValue(v)
self._values.add(real_v.name)
real_v = _SwitchRefOrTensor(real_v, self._pred)[self._branch]
self._external_values[v.name] = real_v
else:
external_v = self._external_values.get(v.name)
if external_v is not None:
real_v = external_v
result.append(real_v)
return original_r, result | [
"def",
"BuildCondBranch",
"(",
"self",
",",
"fn",
")",
":",
"r",
"=",
"fn",
"(",
")",
"original_r",
"=",
"r",
"result",
"=",
"[",
"]",
"if",
"r",
"is",
"not",
"None",
":",
"if",
"not",
"isinstance",
"(",
"r",
",",
"list",
")",
"and",
"not",
"isinstance",
"(",
"r",
",",
"_basetuple",
")",
":",
"r",
"=",
"[",
"r",
"]",
"original_r",
"=",
"[",
"original_r",
"]",
"r",
"=",
"_convert_tensorarrays_to_flows",
"(",
"r",
")",
"for",
"v",
"in",
"r",
":",
"real_v",
"=",
"v",
"if",
"isinstance",
"(",
"v",
",",
"ops",
".",
"Operation",
")",
":",
"# Use pivot as the proxy for this op.",
"real_v",
"=",
"with_dependencies",
"(",
"[",
"v",
"]",
",",
"self",
".",
"_pivot",
")",
"elif",
"v",
".",
"name",
"not",
"in",
"self",
".",
"_values",
":",
"# Handle the special case of lambda: x",
"self",
".",
"_values",
".",
"add",
"(",
"v",
".",
"name",
")",
"if",
"self",
".",
"_outer_context",
":",
"real_v",
"=",
"self",
".",
"_outer_context",
".",
"AddValue",
"(",
"v",
")",
"self",
".",
"_values",
".",
"add",
"(",
"real_v",
".",
"name",
")",
"real_v",
"=",
"_SwitchRefOrTensor",
"(",
"real_v",
",",
"self",
".",
"_pred",
")",
"[",
"self",
".",
"_branch",
"]",
"self",
".",
"_external_values",
"[",
"v",
".",
"name",
"]",
"=",
"real_v",
"else",
":",
"external_v",
"=",
"self",
".",
"_external_values",
".",
"get",
"(",
"v",
".",
"name",
")",
"if",
"external_v",
"is",
"not",
"None",
":",
"real_v",
"=",
"external_v",
"result",
".",
"append",
"(",
"real_v",
")",
"return",
"original_r",
",",
"result"
] | https://github.com/miyosuda/TensorFlowAndroidMNIST/blob/7b5a4603d2780a8a2834575706e9001977524007/jni-build/jni/include/tensorflow/python/ops/control_flow_ops.py#L1244-L1272 |
|
Tencent/CMONGO | c40380caa14e05509f46993aa8b8da966b09b0b5 | buildscripts/packager.py | python | Distro.repo_os_version | (self, build_os) | Return an OS version suitable for package repo directory
naming - e.g. 5, 6 or 7 for redhat/centos, "precise," "wheezy," etc.
for Ubuntu/Debian, 11 for suse, "2013.03" for amazon | Return an OS version suitable for package repo directory
naming - e.g. 5, 6 or 7 for redhat/centos, "precise," "wheezy," etc.
for Ubuntu/Debian, 11 for suse, "2013.03" for amazon | [
"Return",
"an",
"OS",
"version",
"suitable",
"for",
"package",
"repo",
"directory",
"naming",
"-",
"e",
".",
"g",
".",
"5",
"6",
"or",
"7",
"for",
"redhat",
"/",
"centos",
"precise",
"wheezy",
"etc",
".",
"for",
"Ubuntu",
"/",
"Debian",
"11",
"for",
"suse",
"2013",
".",
"03",
"for",
"amazon"
] | def repo_os_version(self, build_os):
"""Return an OS version suitable for package repo directory
naming - e.g. 5, 6 or 7 for redhat/centos, "precise," "wheezy," etc.
for Ubuntu/Debian, 11 for suse, "2013.03" for amazon"""
if self.n == 'suse':
return re.sub(r'^suse(\d+)$', r'\1', build_os)
if self.n == 'redhat':
return re.sub(r'^rhel(\d).*$', r'\1', build_os)
if self.n == 'amazon':
return "2013.03"
elif self.n == 'ubuntu':
if build_os == 'ubuntu1204':
return "precise"
elif build_os == 'ubuntu1404':
return "trusty"
elif build_os == 'ubuntu1604':
return "xenial"
else:
raise Exception("unsupported build_os: %s" % build_os)
elif self.n == 'debian':
if build_os == 'debian71':
return 'wheezy'
elif build_os == 'debian81':
return 'jessie'
else:
raise Exception("unsupported build_os: %s" % build_os)
else:
raise Exception("unsupported distro: %s" % self.n) | [
"def",
"repo_os_version",
"(",
"self",
",",
"build_os",
")",
":",
"if",
"self",
".",
"n",
"==",
"'suse'",
":",
"return",
"re",
".",
"sub",
"(",
"r'^suse(\\d+)$'",
",",
"r'\\1'",
",",
"build_os",
")",
"if",
"self",
".",
"n",
"==",
"'redhat'",
":",
"return",
"re",
".",
"sub",
"(",
"r'^rhel(\\d).*$'",
",",
"r'\\1'",
",",
"build_os",
")",
"if",
"self",
".",
"n",
"==",
"'amazon'",
":",
"return",
"\"2013.03\"",
"elif",
"self",
".",
"n",
"==",
"'ubuntu'",
":",
"if",
"build_os",
"==",
"'ubuntu1204'",
":",
"return",
"\"precise\"",
"elif",
"build_os",
"==",
"'ubuntu1404'",
":",
"return",
"\"trusty\"",
"elif",
"build_os",
"==",
"'ubuntu1604'",
":",
"return",
"\"xenial\"",
"else",
":",
"raise",
"Exception",
"(",
"\"unsupported build_os: %s\"",
"%",
"build_os",
")",
"elif",
"self",
".",
"n",
"==",
"'debian'",
":",
"if",
"build_os",
"==",
"'debian71'",
":",
"return",
"'wheezy'",
"elif",
"build_os",
"==",
"'debian81'",
":",
"return",
"'jessie'",
"else",
":",
"raise",
"Exception",
"(",
"\"unsupported build_os: %s\"",
"%",
"build_os",
")",
"else",
":",
"raise",
"Exception",
"(",
"\"unsupported distro: %s\"",
"%",
"self",
".",
"n",
")"
] | https://github.com/Tencent/CMONGO/blob/c40380caa14e05509f46993aa8b8da966b09b0b5/buildscripts/packager.py#L197-L224 |
||
infinit/memo | 3a8394d0f647efe03ccb8bfe885a7279cb8be8a6 | elle/drake/src/drake/__init__.py | python | BaseNode.drake_type | (self) | return '%s.%s' % (self.__module__, self.__name__) | The qualified name of this type. | The qualified name of this type. | [
"The",
"qualified",
"name",
"of",
"this",
"type",
"."
] | def drake_type(self):
"""The qualified name of this type."""
return '%s.%s' % (self.__module__, self.__name__) | [
"def",
"drake_type",
"(",
"self",
")",
":",
"return",
"'%s.%s'",
"%",
"(",
"self",
".",
"__module__",
",",
"self",
".",
"__name__",
")"
] | https://github.com/infinit/memo/blob/3a8394d0f647efe03ccb8bfe885a7279cb8be8a6/elle/drake/src/drake/__init__.py#L1391-L1393 |
|
benoitsteiner/tensorflow-opencl | cb7cb40a57fde5cfd4731bc551e82a1e2fef43a5 | tensorflow/contrib/distributions/python/ops/binomial.py | python | Binomial.probs | (self) | return self._probs | Probability of drawing a `1`. | Probability of drawing a `1`. | [
"Probability",
"of",
"drawing",
"a",
"1",
"."
] | def probs(self):
"""Probability of drawing a `1`."""
return self._probs | [
"def",
"probs",
"(",
"self",
")",
":",
"return",
"self",
".",
"_probs"
] | https://github.com/benoitsteiner/tensorflow-opencl/blob/cb7cb40a57fde5cfd4731bc551e82a1e2fef43a5/tensorflow/contrib/distributions/python/ops/binomial.py#L198-L200 |
|
Kitware/kwiver | 7ed70308905698b6e88d27ae3dc028c9b016ca0a | python/kwiver/sprokit/processes/kwiver_process.py | python | KwiverProcess.config_value_using_trait | (self, name) | return self.config_value(ct.name) | Get value from config using trait.
An exception will be thrown if the config trait has not been registered
with the process.
:param name: Name of the configuration trait. | Get value from config using trait.
An exception will be thrown if the config trait has not been registered
with the process. | [
"Get",
"value",
"from",
"config",
"using",
"trait",
".",
"An",
"exception",
"will",
"be",
"thrown",
"if",
"the",
"config",
"trait",
"has",
"not",
"been",
"registered",
"with",
"the",
"process",
"."
] | def config_value_using_trait(self, name):
"""
Get value from config using trait.
An exception will be thrown if the config trait has not been registered
with the process.
:param name: Name of the configuration trait.
"""
ct = self._config_trait_set.get(name, None)
if ct is None:
raise ValueError('config trait name \"%s\" not registered' % name)
return self.config_value(ct.name) | [
"def",
"config_value_using_trait",
"(",
"self",
",",
"name",
")",
":",
"ct",
"=",
"self",
".",
"_config_trait_set",
".",
"get",
"(",
"name",
",",
"None",
")",
"if",
"ct",
"is",
"None",
":",
"raise",
"ValueError",
"(",
"'config trait name \\\"%s\\\" not registered'",
"%",
"name",
")",
"return",
"self",
".",
"config_value",
"(",
"ct",
".",
"name",
")"
] | https://github.com/Kitware/kwiver/blob/7ed70308905698b6e88d27ae3dc028c9b016ca0a/python/kwiver/sprokit/processes/kwiver_process.py#L369-L382 |
|
luliyucoordinate/Leetcode | 96afcdc54807d1d184e881a075d1dbf3371e31fb | src/0019-Remove-Nth-Node-From-End-of-List/0019.py | python | Solution.removeNthFromEnd | (self, head, n) | return h.next | :type head: ListNode
:type n: int
:rtype: ListNode | :type head: ListNode
:type n: int
:rtype: ListNode | [
":",
"type",
"head",
":",
"ListNode",
":",
"type",
"n",
":",
"int",
":",
"rtype",
":",
"ListNode"
] | def removeNthFromEnd(self, head, n):
"""
:type head: ListNode
:type n: int
:rtype: ListNode
"""
h = ListNode(-1)
h.next = head
p, q = h, h
for _ in range(n + 1):
assert (q)
q = q.next
while q != None:
p = p.next
q = q.next
p.next = p.next.next
return h.next | [
"def",
"removeNthFromEnd",
"(",
"self",
",",
"head",
",",
"n",
")",
":",
"h",
"=",
"ListNode",
"(",
"-",
"1",
")",
"h",
".",
"next",
"=",
"head",
"p",
",",
"q",
"=",
"h",
",",
"h",
"for",
"_",
"in",
"range",
"(",
"n",
"+",
"1",
")",
":",
"assert",
"(",
"q",
")",
"q",
"=",
"q",
".",
"next",
"while",
"q",
"!=",
"None",
":",
"p",
"=",
"p",
".",
"next",
"q",
"=",
"q",
".",
"next",
"p",
".",
"next",
"=",
"p",
".",
"next",
".",
"next",
"return",
"h",
".",
"next"
] | https://github.com/luliyucoordinate/Leetcode/blob/96afcdc54807d1d184e881a075d1dbf3371e31fb/src/0019-Remove-Nth-Node-From-End-of-List/0019.py#L8-L27 |
|
stulp/dmpbbo | ca900e3b851d25faaf59ea296650370c70ed7d0f | python/bbo/updaters.py | python | costsToWeights | (costs, weighting_method, eliteness) | return weights | Convert costs to weights using different weighting methods.
\param[in] costs A vector of costs.
\param[in] weighting_method The weighting method ('PI-BB','CMA-ES','CEM')
\param[in] eliteness The eliteness parameter (h in PI-BB, mu in CMA-ES)
\return A vector of weights (they always sum to 1). | Convert costs to weights using different weighting methods.
\param[in] costs A vector of costs.
\param[in] weighting_method The weighting method ('PI-BB','CMA-ES','CEM')
\param[in] eliteness The eliteness parameter (h in PI-BB, mu in CMA-ES)
\return A vector of weights (they always sum to 1). | [
"Convert",
"costs",
"to",
"weights",
"using",
"different",
"weighting",
"methods",
".",
"\\",
"param",
"[",
"in",
"]",
"costs",
"A",
"vector",
"of",
"costs",
".",
"\\",
"param",
"[",
"in",
"]",
"weighting_method",
"The",
"weighting",
"method",
"(",
"PI",
"-",
"BB",
"CMA",
"-",
"ES",
"CEM",
")",
"\\",
"param",
"[",
"in",
"]",
"eliteness",
"The",
"eliteness",
"parameter",
"(",
"h",
"in",
"PI",
"-",
"BB",
"mu",
"in",
"CMA",
"-",
"ES",
")",
"\\",
"return",
"A",
"vector",
"of",
"weights",
"(",
"they",
"always",
"sum",
"to",
"1",
")",
"."
] | def costsToWeights(costs, weighting_method, eliteness):
""" Convert costs to weights using different weighting methods.
\param[in] costs A vector of costs.
\param[in] weighting_method The weighting method ('PI-BB','CMA-ES','CEM')
\param[in] eliteness The eliteness parameter (h in PI-BB, mu in CMA-ES)
\return A vector of weights (they always sum to 1).
"""
# Costs can be a 2D array or a list of lists. In this case, the first
# column is the sum of the other columns (which contain the different cost
# components). In this case, we should use only the first column.
costs = np.asarray([np.atleast_1d(x)[0] for x in costs])
#np.set_printoptions(precision=4, suppress=True)
if weighting_method == 'PI-BB':
# PI^2 style weighting: continuous, cost exponention
h = eliteness # In PI^2, eliteness parameter is known as "h"
costs_range = max(costs)-min(costs)
if costs_range==0:
weights = np.full(costs.shape,1.0)
else:
costs_norm = np.asarray([-h*(x-min(costs))/costs_range for x in costs])
weights = np.exp(costs_norm)
elif weighting_method=='CEM' or weighting_method=='CMA-ES':
# CEM/CMA-ES style weights: rank-based, uses defaults
mu = eliteness # In CMA-ES, eliteness parameter is known as "mu"
indices = np.argsort(costs)
weights = np.full(costs.size,0.0)
if weighting_method=='CEM':
# CEM
weights[indices[0:mu]] = 1.0/mu
else:
# CMA-ES
for ii in range(mu):
weights[indices[ii]] = np.log(mu+0.5)-np.log(ii+1)
else:
print("WARNING: Unknown weighting method '", weighting_method, "'. Calling with PI-BB weighting.");
return costsToWeights(costs, 'PI-BB', eliteness);
#// Relative standard deviation of total costs
#double mean = weights.mean();
#double std = sqrt((weights.array()-mean).pow(2).mean());
#double rel_std = std/mean;
#if (rel_std<1e-10)
#{
# // Special case: all costs are the same
# // Set same weights for all.
# weights.fill(1);
#}
# Normalize weights
weights = weights/sum(weights)
return weights | [
"def",
"costsToWeights",
"(",
"costs",
",",
"weighting_method",
",",
"eliteness",
")",
":",
"# Costs can be a 2D array or a list of lists. In this case, the first",
"# column is the sum of the other columns (which contain the different cost",
"# components). In this case, we should use only the first column.",
"costs",
"=",
"np",
".",
"asarray",
"(",
"[",
"np",
".",
"atleast_1d",
"(",
"x",
")",
"[",
"0",
"]",
"for",
"x",
"in",
"costs",
"]",
")",
"#np.set_printoptions(precision=4, suppress=True)",
"if",
"weighting_method",
"==",
"'PI-BB'",
":",
"# PI^2 style weighting: continuous, cost exponention",
"h",
"=",
"eliteness",
"# In PI^2, eliteness parameter is known as \"h\"",
"costs_range",
"=",
"max",
"(",
"costs",
")",
"-",
"min",
"(",
"costs",
")",
"if",
"costs_range",
"==",
"0",
":",
"weights",
"=",
"np",
".",
"full",
"(",
"costs",
".",
"shape",
",",
"1.0",
")",
"else",
":",
"costs_norm",
"=",
"np",
".",
"asarray",
"(",
"[",
"-",
"h",
"*",
"(",
"x",
"-",
"min",
"(",
"costs",
")",
")",
"/",
"costs_range",
"for",
"x",
"in",
"costs",
"]",
")",
"weights",
"=",
"np",
".",
"exp",
"(",
"costs_norm",
")",
"elif",
"weighting_method",
"==",
"'CEM'",
"or",
"weighting_method",
"==",
"'CMA-ES'",
":",
"# CEM/CMA-ES style weights: rank-based, uses defaults",
"mu",
"=",
"eliteness",
"# In CMA-ES, eliteness parameter is known as \"mu\"",
"indices",
"=",
"np",
".",
"argsort",
"(",
"costs",
")",
"weights",
"=",
"np",
".",
"full",
"(",
"costs",
".",
"size",
",",
"0.0",
")",
"if",
"weighting_method",
"==",
"'CEM'",
":",
"# CEM",
"weights",
"[",
"indices",
"[",
"0",
":",
"mu",
"]",
"]",
"=",
"1.0",
"/",
"mu",
"else",
":",
"# CMA-ES",
"for",
"ii",
"in",
"range",
"(",
"mu",
")",
":",
"weights",
"[",
"indices",
"[",
"ii",
"]",
"]",
"=",
"np",
".",
"log",
"(",
"mu",
"+",
"0.5",
")",
"-",
"np",
".",
"log",
"(",
"ii",
"+",
"1",
")",
"else",
":",
"print",
"(",
"\"WARNING: Unknown weighting method '\"",
",",
"weighting_method",
",",
"\"'. Calling with PI-BB weighting.\"",
")",
"return",
"costsToWeights",
"(",
"costs",
",",
"'PI-BB'",
",",
"eliteness",
")",
"#// Relative standard deviation of total costs",
"#double mean = weights.mean();",
"#double std = sqrt((weights.array()-mean).pow(2).mean());",
"#double rel_std = std/mean;",
"#if (rel_std<1e-10)",
"#{",
"# // Special case: all costs are the same",
"# // Set same weights for all.",
"# weights.fill(1);",
"#}",
"# Normalize weights",
"weights",
"=",
"weights",
"/",
"sum",
"(",
"weights",
")",
"return",
"weights"
] | https://github.com/stulp/dmpbbo/blob/ca900e3b851d25faaf59ea296650370c70ed7d0f/python/bbo/updaters.py#L199-L254 |
|
apple/turicreate | cce55aa5311300e3ce6af93cb45ba791fd1bdf49 | src/external/coremltools_wrap/coremltools/coremltools/converters/mil/frontend/torch/ops.py | python | convert_block | (context, block, inputs) | return outputs | Convert a block (sub-graph) to MIL. Conversion happens within a new
context frame.
Arguments:
context: A TranscriptionContext object to pull node inputs and
assign node outputs.
block: An InternalTorchIRBlock object.
inputs: List of Vars from the outer context that map to the block's
expected inputs. The number of inputs provided must match the
number expected by the block. | Convert a block (sub-graph) to MIL. Conversion happens within a new
context frame. | [
"Convert",
"a",
"block",
"(",
"sub",
"-",
"graph",
")",
"to",
"MIL",
".",
"Conversion",
"happens",
"within",
"a",
"new",
"context",
"frame",
"."
] | def convert_block(context, block, inputs):
"""Convert a block (sub-graph) to MIL. Conversion happens within a new
context frame.
Arguments:
context: A TranscriptionContext object to pull node inputs and
assign node outputs.
block: An InternalTorchIRBlock object.
inputs: List of Vars from the outer context that map to the block's
expected inputs. The number of inputs provided must match the
number expected by the block.
"""
assert len(block.inputs) == len(inputs)
# Start a new context frame.
context.push((block.inputs, inputs))
# Add the block ops.
convert_nodes(context, block)
# Collect the block outputs.
outputs = [context[outp] for outp in block.outputs]
# Return to the previous context frame.
context.pop()
return outputs | [
"def",
"convert_block",
"(",
"context",
",",
"block",
",",
"inputs",
")",
":",
"assert",
"len",
"(",
"block",
".",
"inputs",
")",
"==",
"len",
"(",
"inputs",
")",
"# Start a new context frame.",
"context",
".",
"push",
"(",
"(",
"block",
".",
"inputs",
",",
"inputs",
")",
")",
"# Add the block ops.",
"convert_nodes",
"(",
"context",
",",
"block",
")",
"# Collect the block outputs.",
"outputs",
"=",
"[",
"context",
"[",
"outp",
"]",
"for",
"outp",
"in",
"block",
".",
"outputs",
"]",
"# Return to the previous context frame.",
"context",
".",
"pop",
"(",
")",
"return",
"outputs"
] | https://github.com/apple/turicreate/blob/cce55aa5311300e3ce6af93cb45ba791fd1bdf49/src/external/coremltools_wrap/coremltools/coremltools/converters/mil/frontend/torch/ops.py#L62-L88 |
|
google/llvm-propeller | 45c226984fe8377ebfb2ad7713c680d652ba678d | clang/bindings/python/clang/cindex.py | python | Type.get_ref_qualifier | (self) | return RefQualifierKind.from_id(
conf.lib.clang_Type_getCXXRefQualifier(self)) | Retrieve the ref-qualifier of the type. | Retrieve the ref-qualifier of the type. | [
"Retrieve",
"the",
"ref",
"-",
"qualifier",
"of",
"the",
"type",
"."
] | def get_ref_qualifier(self):
"""
Retrieve the ref-qualifier of the type.
"""
return RefQualifierKind.from_id(
conf.lib.clang_Type_getCXXRefQualifier(self)) | [
"def",
"get_ref_qualifier",
"(",
"self",
")",
":",
"return",
"RefQualifierKind",
".",
"from_id",
"(",
"conf",
".",
"lib",
".",
"clang_Type_getCXXRefQualifier",
"(",
"self",
")",
")"
] | https://github.com/google/llvm-propeller/blob/45c226984fe8377ebfb2ad7713c680d652ba678d/clang/bindings/python/clang/cindex.py#L2396-L2401 |
|
MegEngine/MegEngine | ce9ad07a27ec909fb8db4dd67943d24ba98fb93a | imperative/python/megengine/functional/math.py | python | dot | (inp1: Tensor, inp2: Tensor) | return result | r"""Computes dot-product of two vectors ``inp1`` and ``inp2``.
inputs must be 1-dimensional or scalar. A scalar input is automatically broadcasted.
Refer to :func:`~.matmul` for more general usage.
Args:
inp1: first vector.
inp2: second vector.
Returns:
output value.
Examples:
.. testcode::
import numpy as np
from megengine import tensor
import megengine.functional as F
data1 = tensor(np.arange(0, 6, dtype=np.float32))
data2 = tensor(np.arange(0, 6, dtype=np.float32))
out = F.dot(data1, data2)
print(out.numpy())
Outputs:
.. testoutput::
55. | r"""Computes dot-product of two vectors ``inp1`` and ``inp2``.
inputs must be 1-dimensional or scalar. A scalar input is automatically broadcasted.
Refer to :func:`~.matmul` for more general usage. | [
"r",
"Computes",
"dot",
"-",
"product",
"of",
"two",
"vectors",
"inp1",
"and",
"inp2",
".",
"inputs",
"must",
"be",
"1",
"-",
"dimensional",
"or",
"scalar",
".",
"A",
"scalar",
"input",
"is",
"automatically",
"broadcasted",
".",
"Refer",
"to",
":",
"func",
":",
"~",
".",
"matmul",
"for",
"more",
"general",
"usage",
"."
] | def dot(inp1: Tensor, inp2: Tensor) -> Tensor:
r"""Computes dot-product of two vectors ``inp1`` and ``inp2``.
inputs must be 1-dimensional or scalar. A scalar input is automatically broadcasted.
Refer to :func:`~.matmul` for more general usage.
Args:
inp1: first vector.
inp2: second vector.
Returns:
output value.
Examples:
.. testcode::
import numpy as np
from megengine import tensor
import megengine.functional as F
data1 = tensor(np.arange(0, 6, dtype=np.float32))
data2 = tensor(np.arange(0, 6, dtype=np.float32))
out = F.dot(data1, data2)
print(out.numpy())
Outputs:
.. testoutput::
55.
"""
op = builtin.Dot()
assert (
inp1.ndim <= 1 and inp2.ndim <= 1
), "Input tensors for dot must be 1-dimensional or scalar"
(result,) = apply(op, inp1, inp2)
return result | [
"def",
"dot",
"(",
"inp1",
":",
"Tensor",
",",
"inp2",
":",
"Tensor",
")",
"->",
"Tensor",
":",
"op",
"=",
"builtin",
".",
"Dot",
"(",
")",
"assert",
"(",
"inp1",
".",
"ndim",
"<=",
"1",
"and",
"inp2",
".",
"ndim",
"<=",
"1",
")",
",",
"\"Input tensors for dot must be 1-dimensional or scalar\"",
"(",
"result",
",",
")",
"=",
"apply",
"(",
"op",
",",
"inp1",
",",
"inp2",
")",
"return",
"result"
] | https://github.com/MegEngine/MegEngine/blob/ce9ad07a27ec909fb8db4dd67943d24ba98fb93a/imperative/python/megengine/functional/math.py#L1116-L1152 |
|
microsoft/DirectXShaderCompiler | 8348ff8d9e0287610ba05d3a828e10af981a1c05 | tools/clang/bindings/python/clang/cindex.py | python | SourceLocation.offset | (self) | return self._get_instantiation()[3] | Get the file offset represented by this source location. | Get the file offset represented by this source location. | [
"Get",
"the",
"file",
"offset",
"represented",
"by",
"this",
"source",
"location",
"."
] | def offset(self):
"""Get the file offset represented by this source location."""
return self._get_instantiation()[3] | [
"def",
"offset",
"(",
"self",
")",
":",
"return",
"self",
".",
"_get_instantiation",
"(",
")",
"[",
"3",
"]"
] | https://github.com/microsoft/DirectXShaderCompiler/blob/8348ff8d9e0287610ba05d3a828e10af981a1c05/tools/clang/bindings/python/clang/cindex.py#L213-L215 |
|
aws/lumberyard | f85344403c1c2e77ec8c75deb2c116e97b713217 | dev/Tools/Python/3.7.10/mac/Python.framework/Versions/3.7/lib/python3.7/distutils/command/build_py.py | python | build_py.build_package_data | (self) | Copy data files into build directory | Copy data files into build directory | [
"Copy",
"data",
"files",
"into",
"build",
"directory"
] | def build_package_data(self):
"""Copy data files into build directory"""
lastdir = None
for package, src_dir, build_dir, filenames in self.data_files:
for filename in filenames:
target = os.path.join(build_dir, filename)
self.mkpath(os.path.dirname(target))
self.copy_file(os.path.join(src_dir, filename), target,
preserve_mode=False) | [
"def",
"build_package_data",
"(",
"self",
")",
":",
"lastdir",
"=",
"None",
"for",
"package",
",",
"src_dir",
",",
"build_dir",
",",
"filenames",
"in",
"self",
".",
"data_files",
":",
"for",
"filename",
"in",
"filenames",
":",
"target",
"=",
"os",
".",
"path",
".",
"join",
"(",
"build_dir",
",",
"filename",
")",
"self",
".",
"mkpath",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"target",
")",
")",
"self",
".",
"copy_file",
"(",
"os",
".",
"path",
".",
"join",
"(",
"src_dir",
",",
"filename",
")",
",",
"target",
",",
"preserve_mode",
"=",
"False",
")"
] | https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/mac/Python.framework/Versions/3.7/lib/python3.7/distutils/command/build_py.py#L134-L142 |
||
google/brunsli | e811197ab1ad8ddde3e3cf444548e42e2bdacf92 | contrib/py/jxl_library_patches/jxl_utils.py | python | is_jpegxl_recompressed_jpeg_file | (filename) | Returns True iff the given filename is a genuine JPEG-XL file. | Returns True iff the given filename is a genuine JPEG-XL file. | [
"Returns",
"True",
"iff",
"the",
"given",
"filename",
"is",
"a",
"genuine",
"JPEG",
"-",
"XL",
"file",
"."
] | def is_jpegxl_recompressed_jpeg_file(filename):
"""Returns True iff the given filename is a genuine JPEG-XL file."""
try:
with open(filename, 'rb') as h:
header = h.read(len(JPEGXL_RECOMPRESSED_JPEG_HEADER))
# Cf. https://arxiv.org/pdf/1908.03565.pdf, section 9.1,
# on recompressed-JPEG header.
return header == JPEGXL_RECOMPRESSED_JPEG_HEADER
except: # pylint:disable=bare-except
# If anything failed, this means that we cannot establish that the file
# has the expected header, so we return False.
return False | [
"def",
"is_jpegxl_recompressed_jpeg_file",
"(",
"filename",
")",
":",
"try",
":",
"with",
"open",
"(",
"filename",
",",
"'rb'",
")",
"as",
"h",
":",
"header",
"=",
"h",
".",
"read",
"(",
"len",
"(",
"JPEGXL_RECOMPRESSED_JPEG_HEADER",
")",
")",
"# Cf. https://arxiv.org/pdf/1908.03565.pdf, section 9.1,",
"# on recompressed-JPEG header.",
"return",
"header",
"==",
"JPEGXL_RECOMPRESSED_JPEG_HEADER",
"except",
":",
"# pylint:disable=bare-except",
"# If anything failed, this means that we cannot establish that the file",
"# has the expected header, so we return False.",
"return",
"False"
] | https://github.com/google/brunsli/blob/e811197ab1ad8ddde3e3cf444548e42e2bdacf92/contrib/py/jxl_library_patches/jxl_utils.py#L44-L55 |
||
sdhash/sdhash | b9eff63e4e5867e910f41fd69032bbb1c94a2a5e | sdhash-ui/cherrypy/process/servers.py | python | client_host | (server_host) | return server_host | Return the host on which a client can connect to the given listener. | Return the host on which a client can connect to the given listener. | [
"Return",
"the",
"host",
"on",
"which",
"a",
"client",
"can",
"connect",
"to",
"the",
"given",
"listener",
"."
] | def client_host(server_host):
"""Return the host on which a client can connect to the given listener."""
if server_host == '0.0.0.0':
# 0.0.0.0 is INADDR_ANY, which should answer on localhost.
return '127.0.0.1'
if server_host in ('::', '::0', '::0.0.0.0'):
# :: is IN6ADDR_ANY, which should answer on localhost.
# ::0 and ::0.0.0.0 are non-canonical but common ways to write IN6ADDR_ANY.
return '::1'
return server_host | [
"def",
"client_host",
"(",
"server_host",
")",
":",
"if",
"server_host",
"==",
"'0.0.0.0'",
":",
"# 0.0.0.0 is INADDR_ANY, which should answer on localhost.",
"return",
"'127.0.0.1'",
"if",
"server_host",
"in",
"(",
"'::'",
",",
"'::0'",
",",
"'::0.0.0.0'",
")",
":",
"# :: is IN6ADDR_ANY, which should answer on localhost.",
"# ::0 and ::0.0.0.0 are non-canonical but common ways to write IN6ADDR_ANY.",
"return",
"'::1'",
"return",
"server_host"
] | https://github.com/sdhash/sdhash/blob/b9eff63e4e5867e910f41fd69032bbb1c94a2a5e/sdhash-ui/cherrypy/process/servers.py#L340-L349 |
|
catboost/catboost | 167f64f237114a4d10b2b4ee42adb4569137debe | contrib/python/scipy/py2/scipy/signal/waveforms.py | python | unit_impulse | (shape, idx=None, dtype=float) | return out | Unit impulse signal (discrete delta function) or unit basis vector.
Parameters
----------
shape : int or tuple of int
Number of samples in the output (1-D), or a tuple that represents the
shape of the output (N-D).
idx : None or int or tuple of int or 'mid', optional
Index at which the value is 1. If None, defaults to the 0th element.
If ``idx='mid'``, the impulse will be centered at ``shape // 2`` in
all dimensions. If an int, the impulse will be at `idx` in all
dimensions.
dtype : data-type, optional
The desired data-type for the array, e.g., `numpy.int8`. Default is
`numpy.float64`.
Returns
-------
y : ndarray
Output array containing an impulse signal.
Notes
-----
The 1D case is also known as the Kronecker delta.
.. versionadded:: 0.19.0
Examples
--------
An impulse at the 0th element (:math:`\\delta[n]`):
>>> from scipy import signal
>>> signal.unit_impulse(8)
array([ 1., 0., 0., 0., 0., 0., 0., 0.])
Impulse offset by 2 samples (:math:`\\delta[n-2]`):
>>> signal.unit_impulse(7, 2)
array([ 0., 0., 1., 0., 0., 0., 0.])
2-dimensional impulse, centered:
>>> signal.unit_impulse((3, 3), 'mid')
array([[ 0., 0., 0.],
[ 0., 1., 0.],
[ 0., 0., 0.]])
Impulse at (2, 2), using broadcasting:
>>> signal.unit_impulse((4, 4), 2)
array([[ 0., 0., 0., 0.],
[ 0., 0., 0., 0.],
[ 0., 0., 1., 0.],
[ 0., 0., 0., 0.]])
Plot the impulse response of a 4th-order Butterworth lowpass filter:
>>> imp = signal.unit_impulse(100, 'mid')
>>> b, a = signal.butter(4, 0.2)
>>> response = signal.lfilter(b, a, imp)
>>> import matplotlib.pyplot as plt
>>> plt.plot(np.arange(-50, 50), imp)
>>> plt.plot(np.arange(-50, 50), response)
>>> plt.margins(0.1, 0.1)
>>> plt.xlabel('Time [samples]')
>>> plt.ylabel('Amplitude')
>>> plt.grid(True)
>>> plt.show() | Unit impulse signal (discrete delta function) or unit basis vector. | [
"Unit",
"impulse",
"signal",
"(",
"discrete",
"delta",
"function",
")",
"or",
"unit",
"basis",
"vector",
"."
] | def unit_impulse(shape, idx=None, dtype=float):
"""
Unit impulse signal (discrete delta function) or unit basis vector.
Parameters
----------
shape : int or tuple of int
Number of samples in the output (1-D), or a tuple that represents the
shape of the output (N-D).
idx : None or int or tuple of int or 'mid', optional
Index at which the value is 1. If None, defaults to the 0th element.
If ``idx='mid'``, the impulse will be centered at ``shape // 2`` in
all dimensions. If an int, the impulse will be at `idx` in all
dimensions.
dtype : data-type, optional
The desired data-type for the array, e.g., `numpy.int8`. Default is
`numpy.float64`.
Returns
-------
y : ndarray
Output array containing an impulse signal.
Notes
-----
The 1D case is also known as the Kronecker delta.
.. versionadded:: 0.19.0
Examples
--------
An impulse at the 0th element (:math:`\\delta[n]`):
>>> from scipy import signal
>>> signal.unit_impulse(8)
array([ 1., 0., 0., 0., 0., 0., 0., 0.])
Impulse offset by 2 samples (:math:`\\delta[n-2]`):
>>> signal.unit_impulse(7, 2)
array([ 0., 0., 1., 0., 0., 0., 0.])
2-dimensional impulse, centered:
>>> signal.unit_impulse((3, 3), 'mid')
array([[ 0., 0., 0.],
[ 0., 1., 0.],
[ 0., 0., 0.]])
Impulse at (2, 2), using broadcasting:
>>> signal.unit_impulse((4, 4), 2)
array([[ 0., 0., 0., 0.],
[ 0., 0., 0., 0.],
[ 0., 0., 1., 0.],
[ 0., 0., 0., 0.]])
Plot the impulse response of a 4th-order Butterworth lowpass filter:
>>> imp = signal.unit_impulse(100, 'mid')
>>> b, a = signal.butter(4, 0.2)
>>> response = signal.lfilter(b, a, imp)
>>> import matplotlib.pyplot as plt
>>> plt.plot(np.arange(-50, 50), imp)
>>> plt.plot(np.arange(-50, 50), response)
>>> plt.margins(0.1, 0.1)
>>> plt.xlabel('Time [samples]')
>>> plt.ylabel('Amplitude')
>>> plt.grid(True)
>>> plt.show()
"""
out = zeros(shape, dtype)
shape = np.atleast_1d(shape)
if idx is None:
idx = (0,) * len(shape)
elif idx == 'mid':
idx = tuple(shape // 2)
elif not hasattr(idx, "__iter__"):
idx = (idx,) * len(shape)
out[idx] = 1
return out | [
"def",
"unit_impulse",
"(",
"shape",
",",
"idx",
"=",
"None",
",",
"dtype",
"=",
"float",
")",
":",
"out",
"=",
"zeros",
"(",
"shape",
",",
"dtype",
")",
"shape",
"=",
"np",
".",
"atleast_1d",
"(",
"shape",
")",
"if",
"idx",
"is",
"None",
":",
"idx",
"=",
"(",
"0",
",",
")",
"*",
"len",
"(",
"shape",
")",
"elif",
"idx",
"==",
"'mid'",
":",
"idx",
"=",
"tuple",
"(",
"shape",
"//",
"2",
")",
"elif",
"not",
"hasattr",
"(",
"idx",
",",
"\"__iter__\"",
")",
":",
"idx",
"=",
"(",
"idx",
",",
")",
"*",
"len",
"(",
"shape",
")",
"out",
"[",
"idx",
"]",
"=",
"1",
"return",
"out"
] | https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/scipy/py2/scipy/signal/waveforms.py#L596-L681 |
|
psnonis/FinBERT | c0c555d833a14e2316a3701e59c0b5156f804b4e | bert/optimization.py | python | AdamWeightDecayOptimizer._do_use_weight_decay | (self, param_name) | return True | Whether to use L2 weight decay for `param_name`. | Whether to use L2 weight decay for `param_name`. | [
"Whether",
"to",
"use",
"L2",
"weight",
"decay",
"for",
"param_name",
"."
] | def _do_use_weight_decay(self, param_name):
"""Whether to use L2 weight decay for `param_name`."""
if not self.weight_decay_rate:
return False
if self.exclude_from_weight_decay:
for r in self.exclude_from_weight_decay:
if re.search(r, param_name) is not None:
return False
return True | [
"def",
"_do_use_weight_decay",
"(",
"self",
",",
"param_name",
")",
":",
"if",
"not",
"self",
".",
"weight_decay_rate",
":",
"return",
"False",
"if",
"self",
".",
"exclude_from_weight_decay",
":",
"for",
"r",
"in",
"self",
".",
"exclude_from_weight_decay",
":",
"if",
"re",
".",
"search",
"(",
"r",
",",
"param_name",
")",
"is",
"not",
"None",
":",
"return",
"False",
"return",
"True"
] | https://github.com/psnonis/FinBERT/blob/c0c555d833a14e2316a3701e59c0b5156f804b4e/bert/optimization.py#L159-L167 |
|
Xilinx/Vitis-AI | fc74d404563d9951b57245443c73bef389f3657f | tools/Vitis-AI-Quantizer/vai_q_pytorch/nndct_shared/quantization/quant_strategy.py | python | QuantStrategyBase._get_default_quant_config | (self,
quant_info_mgr,
lstm=False) | return config | 1. unified activation bits
2 .mixed bits for lstm | 1. unified activation bits
2 .mixed bits for lstm | [
"1",
".",
"unified",
"activation",
"bits",
"2",
".",
"mixed",
"bits",
"for",
"lstm"
] | def _get_default_quant_config(self,
quant_info_mgr,
lstm=False):
"""
1. unified activation bits
2 .mixed bits for lstm
"""
# import ipdb
# ipdb.set_trace()
config = {'param': {}, 'output': {}, 'input': {}}
for node in quant_info_mgr.Nndctgraph.nodes:
# print('---- Handling node %s type: %s' % (node.name, node.op.type))
if quant_info_mgr.is_node_quantizable(node, lstm):
# parameters
for k in quant_info_mgr.quant_node_params(node).keys():
p = quant_info_mgr.quant_node_params(node)[k]
# for mix precision quantization
bw = self._bits_act
if (node.has_bound_params() and
(hasattr(node.op.ParamName, 'WEIGHTS') and k == node.op.ParamName.WEIGHTS or
hasattr(node.op.ParamName, 'GAMMA') and k == node.op.ParamName.GAMMA)):
bw = self._bits_weight
config['param'][p.name] = [bw, None]
# print('---- Add fix of param %s' % p.name)
# output blobs
end = quant_info_mgr.quant_output(node.name).name
if end not in config['output']:
config['output'][end] = [self._bits_act, None]
# print('---- Add fix of output blob %s' % end)
# input blobs (for mix precision quantization)
if self._bits_weight != self._bits_act:
if node.op.type in [NNDCT_OP.DENSE, NNDCT_OP.CONV2D]:
config['input'][node.name] = [self._bits_weight, None]
# print('---- Add fix of input blob %s' % end)
elif (lstm and (node in quant_info_mgr.Nndctgraph.inputs)):
# print('---- Handling input node %s' % (node.name))
# this path is only for quantizing a whole graph without quant stub OP
# for lstm, check the following node type
if (node.in_quant_part or (any(
(quant_info_mgr.is_node_quantizable(c, lstm) and
c.op.type is not NNDCT_OP.QUANT_STUB)
for c in quant_info_mgr.Nndctgraph.children(node.name)))):
end = quant_info_mgr.quant_output(node.name).name
if end not in config['output']:
config['output'][end] = [self._bits_act, None]
# print('---- Add fix of quant net input blob %s' % end)
# check the input fix of all quantized ops
# import ipdb
# ipdb.set_trace()
if not lstm:
for node in quant_info_mgr.Nndctgraph.nodes:
if quant_info_mgr.is_node_quantizable(node, lstm):
#print('---- Check input of node %s type: %s' % (node.name, node.op.type))
if node.op.type not in [NNDCT_OP.INPUT, NNDCT_OP.QUANT_STUB, NNDCT_OP.CONCAT]:
for p_n in quant_info_mgr.Nndctgraph.parents(node):
# if not quant_info_mgr.op_unquantizable(p_n.op.type):
end = quant_info_mgr.quant_output(p_n.name).name
end_node = quant_info_mgr.Nndctgraph.node(end)
out_is_tensor = True
for tensor in end_node.out_tensors:
if tensor.shape == None:
out_is_tensor = False
if end not in config['output'] and out_is_tensor:
config['output'][end] = [self._bits_act, None]
#print('---- Add fix of output blob %s type: %s' % (end, end_node.op.type))
elif node.op.type in [NNDCT_OP.INPUT]:
cn_nodes = quant_info_mgr.Nndctgraph.children(node)
if len(cn_nodes) == 1 and cn_nodes[0].op.is_custom_op:
end = quant_info_mgr.quant_output(node.name).name
if end in config['output']:
del config['output'][end]
node.in_quant_part = False
return config | [
"def",
"_get_default_quant_config",
"(",
"self",
",",
"quant_info_mgr",
",",
"lstm",
"=",
"False",
")",
":",
"# import ipdb",
"# ipdb.set_trace()",
"config",
"=",
"{",
"'param'",
":",
"{",
"}",
",",
"'output'",
":",
"{",
"}",
",",
"'input'",
":",
"{",
"}",
"}",
"for",
"node",
"in",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"nodes",
":",
"# print('---- Handling node %s type: %s' % (node.name, node.op.type))",
"if",
"quant_info_mgr",
".",
"is_node_quantizable",
"(",
"node",
",",
"lstm",
")",
":",
"# parameters",
"for",
"k",
"in",
"quant_info_mgr",
".",
"quant_node_params",
"(",
"node",
")",
".",
"keys",
"(",
")",
":",
"p",
"=",
"quant_info_mgr",
".",
"quant_node_params",
"(",
"node",
")",
"[",
"k",
"]",
"# for mix precision quantization",
"bw",
"=",
"self",
".",
"_bits_act",
"if",
"(",
"node",
".",
"has_bound_params",
"(",
")",
"and",
"(",
"hasattr",
"(",
"node",
".",
"op",
".",
"ParamName",
",",
"'WEIGHTS'",
")",
"and",
"k",
"==",
"node",
".",
"op",
".",
"ParamName",
".",
"WEIGHTS",
"or",
"hasattr",
"(",
"node",
".",
"op",
".",
"ParamName",
",",
"'GAMMA'",
")",
"and",
"k",
"==",
"node",
".",
"op",
".",
"ParamName",
".",
"GAMMA",
")",
")",
":",
"bw",
"=",
"self",
".",
"_bits_weight",
"config",
"[",
"'param'",
"]",
"[",
"p",
".",
"name",
"]",
"=",
"[",
"bw",
",",
"None",
"]",
"# print('---- Add fix of param %s' % p.name)",
"# output blobs",
"end",
"=",
"quant_info_mgr",
".",
"quant_output",
"(",
"node",
".",
"name",
")",
".",
"name",
"if",
"end",
"not",
"in",
"config",
"[",
"'output'",
"]",
":",
"config",
"[",
"'output'",
"]",
"[",
"end",
"]",
"=",
"[",
"self",
".",
"_bits_act",
",",
"None",
"]",
"# print('---- Add fix of output blob %s' % end)",
"# input blobs (for mix precision quantization)",
"if",
"self",
".",
"_bits_weight",
"!=",
"self",
".",
"_bits_act",
":",
"if",
"node",
".",
"op",
".",
"type",
"in",
"[",
"NNDCT_OP",
".",
"DENSE",
",",
"NNDCT_OP",
".",
"CONV2D",
"]",
":",
"config",
"[",
"'input'",
"]",
"[",
"node",
".",
"name",
"]",
"=",
"[",
"self",
".",
"_bits_weight",
",",
"None",
"]",
"# print('---- Add fix of input blob %s' % end)",
"elif",
"(",
"lstm",
"and",
"(",
"node",
"in",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"inputs",
")",
")",
":",
"# print('---- Handling input node %s' % (node.name))",
"# this path is only for quantizing a whole graph without quant stub OP",
"# for lstm, check the following node type",
"if",
"(",
"node",
".",
"in_quant_part",
"or",
"(",
"any",
"(",
"(",
"quant_info_mgr",
".",
"is_node_quantizable",
"(",
"c",
",",
"lstm",
")",
"and",
"c",
".",
"op",
".",
"type",
"is",
"not",
"NNDCT_OP",
".",
"QUANT_STUB",
")",
"for",
"c",
"in",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"children",
"(",
"node",
".",
"name",
")",
")",
")",
")",
":",
"end",
"=",
"quant_info_mgr",
".",
"quant_output",
"(",
"node",
".",
"name",
")",
".",
"name",
"if",
"end",
"not",
"in",
"config",
"[",
"'output'",
"]",
":",
"config",
"[",
"'output'",
"]",
"[",
"end",
"]",
"=",
"[",
"self",
".",
"_bits_act",
",",
"None",
"]",
"# print('---- Add fix of quant net input blob %s' % end)",
"# check the input fix of all quantized ops ",
"# import ipdb",
"# ipdb.set_trace()",
"if",
"not",
"lstm",
":",
"for",
"node",
"in",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"nodes",
":",
"if",
"quant_info_mgr",
".",
"is_node_quantizable",
"(",
"node",
",",
"lstm",
")",
":",
"#print('---- Check input of node %s type: %s' % (node.name, node.op.type))",
"if",
"node",
".",
"op",
".",
"type",
"not",
"in",
"[",
"NNDCT_OP",
".",
"INPUT",
",",
"NNDCT_OP",
".",
"QUANT_STUB",
",",
"NNDCT_OP",
".",
"CONCAT",
"]",
":",
"for",
"p_n",
"in",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"parents",
"(",
"node",
")",
":",
"# if not quant_info_mgr.op_unquantizable(p_n.op.type):",
"end",
"=",
"quant_info_mgr",
".",
"quant_output",
"(",
"p_n",
".",
"name",
")",
".",
"name",
"end_node",
"=",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"node",
"(",
"end",
")",
"out_is_tensor",
"=",
"True",
"for",
"tensor",
"in",
"end_node",
".",
"out_tensors",
":",
"if",
"tensor",
".",
"shape",
"==",
"None",
":",
"out_is_tensor",
"=",
"False",
"if",
"end",
"not",
"in",
"config",
"[",
"'output'",
"]",
"and",
"out_is_tensor",
":",
"config",
"[",
"'output'",
"]",
"[",
"end",
"]",
"=",
"[",
"self",
".",
"_bits_act",
",",
"None",
"]",
"#print('---- Add fix of output blob %s type: %s' % (end, end_node.op.type))",
"elif",
"node",
".",
"op",
".",
"type",
"in",
"[",
"NNDCT_OP",
".",
"INPUT",
"]",
":",
"cn_nodes",
"=",
"quant_info_mgr",
".",
"Nndctgraph",
".",
"children",
"(",
"node",
")",
"if",
"len",
"(",
"cn_nodes",
")",
"==",
"1",
"and",
"cn_nodes",
"[",
"0",
"]",
".",
"op",
".",
"is_custom_op",
":",
"end",
"=",
"quant_info_mgr",
".",
"quant_output",
"(",
"node",
".",
"name",
")",
".",
"name",
"if",
"end",
"in",
"config",
"[",
"'output'",
"]",
":",
"del",
"config",
"[",
"'output'",
"]",
"[",
"end",
"]",
"node",
".",
"in_quant_part",
"=",
"False",
"return",
"config"
] | https://github.com/Xilinx/Vitis-AI/blob/fc74d404563d9951b57245443c73bef389f3657f/tools/Vitis-AI-Quantizer/vai_q_pytorch/nndct_shared/quantization/quant_strategy.py#L56-L132 |
|
mantidproject/mantid | 03deeb89254ec4289edb8771e0188c2090a02f32 | scripts/reduction_gui/reduction/scripter.py | python | BaseScriptElement.apply | (self) | return NotImplemented | Method called to apply the reduction script element
to a Mantid Reducer | Method called to apply the reduction script element
to a Mantid Reducer | [
"Method",
"called",
"to",
"apply",
"the",
"reduction",
"script",
"element",
"to",
"a",
"Mantid",
"Reducer"
] | def apply(self):
"""
Method called to apply the reduction script element
to a Mantid Reducer
"""
return NotImplemented | [
"def",
"apply",
"(",
"self",
")",
":",
"return",
"NotImplemented"
] | https://github.com/mantidproject/mantid/blob/03deeb89254ec4289edb8771e0188c2090a02f32/scripts/reduction_gui/reduction/scripter.py#L66-L71 |
|
ycm-core/ycmd | fc0fb7e5e15176cc5a2a30c80956335988c6b59a | ycmd/completers/language_server/language_server_completer.py | python | LanguageServerCompleter.WorkspaceConfigurationResponse | ( self, request ) | return None | If the concrete completer wants to respond to workspace/configuration
requests, it should override this method. | If the concrete completer wants to respond to workspace/configuration
requests, it should override this method. | [
"If",
"the",
"concrete",
"completer",
"wants",
"to",
"respond",
"to",
"workspace",
"/",
"configuration",
"requests",
"it",
"should",
"override",
"this",
"method",
"."
] | def WorkspaceConfigurationResponse( self, request ):
"""If the concrete completer wants to respond to workspace/configuration
requests, it should override this method."""
return None | [
"def",
"WorkspaceConfigurationResponse",
"(",
"self",
",",
"request",
")",
":",
"return",
"None"
] | https://github.com/ycm-core/ycmd/blob/fc0fb7e5e15176cc5a2a30c80956335988c6b59a/ycmd/completers/language_server/language_server_completer.py#L1597-L1600 |
|
echronos/echronos | c996f1d2c8af6c6536205eb319c1bf1d4d84569c | external_tools/pystache/renderer.py | python | Renderer._bytes_to_str | (self, _bytes) | return str(_bytes, self.string_encoding, self.decode_errors) | Convert a byte string to str, using string_encoding and decode_errors. | Convert a byte string to str, using string_encoding and decode_errors. | [
"Convert",
"a",
"byte",
"string",
"to",
"str",
"using",
"string_encoding",
"and",
"decode_errors",
"."
] | def _bytes_to_str(self, _bytes):
"""Convert a byte string to str, using string_encoding and decode_errors.
"""
assert type(_bytes) == bytes
return str(_bytes, self.string_encoding, self.decode_errors) | [
"def",
"_bytes_to_str",
"(",
"self",
",",
"_bytes",
")",
":",
"assert",
"type",
"(",
"_bytes",
")",
"==",
"bytes",
"return",
"str",
"(",
"_bytes",
",",
"self",
".",
"string_encoding",
",",
"self",
".",
"decode_errors",
")"
] | https://github.com/echronos/echronos/blob/c996f1d2c8af6c6536205eb319c1bf1d4d84569c/external_tools/pystache/renderer.py#L182-L187 |
|
catboost/catboost | 167f64f237114a4d10b2b4ee42adb4569137debe | contrib/python/pexpect/pexpect/screen.py | python | screen._decode | (self, s) | This converts from the external coding system (as passed to
the constructor) to the internal one (unicode). | This converts from the external coding system (as passed to
the constructor) to the internal one (unicode). | [
"This",
"converts",
"from",
"the",
"external",
"coding",
"system",
"(",
"as",
"passed",
"to",
"the",
"constructor",
")",
"to",
"the",
"internal",
"one",
"(",
"unicode",
")",
"."
] | def _decode(self, s):
'''This converts from the external coding system (as passed to
the constructor) to the internal one (unicode). '''
if self.decoder is not None:
return self.decoder.decode(s)
else:
raise TypeError("This screen was constructed with encoding=None, "
"so it does not handle bytes.") | [
"def",
"_decode",
"(",
"self",
",",
"s",
")",
":",
"if",
"self",
".",
"decoder",
"is",
"not",
"None",
":",
"return",
"self",
".",
"decoder",
".",
"decode",
"(",
"s",
")",
"else",
":",
"raise",
"TypeError",
"(",
"\"This screen was constructed with encoding=None, \"",
"\"so it does not handle bytes.\"",
")"
] | https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/pexpect/pexpect/screen.py#L104-L111 |
||
trilinos/Trilinos | 6168be6dd51e35e1cd681e9c4b24433e709df140 | packages/muelu/utils/analysis/tableau.py | python | tableau10 | () | return rgb2float(colors) | Tableau 10' colors as RGB | Tableau 10' colors as RGB | [
"Tableau",
"10",
"colors",
"as",
"RGB"
] | def tableau10():
"""'Tableau 10' colors as RGB"""
colors = [
( 31, 119, 180), (255, 127, 14), ( 44, 160, 44), (214, 39, 40),
(148, 103, 189), (140, 86, 75), (227, 119, 194), (127, 127, 127),
(188, 189, 34), ( 23, 190, 207)
]
return rgb2float(colors) | [
"def",
"tableau10",
"(",
")",
":",
"colors",
"=",
"[",
"(",
"31",
",",
"119",
",",
"180",
")",
",",
"(",
"255",
",",
"127",
",",
"14",
")",
",",
"(",
"44",
",",
"160",
",",
"44",
")",
",",
"(",
"214",
",",
"39",
",",
"40",
")",
",",
"(",
"148",
",",
"103",
",",
"189",
")",
",",
"(",
"140",
",",
"86",
",",
"75",
")",
",",
"(",
"227",
",",
"119",
",",
"194",
")",
",",
"(",
"127",
",",
"127",
",",
"127",
")",
",",
"(",
"188",
",",
"189",
",",
"34",
")",
",",
"(",
"23",
",",
"190",
",",
"207",
")",
"]",
"return",
"rgb2float",
"(",
"colors",
")"
] | https://github.com/trilinos/Trilinos/blob/6168be6dd51e35e1cd681e9c4b24433e709df140/packages/muelu/utils/analysis/tableau.py#L10-L17 |
|
takemaru/graphillion | 51879f92bb96b53ef8f914ef37a05252ce383617 | graphillion/graphset.py | python | GraphSet.paths | (terminal1, terminal2, is_hamilton=False, graphset=None) | return GraphSet.graphs(vertex_groups=[[terminal1, terminal2]],
degree_constraints=dc,
no_loop=True, graphset=graphset) | Returns a GraphSet of paths.
This method can be parallelized with OpenMP by specifying the
environmental variable `OMP_NUM_THREADS`:
`$ OMP_NUM_THREADS=4 python your_graphillion_script.py`
Examples:
>>> GraphSet.paths(1, 6)
GraphSet([[(1, 2), (2, 3), (3, 6)], [(1, 2), (2, 5), (5, 6)], [(1, 4), (4, 5 ...
Args:
terminal1 and terminal2: Both end vertices of a paths.
graphset: Optional. A GraphSet object. Paths to be stored
are selected from this object.
Returns:
A new GraphSet object.
See Also:
graphs() | Returns a GraphSet of paths. | [
"Returns",
"a",
"GraphSet",
"of",
"paths",
"."
] | def paths(terminal1, terminal2, is_hamilton=False, graphset=None):
"""Returns a GraphSet of paths.
This method can be parallelized with OpenMP by specifying the
environmental variable `OMP_NUM_THREADS`:
`$ OMP_NUM_THREADS=4 python your_graphillion_script.py`
Examples:
>>> GraphSet.paths(1, 6)
GraphSet([[(1, 2), (2, 3), (3, 6)], [(1, 2), (2, 5), (5, 6)], [(1, 4), (4, 5 ...
Args:
terminal1 and terminal2: Both end vertices of a paths.
graphset: Optional. A GraphSet object. Paths to be stored
are selected from this object.
Returns:
A new GraphSet object.
See Also:
graphs()
"""
dc = {}
for v in GraphSet._vertices:
if v in (terminal1, terminal2):
dc[v] = 1
else:
dc[v] = 2 if is_hamilton else range(0, 3, 2)
return GraphSet.graphs(vertex_groups=[[terminal1, terminal2]],
degree_constraints=dc,
no_loop=True, graphset=graphset) | [
"def",
"paths",
"(",
"terminal1",
",",
"terminal2",
",",
"is_hamilton",
"=",
"False",
",",
"graphset",
"=",
"None",
")",
":",
"dc",
"=",
"{",
"}",
"for",
"v",
"in",
"GraphSet",
".",
"_vertices",
":",
"if",
"v",
"in",
"(",
"terminal1",
",",
"terminal2",
")",
":",
"dc",
"[",
"v",
"]",
"=",
"1",
"else",
":",
"dc",
"[",
"v",
"]",
"=",
"2",
"if",
"is_hamilton",
"else",
"range",
"(",
"0",
",",
"3",
",",
"2",
")",
"return",
"GraphSet",
".",
"graphs",
"(",
"vertex_groups",
"=",
"[",
"[",
"terminal1",
",",
"terminal2",
"]",
"]",
",",
"degree_constraints",
"=",
"dc",
",",
"no_loop",
"=",
"True",
",",
"graphset",
"=",
"graphset",
")"
] | https://github.com/takemaru/graphillion/blob/51879f92bb96b53ef8f914ef37a05252ce383617/graphillion/graphset.py#L1941-L1973 |
|
aws/lumberyard | f85344403c1c2e77ec8c75deb2c116e97b713217 | dev/Tools/Python/3.7.10/mac/Python.framework/Versions/3.7/lib/python3.7/plistlib.py | python | _BinaryPlistParser._get_size | (self, tokenL) | return tokenL | return the size of the next object. | return the size of the next object. | [
"return",
"the",
"size",
"of",
"the",
"next",
"object",
"."
] | def _get_size(self, tokenL):
""" return the size of the next object."""
if tokenL == 0xF:
m = self._fp.read(1)[0] & 0x3
s = 1 << m
f = '>' + _BINARY_FORMAT[s]
return struct.unpack(f, self._fp.read(s))[0]
return tokenL | [
"def",
"_get_size",
"(",
"self",
",",
"tokenL",
")",
":",
"if",
"tokenL",
"==",
"0xF",
":",
"m",
"=",
"self",
".",
"_fp",
".",
"read",
"(",
"1",
")",
"[",
"0",
"]",
"&",
"0x3",
"s",
"=",
"1",
"<<",
"m",
"f",
"=",
"'>'",
"+",
"_BINARY_FORMAT",
"[",
"s",
"]",
"return",
"struct",
".",
"unpack",
"(",
"f",
",",
"self",
".",
"_fp",
".",
"read",
"(",
"s",
")",
")",
"[",
"0",
"]",
"return",
"tokenL"
] | https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/mac/Python.framework/Versions/3.7/lib/python3.7/plistlib.py#L574-L582 |
|
catboost/catboost | 167f64f237114a4d10b2b4ee42adb4569137debe | contrib/python/prompt-toolkit/py3/prompt_toolkit/input/win32_pipe.py | python | Win32PipeInput.send_bytes | (self, data: bytes) | Send bytes to the input. | Send bytes to the input. | [
"Send",
"bytes",
"to",
"the",
"input",
"."
] | def send_bytes(self, data: bytes) -> None:
"Send bytes to the input."
self.send_text(data.decode("utf-8", "ignore")) | [
"def",
"send_bytes",
"(",
"self",
",",
"data",
":",
"bytes",
")",
"->",
"None",
":",
"self",
".",
"send_text",
"(",
"data",
".",
"decode",
"(",
"\"utf-8\"",
",",
"\"ignore\"",
")",
")"
] | https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/prompt-toolkit/py3/prompt_toolkit/input/win32_pipe.py#L108-L110 |
||
baidu-research/tensorflow-allreduce | 66d5b855e90b0949e9fa5cca5599fd729a70e874 | tensorflow/contrib/tensor_forest/client/eval_metrics.py | python | get_metric | (metric_name) | return _EVAL_METRICS[metric_name] | Given a metric name, return the corresponding metric function. | Given a metric name, return the corresponding metric function. | [
"Given",
"a",
"metric",
"name",
"return",
"the",
"corresponding",
"metric",
"function",
"."
] | def get_metric(metric_name):
"""Given a metric name, return the corresponding metric function."""
return _EVAL_METRICS[metric_name] | [
"def",
"get_metric",
"(",
"metric_name",
")",
":",
"return",
"_EVAL_METRICS",
"[",
"metric_name",
"]"
] | https://github.com/baidu-research/tensorflow-allreduce/blob/66d5b855e90b0949e9fa5cca5599fd729a70e874/tensorflow/contrib/tensor_forest/client/eval_metrics.py#L154-L156 |
|
pytorch/pytorch | 7176c92687d3cc847cc046bf002269c6949a21c2 | torch/distributed/pipeline/sync/phony.py | python | get_phony | (device: torch.device, *, requires_grad: bool) | return phony | Gets a phony. Phony is tensor without space. It is useful to make
arbitrary dependency in a autograd graph because it doesn't require any
gradient accumulation.
.. note::
Phonies for each device are cached. If an autograd function gets a phony
internally, the phony must be detached to be returned. Otherwise, the
autograd engine will mutate the cached phony in-place::
class Phonify(torch.autograd.Function):
@staticmethod
def forward(ctx, input):
phony = get_phony(input.device, requires_grad=False)
return phony.detach() # detach() is necessary. | Gets a phony. Phony is tensor without space. It is useful to make
arbitrary dependency in a autograd graph because it doesn't require any
gradient accumulation. | [
"Gets",
"a",
"phony",
".",
"Phony",
"is",
"tensor",
"without",
"space",
".",
"It",
"is",
"useful",
"to",
"make",
"arbitrary",
"dependency",
"in",
"a",
"autograd",
"graph",
"because",
"it",
"doesn",
"t",
"require",
"any",
"gradient",
"accumulation",
"."
] | def get_phony(device: torch.device, *, requires_grad: bool) -> Tensor:
"""Gets a phony. Phony is tensor without space. It is useful to make
arbitrary dependency in a autograd graph because it doesn't require any
gradient accumulation.
.. note::
Phonies for each device are cached. If an autograd function gets a phony
internally, the phony must be detached to be returned. Otherwise, the
autograd engine will mutate the cached phony in-place::
class Phonify(torch.autograd.Function):
@staticmethod
def forward(ctx, input):
phony = get_phony(input.device, requires_grad=False)
return phony.detach() # detach() is necessary.
"""
key = (device, requires_grad)
try:
phony = _phonies[key]
except KeyError:
with use_stream(default_stream(device)):
phony = torch.empty(0, device=device, requires_grad=requires_grad)
_phonies[key] = phony
return phony | [
"def",
"get_phony",
"(",
"device",
":",
"torch",
".",
"device",
",",
"*",
",",
"requires_grad",
":",
"bool",
")",
"->",
"Tensor",
":",
"key",
"=",
"(",
"device",
",",
"requires_grad",
")",
"try",
":",
"phony",
"=",
"_phonies",
"[",
"key",
"]",
"except",
"KeyError",
":",
"with",
"use_stream",
"(",
"default_stream",
"(",
"device",
")",
")",
":",
"phony",
"=",
"torch",
".",
"empty",
"(",
"0",
",",
"device",
"=",
"device",
",",
"requires_grad",
"=",
"requires_grad",
")",
"_phonies",
"[",
"key",
"]",
"=",
"phony",
"return",
"phony"
] | https://github.com/pytorch/pytorch/blob/7176c92687d3cc847cc046bf002269c6949a21c2/torch/distributed/pipeline/sync/phony.py#L21-L49 |
|
wxWidgets/wxPython-Classic | 19571e1ae65f1ac445f5491474121998c97a1bf0 | src/gtk/_core.py | python | Control_RemoveMnemonics | (*args, **kwargs) | return _core_.Control_RemoveMnemonics(*args, **kwargs) | Control_RemoveMnemonics(String str) -> String
removes the mnemonics characters | Control_RemoveMnemonics(String str) -> String | [
"Control_RemoveMnemonics",
"(",
"String",
"str",
")",
"-",
">",
"String"
] | def Control_RemoveMnemonics(*args, **kwargs):
"""
Control_RemoveMnemonics(String str) -> String
removes the mnemonics characters
"""
return _core_.Control_RemoveMnemonics(*args, **kwargs) | [
"def",
"Control_RemoveMnemonics",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"_core_",
".",
"Control_RemoveMnemonics",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/gtk/_core.py#L12781-L12787 |
|
Xilinx/Vitis-AI | fc74d404563d9951b57245443c73bef389f3657f | tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/training/input.py | python | slice_input_producer | (tensor_list, num_epochs=None, shuffle=True, seed=None,
capacity=32, shared_name=None, name=None) | Produces a slice of each `Tensor` in `tensor_list`.
Implemented using a Queue -- a `QueueRunner` for the Queue
is added to the current `Graph`'s `QUEUE_RUNNER` collection.
Args:
tensor_list: A list of `Tensor` objects. Every `Tensor` in
`tensor_list` must have the same size in the first dimension.
num_epochs: An integer (optional). If specified, `slice_input_producer`
produces each slice `num_epochs` times before generating
an `OutOfRange` error. If not specified, `slice_input_producer` can cycle
through the slices an unlimited number of times.
shuffle: Boolean. If true, the integers are randomly shuffled within each
epoch.
seed: An integer (optional). Seed used if shuffle == True.
capacity: An integer. Sets the queue capacity.
shared_name: (optional). If set, this queue will be shared under the given
name across multiple sessions.
name: A name for the operations (optional).
Returns:
A list of tensors, one for each element of `tensor_list`. If the tensor
in `tensor_list` has shape `[N, a, b, .., z]`, then the corresponding output
tensor will have shape `[a, b, ..., z]`.
Raises:
ValueError: if `slice_input_producer` produces nothing from `tensor_list`.
@compatibility(eager)
Input pipelines based on Queues are not supported when eager execution is
enabled. Please use the `tf.data` API to ingest data under eager execution.
@end_compatibility | Produces a slice of each `Tensor` in `tensor_list`. | [
"Produces",
"a",
"slice",
"of",
"each",
"Tensor",
"in",
"tensor_list",
"."
] | def slice_input_producer(tensor_list, num_epochs=None, shuffle=True, seed=None,
capacity=32, shared_name=None, name=None):
"""Produces a slice of each `Tensor` in `tensor_list`.
Implemented using a Queue -- a `QueueRunner` for the Queue
is added to the current `Graph`'s `QUEUE_RUNNER` collection.
Args:
tensor_list: A list of `Tensor` objects. Every `Tensor` in
`tensor_list` must have the same size in the first dimension.
num_epochs: An integer (optional). If specified, `slice_input_producer`
produces each slice `num_epochs` times before generating
an `OutOfRange` error. If not specified, `slice_input_producer` can cycle
through the slices an unlimited number of times.
shuffle: Boolean. If true, the integers are randomly shuffled within each
epoch.
seed: An integer (optional). Seed used if shuffle == True.
capacity: An integer. Sets the queue capacity.
shared_name: (optional). If set, this queue will be shared under the given
name across multiple sessions.
name: A name for the operations (optional).
Returns:
A list of tensors, one for each element of `tensor_list`. If the tensor
in `tensor_list` has shape `[N, a, b, .., z]`, then the corresponding output
tensor will have shape `[a, b, ..., z]`.
Raises:
ValueError: if `slice_input_producer` produces nothing from `tensor_list`.
@compatibility(eager)
Input pipelines based on Queues are not supported when eager execution is
enabled. Please use the `tf.data` API to ingest data under eager execution.
@end_compatibility
"""
with ops.name_scope(name, "input_producer", tensor_list):
tensor_list = ops.convert_n_to_tensor_or_indexed_slices(tensor_list)
if not tensor_list:
raise ValueError(
"Expected at least one tensor in slice_input_producer().")
range_size = array_ops.shape(tensor_list[0])[0]
# TODO(josh11b): Add an assertion that the first dimension of
# everything in TensorList matches. Maybe just check the inferred shapes?
queue = range_input_producer(range_size, num_epochs=num_epochs,
shuffle=shuffle, seed=seed, capacity=capacity,
shared_name=shared_name)
index = queue.dequeue()
output = [array_ops.gather(t, index) for t in tensor_list]
return output | [
"def",
"slice_input_producer",
"(",
"tensor_list",
",",
"num_epochs",
"=",
"None",
",",
"shuffle",
"=",
"True",
",",
"seed",
"=",
"None",
",",
"capacity",
"=",
"32",
",",
"shared_name",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"with",
"ops",
".",
"name_scope",
"(",
"name",
",",
"\"input_producer\"",
",",
"tensor_list",
")",
":",
"tensor_list",
"=",
"ops",
".",
"convert_n_to_tensor_or_indexed_slices",
"(",
"tensor_list",
")",
"if",
"not",
"tensor_list",
":",
"raise",
"ValueError",
"(",
"\"Expected at least one tensor in slice_input_producer().\"",
")",
"range_size",
"=",
"array_ops",
".",
"shape",
"(",
"tensor_list",
"[",
"0",
"]",
")",
"[",
"0",
"]",
"# TODO(josh11b): Add an assertion that the first dimension of",
"# everything in TensorList matches. Maybe just check the inferred shapes?",
"queue",
"=",
"range_input_producer",
"(",
"range_size",
",",
"num_epochs",
"=",
"num_epochs",
",",
"shuffle",
"=",
"shuffle",
",",
"seed",
"=",
"seed",
",",
"capacity",
"=",
"capacity",
",",
"shared_name",
"=",
"shared_name",
")",
"index",
"=",
"queue",
".",
"dequeue",
"(",
")",
"output",
"=",
"[",
"array_ops",
".",
"gather",
"(",
"t",
",",
"index",
")",
"for",
"t",
"in",
"tensor_list",
"]",
"return",
"output"
] | https://github.com/Xilinx/Vitis-AI/blob/fc74d404563d9951b57245443c73bef389f3657f/tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/training/input.py#L328-L376 |
||
wxWidgets/wxPython-Classic | 19571e1ae65f1ac445f5491474121998c97a1bf0 | src/osx_cocoa/_misc.py | python | DataObjectComposite.GetObject | (*args, **kwargs) | return _misc_.DataObjectComposite_GetObject(*args, **kwargs) | GetObject(self, DataFormat format, wxDataObjectBase::Direction dir=Get) -> DataObjectSimple
Returns the pointer to the object which supports this format or None.
TODO: Fix this to use OOR and return the right object type. | GetObject(self, DataFormat format, wxDataObjectBase::Direction dir=Get) -> DataObjectSimple | [
"GetObject",
"(",
"self",
"DataFormat",
"format",
"wxDataObjectBase",
"::",
"Direction",
"dir",
"=",
"Get",
")",
"-",
">",
"DataObjectSimple"
] | def GetObject(*args, **kwargs):
"""
GetObject(self, DataFormat format, wxDataObjectBase::Direction dir=Get) -> DataObjectSimple
Returns the pointer to the object which supports this format or None.
TODO: Fix this to use OOR and return the right object type.
"""
return _misc_.DataObjectComposite_GetObject(*args, **kwargs) | [
"def",
"GetObject",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"_misc_",
".",
"DataObjectComposite_GetObject",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_cocoa/_misc.py#L5154-L5161 |
|
baidu-research/tensorflow-allreduce | 66d5b855e90b0949e9fa5cca5599fd729a70e874 | tensorflow/contrib/data/python/ops/dataset_ops.py | python | ZipDataset.__init__ | (self, datasets) | See `Dataset.zip()` for details. | See `Dataset.zip()` for details. | [
"See",
"Dataset",
".",
"zip",
"()",
"for",
"details",
"."
] | def __init__(self, datasets):
"""See `Dataset.zip()` for details."""
super(ZipDataset, self).__init__()
self._datasets = datasets | [
"def",
"__init__",
"(",
"self",
",",
"datasets",
")",
":",
"super",
"(",
"ZipDataset",
",",
"self",
")",
".",
"__init__",
"(",
")",
"self",
".",
"_datasets",
"=",
"datasets"
] | https://github.com/baidu-research/tensorflow-allreduce/blob/66d5b855e90b0949e9fa5cca5599fd729a70e874/tensorflow/contrib/data/python/ops/dataset_ops.py#L1171-L1174 |
||
hfinkel/llvm-project-cxxjit | 91084ef018240bbb8e24235ff5cd8c355a9c1a1e | clang/bindings/python/clang/cindex.py | python | Type.get_offset | (self, fieldname) | return conf.lib.clang_Type_getOffsetOf(self, fieldname) | Retrieve the offset of a field in the record. | Retrieve the offset of a field in the record. | [
"Retrieve",
"the",
"offset",
"of",
"a",
"field",
"in",
"the",
"record",
"."
] | def get_offset(self, fieldname):
"""
Retrieve the offset of a field in the record.
"""
return conf.lib.clang_Type_getOffsetOf(self, fieldname) | [
"def",
"get_offset",
"(",
"self",
",",
"fieldname",
")",
":",
"return",
"conf",
".",
"lib",
".",
"clang_Type_getOffsetOf",
"(",
"self",
",",
"fieldname",
")"
] | https://github.com/hfinkel/llvm-project-cxxjit/blob/91084ef018240bbb8e24235ff5cd8c355a9c1a1e/clang/bindings/python/clang/cindex.py#L2384-L2388 |
|
strasdat/Sophus | 36b08885e094fda63e92ad89d65be380c288265a | sympy/sophus/complex.py | python | Complex.Da_a_mul_b | (a, b) | return sympy.Matrix([[b.real, -b.imag],
[b.imag, b.real]]) | derivatice of complex muliplication wrt left multiplier a | derivatice of complex muliplication wrt left multiplier a | [
"derivatice",
"of",
"complex",
"muliplication",
"wrt",
"left",
"multiplier",
"a"
] | def Da_a_mul_b(a, b):
""" derivatice of complex muliplication wrt left multiplier a """
return sympy.Matrix([[b.real, -b.imag],
[b.imag, b.real]]) | [
"def",
"Da_a_mul_b",
"(",
"a",
",",
"b",
")",
":",
"return",
"sympy",
".",
"Matrix",
"(",
"[",
"[",
"b",
".",
"real",
",",
"-",
"b",
".",
"imag",
"]",
",",
"[",
"b",
".",
"imag",
",",
"b",
".",
"real",
"]",
"]",
")"
] | https://github.com/strasdat/Sophus/blob/36b08885e094fda63e92ad89d65be380c288265a/sympy/sophus/complex.py#L72-L75 |
|
BlzFans/wke | b0fa21158312e40c5fbd84682d643022b6c34a93 | cygwin/lib/python2.6/platform.py | python | linux_distribution | (distname='', version='', id='',
supported_dists=_supported_dists,
full_distribution_name=1) | return distname, version, id | Tries to determine the name of the Linux OS distribution name.
The function first looks for a distribution release file in
/etc and then reverts to _dist_try_harder() in case no
suitable files are found.
supported_dists may be given to define the set of Linux
distributions to look for. It defaults to a list of currently
supported Linux distributions identified by their release file
name.
If full_distribution_name is true (default), the full
distribution read from the OS is returned. Otherwise the short
name taken from supported_dists is used.
Returns a tuple (distname,version,id) which default to the
args given as parameters. | Tries to determine the name of the Linux OS distribution name. | [
"Tries",
"to",
"determine",
"the",
"name",
"of",
"the",
"Linux",
"OS",
"distribution",
"name",
"."
] | def linux_distribution(distname='', version='', id='',
supported_dists=_supported_dists,
full_distribution_name=1):
""" Tries to determine the name of the Linux OS distribution name.
The function first looks for a distribution release file in
/etc and then reverts to _dist_try_harder() in case no
suitable files are found.
supported_dists may be given to define the set of Linux
distributions to look for. It defaults to a list of currently
supported Linux distributions identified by their release file
name.
If full_distribution_name is true (default), the full
distribution read from the OS is returned. Otherwise the short
name taken from supported_dists is used.
Returns a tuple (distname,version,id) which default to the
args given as parameters.
"""
try:
etc = os.listdir('/etc')
except os.error:
# Probably not a Unix system
return distname,version,id
etc.sort()
for file in etc:
m = _release_filename.match(file)
if m is not None:
_distname,dummy = m.groups()
if _distname in supported_dists:
distname = _distname
break
else:
return _dist_try_harder(distname,version,id)
# Read the first line
f = open('/etc/'+file, 'r')
firstline = f.readline()
f.close()
_distname, _version, _id = _parse_release_file(firstline)
if _distname and full_distribution_name:
distname = _distname
if _version:
version = _version
if _id:
id = _id
return distname, version, id | [
"def",
"linux_distribution",
"(",
"distname",
"=",
"''",
",",
"version",
"=",
"''",
",",
"id",
"=",
"''",
",",
"supported_dists",
"=",
"_supported_dists",
",",
"full_distribution_name",
"=",
"1",
")",
":",
"try",
":",
"etc",
"=",
"os",
".",
"listdir",
"(",
"'/etc'",
")",
"except",
"os",
".",
"error",
":",
"# Probably not a Unix system",
"return",
"distname",
",",
"version",
",",
"id",
"etc",
".",
"sort",
"(",
")",
"for",
"file",
"in",
"etc",
":",
"m",
"=",
"_release_filename",
".",
"match",
"(",
"file",
")",
"if",
"m",
"is",
"not",
"None",
":",
"_distname",
",",
"dummy",
"=",
"m",
".",
"groups",
"(",
")",
"if",
"_distname",
"in",
"supported_dists",
":",
"distname",
"=",
"_distname",
"break",
"else",
":",
"return",
"_dist_try_harder",
"(",
"distname",
",",
"version",
",",
"id",
")",
"# Read the first line",
"f",
"=",
"open",
"(",
"'/etc/'",
"+",
"file",
",",
"'r'",
")",
"firstline",
"=",
"f",
".",
"readline",
"(",
")",
"f",
".",
"close",
"(",
")",
"_distname",
",",
"_version",
",",
"_id",
"=",
"_parse_release_file",
"(",
"firstline",
")",
"if",
"_distname",
"and",
"full_distribution_name",
":",
"distname",
"=",
"_distname",
"if",
"_version",
":",
"version",
"=",
"_version",
"if",
"_id",
":",
"id",
"=",
"_id",
"return",
"distname",
",",
"version",
",",
"id"
] | https://github.com/BlzFans/wke/blob/b0fa21158312e40c5fbd84682d643022b6c34a93/cygwin/lib/python2.6/platform.py#L293-L345 |
|
apple/swift | 469f72fdae2ea828b3b6c0d7d62d7e4cf98c4893 | utils/swift_build_support/swift_build_support/products/swiftsyntax.py | python | SwiftSyntax.product_source_name | (cls) | return "swift-syntax" | product_source_name() -> str
The name of the source code directory of this product. | product_source_name() -> str | [
"product_source_name",
"()",
"-",
">",
"str"
] | def product_source_name(cls):
"""product_source_name() -> str
The name of the source code directory of this product.
"""
return "swift-syntax" | [
"def",
"product_source_name",
"(",
"cls",
")",
":",
"return",
"\"swift-syntax\""
] | https://github.com/apple/swift/blob/469f72fdae2ea828b3b6c0d7d62d7e4cf98c4893/utils/swift_build_support/swift_build_support/products/swiftsyntax.py#L33-L38 |
|
krishauser/Klampt | 972cc83ea5befac3f653c1ba20f80155768ad519 | Python/control-examples/system_id.py | python | LinearSystemID.fixC | (self,i,value) | Sets the i'th entry of the C vector to a fixed value | Sets the i'th entry of the C vector to a fixed value | [
"Sets",
"the",
"i",
"th",
"entry",
"of",
"the",
"C",
"vector",
"to",
"a",
"fixed",
"value"
] | def fixC(self,i,value):
"""Sets the i'th entry of the C vector to a fixed value"""
if self.coeffPattern[2] == None:
m,n=self.m,self.n
self.coeffPattern[2] = [None]*m
self.coeffPattern[2][i]=value
self._updateEstimatorSize(i) | [
"def",
"fixC",
"(",
"self",
",",
"i",
",",
"value",
")",
":",
"if",
"self",
".",
"coeffPattern",
"[",
"2",
"]",
"==",
"None",
":",
"m",
",",
"n",
"=",
"self",
".",
"m",
",",
"self",
".",
"n",
"self",
".",
"coeffPattern",
"[",
"2",
"]",
"=",
"[",
"None",
"]",
"*",
"m",
"self",
".",
"coeffPattern",
"[",
"2",
"]",
"[",
"i",
"]",
"=",
"value",
"self",
".",
"_updateEstimatorSize",
"(",
"i",
")"
] | https://github.com/krishauser/Klampt/blob/972cc83ea5befac3f653c1ba20f80155768ad519/Python/control-examples/system_id.py#L39-L45 |
||
facebook/proxygen | a9ca025af207787815cb01eee1971cd572c7a81e | build/fbcode_builder/shell_quoting.py | python | ShellQuoted.__new__ | (cls, s) | return super(ShellQuoted, cls).__new__(
cls, s.do_not_use_raw_str if isinstance(s, ShellQuoted) else s
) | No need to nest ShellQuoted. | No need to nest ShellQuoted. | [
"No",
"need",
"to",
"nest",
"ShellQuoted",
"."
] | def __new__(cls, s):
"No need to nest ShellQuoted."
return super(ShellQuoted, cls).__new__(
cls, s.do_not_use_raw_str if isinstance(s, ShellQuoted) else s
) | [
"def",
"__new__",
"(",
"cls",
",",
"s",
")",
":",
"return",
"super",
"(",
"ShellQuoted",
",",
"cls",
")",
".",
"__new__",
"(",
"cls",
",",
"s",
".",
"do_not_use_raw_str",
"if",
"isinstance",
"(",
"s",
",",
"ShellQuoted",
")",
"else",
"s",
")"
] | https://github.com/facebook/proxygen/blob/a9ca025af207787815cb01eee1971cd572c7a81e/build/fbcode_builder/shell_quoting.py#L34-L38 |
|
Polidea/SiriusObfuscator | b0e590d8130e97856afe578869b83a209e2b19be | SymbolExtractorAndRenamer/lldb/third_party/Python/module/pexpect-2.4/screen.py | python | screen.cr | (self) | This moves the cursor to the beginning (col 1) of the current row. | This moves the cursor to the beginning (col 1) of the current row. | [
"This",
"moves",
"the",
"cursor",
"to",
"the",
"beginning",
"(",
"col",
"1",
")",
"of",
"the",
"current",
"row",
"."
] | def cr(self):
"""This moves the cursor to the beginning (col 1) of the current row.
"""
self.cursor_home(self.cur_r, 1) | [
"def",
"cr",
"(",
"self",
")",
":",
"self",
".",
"cursor_home",
"(",
"self",
".",
"cur_r",
",",
"1",
")"
] | https://github.com/Polidea/SiriusObfuscator/blob/b0e590d8130e97856afe578869b83a209e2b19be/SymbolExtractorAndRenamer/lldb/third_party/Python/module/pexpect-2.4/screen.py#L101-L105 |
||
zju3dv/clean-pvnet | 5870c509e3cc205e1bb28910a7b1a9a3c8add9a8 | lib/utils/meshrenderer/gl_utils/camera.py | python | Camera.setIntrinsic | (self, I, W, H, near, far, scale=1.0, originIsInTopLeft=True) | Args:
I: 3x3 intrinsic camera matrix from real camera (without any OpenGL stuff)
W: Width of the camera image
H: Height of the camera image
near: Near plane
far: Far plane
originIsInTopLeft: If True then the image origin is in top left
if False the image origin is in image center
Source: http://ksimek.github.io/2013/06/03/calibrated_cameras_in_opengl/ | Args:
I: 3x3 intrinsic camera matrix from real camera (without any OpenGL stuff)
W: Width of the camera image
H: Height of the camera image
near: Near plane
far: Far plane
originIsInTopLeft: If True then the image origin is in top left
if False the image origin is in image center
Source: http://ksimek.github.io/2013/06/03/calibrated_cameras_in_opengl/ | [
"Args",
":",
"I",
":",
"3x3",
"intrinsic",
"camera",
"matrix",
"from",
"real",
"camera",
"(",
"without",
"any",
"OpenGL",
"stuff",
")",
"W",
":",
"Width",
"of",
"the",
"camera",
"image",
"H",
":",
"Height",
"of",
"the",
"camera",
"image",
"near",
":",
"Near",
"plane",
"far",
":",
"Far",
"plane",
"originIsInTopLeft",
":",
"If",
"True",
"then",
"the",
"image",
"origin",
"is",
"in",
"top",
"left",
"if",
"False",
"the",
"image",
"origin",
"is",
"in",
"image",
"center",
"Source",
":",
"http",
":",
"//",
"ksimek",
".",
"github",
".",
"io",
"/",
"2013",
"/",
"06",
"/",
"03",
"/",
"calibrated_cameras_in_opengl",
"/"
] | def setIntrinsic(self, I, W, H, near, far, scale=1.0, originIsInTopLeft=True):
'''
Args:
I: 3x3 intrinsic camera matrix from real camera (without any OpenGL stuff)
W: Width of the camera image
H: Height of the camera image
near: Near plane
far: Far plane
originIsInTopLeft: If True then the image origin is in top left
if False the image origin is in image center
Source: http://ksimek.github.io/2013/06/03/calibrated_cameras_in_opengl/
'''
Camera.__check_matrix__(I)
A = near + far
B = near * far
persp = np.array( [ [ I[0,0]*scale, I[0,1]*scale, -I[0,2]*scale, 0 ],
[ 0 , I[1,1]*scale, -I[1,2]*scale, 0 ],
[ 0 , 0 , A , B ],
[ 0 , 0 , -1 , 0 ] ] , dtype=np.float64)
ortho = Camera.__glOrtho__(0, W, H, 0, near, far) if originIsInTopLeft else\
Camera.__glOrtho__(-W/2., W/2., -H/2., H/2., near, far)
self.__T_proj_view[:] = np.dot( ortho, persp ).astype(np.float32)
self.__T_view_proj[:] = np.linalg.inv(self.__T_proj_view)
self.__T_proj_world[:] = np.dot(self.__T_proj_view, self.__T_view_world)
self.dirty = True | [
"def",
"setIntrinsic",
"(",
"self",
",",
"I",
",",
"W",
",",
"H",
",",
"near",
",",
"far",
",",
"scale",
"=",
"1.0",
",",
"originIsInTopLeft",
"=",
"True",
")",
":",
"Camera",
".",
"__check_matrix__",
"(",
"I",
")",
"A",
"=",
"near",
"+",
"far",
"B",
"=",
"near",
"*",
"far",
"persp",
"=",
"np",
".",
"array",
"(",
"[",
"[",
"I",
"[",
"0",
",",
"0",
"]",
"*",
"scale",
",",
"I",
"[",
"0",
",",
"1",
"]",
"*",
"scale",
",",
"-",
"I",
"[",
"0",
",",
"2",
"]",
"*",
"scale",
",",
"0",
"]",
",",
"[",
"0",
",",
"I",
"[",
"1",
",",
"1",
"]",
"*",
"scale",
",",
"-",
"I",
"[",
"1",
",",
"2",
"]",
"*",
"scale",
",",
"0",
"]",
",",
"[",
"0",
",",
"0",
",",
"A",
",",
"B",
"]",
",",
"[",
"0",
",",
"0",
",",
"-",
"1",
",",
"0",
"]",
"]",
",",
"dtype",
"=",
"np",
".",
"float64",
")",
"ortho",
"=",
"Camera",
".",
"__glOrtho__",
"(",
"0",
",",
"W",
",",
"H",
",",
"0",
",",
"near",
",",
"far",
")",
"if",
"originIsInTopLeft",
"else",
"Camera",
".",
"__glOrtho__",
"(",
"-",
"W",
"/",
"2.",
",",
"W",
"/",
"2.",
",",
"-",
"H",
"/",
"2.",
",",
"H",
"/",
"2.",
",",
"near",
",",
"far",
")",
"self",
".",
"__T_proj_view",
"[",
":",
"]",
"=",
"np",
".",
"dot",
"(",
"ortho",
",",
"persp",
")",
".",
"astype",
"(",
"np",
".",
"float32",
")",
"self",
".",
"__T_view_proj",
"[",
":",
"]",
"=",
"np",
".",
"linalg",
".",
"inv",
"(",
"self",
".",
"__T_proj_view",
")",
"self",
".",
"__T_proj_world",
"[",
":",
"]",
"=",
"np",
".",
"dot",
"(",
"self",
".",
"__T_proj_view",
",",
"self",
".",
"__T_view_world",
")",
"self",
".",
"dirty",
"=",
"True"
] | https://github.com/zju3dv/clean-pvnet/blob/5870c509e3cc205e1bb28910a7b1a9a3c8add9a8/lib/utils/meshrenderer/gl_utils/camera.py#L139-L166 |
||
wxWidgets/wxPython-Classic | 19571e1ae65f1ac445f5491474121998c97a1bf0 | wx/py/pseudo.py | python | PseudoFile.__init__ | (self) | Create a file-like object. | Create a file-like object. | [
"Create",
"a",
"file",
"-",
"like",
"object",
"."
] | def __init__(self):
"""Create a file-like object."""
pass | [
"def",
"__init__",
"(",
"self",
")",
":",
"pass"
] | https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/wx/py/pseudo.py#L46-L48 |
||
catboost/catboost | 167f64f237114a4d10b2b4ee42adb4569137debe | contrib/python/ipython/py2/IPython/utils/text.py | python | strip_ansi | (source) | return re.sub(r'\033\[(\d|;)+?m', '', source) | Remove ansi escape codes from text.
Parameters
----------
source : str
Source to remove the ansi from | Remove ansi escape codes from text.
Parameters
----------
source : str
Source to remove the ansi from | [
"Remove",
"ansi",
"escape",
"codes",
"from",
"text",
".",
"Parameters",
"----------",
"source",
":",
"str",
"Source",
"to",
"remove",
"the",
"ansi",
"from"
] | def strip_ansi(source):
"""
Remove ansi escape codes from text.
Parameters
----------
source : str
Source to remove the ansi from
"""
return re.sub(r'\033\[(\d|;)+?m', '', source) | [
"def",
"strip_ansi",
"(",
"source",
")",
":",
"return",
"re",
".",
"sub",
"(",
"r'\\033\\[(\\d|;)+?m'",
",",
"''",
",",
"source",
")"
] | https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/ipython/py2/IPython/utils/text.py#L479-L488 |
|
ApolloAuto/apollo | 463fb82f9e979d02dcb25044e60931293ab2dba0 | modules/tools/sensor_calibration/extract_data.py | python | Extractor.generate_compressed_file | (input_path,
input_name,
output_path,
compressed_file='sensor_data') | Compress data extraction directory as a single tar.gz archive | Compress data extraction directory as a single tar.gz archive | [
"Compress",
"data",
"extraction",
"directory",
"as",
"a",
"single",
"tar",
".",
"gz",
"archive"
] | def generate_compressed_file(input_path,
input_name,
output_path,
compressed_file='sensor_data'):
"""
Compress data extraction directory as a single tar.gz archive
"""
cwd_path = os.getcwd()
os.chdir(input_path)
shutil.make_archive(base_name=os.path.join(output_path,
compressed_file),
format='gztar',
root_dir=input_path,
base_dir=input_name)
os.chdir(cwd_path) | [
"def",
"generate_compressed_file",
"(",
"input_path",
",",
"input_name",
",",
"output_path",
",",
"compressed_file",
"=",
"'sensor_data'",
")",
":",
"cwd_path",
"=",
"os",
".",
"getcwd",
"(",
")",
"os",
".",
"chdir",
"(",
"input_path",
")",
"shutil",
".",
"make_archive",
"(",
"base_name",
"=",
"os",
".",
"path",
".",
"join",
"(",
"output_path",
",",
"compressed_file",
")",
",",
"format",
"=",
"'gztar'",
",",
"root_dir",
"=",
"input_path",
",",
"base_dir",
"=",
"input_name",
")",
"os",
".",
"chdir",
"(",
"cwd_path",
")"
] | https://github.com/ApolloAuto/apollo/blob/463fb82f9e979d02dcb25044e60931293ab2dba0/modules/tools/sensor_calibration/extract_data.py#L323-L337 |
||
DmitryKoterov/dklab_realplexor | 01281d42fddcf7b9efe763b3ab50191c4429debc | api/python/Dklab/realplexor.py | python | Dklab_Realplexor.send | (self, idsAndCursors, data, showOnlyForIds=None) | Send data to realplexor.
Throw Dklab_Realplexor_Exception in case of error.
idsAndCursors -- Target IDs in form of: dictionary(id1 => cursor1, id2 => cursor2, ...)
of dictionary(id1, id2, id3, ...). If sending to a single ID,
you may pass it as a plain string, not dictionary.
data -- Data to be sent (any format, e.g. nested dictionaries are OK).
showOnlyForIds -- Send this message to only those who also listen any of these IDs.
This parameter may be used to limit the visibility to a closed
number of cliens: give each client an unique ID and enumerate
client IDs in $showOnlyForIds to not to send messages to others. | Send data to realplexor.
Throw Dklab_Realplexor_Exception in case of error. | [
"Send",
"data",
"to",
"realplexor",
".",
"Throw",
"Dklab_Realplexor_Exception",
"in",
"case",
"of",
"error",
"."
] | def send(self, idsAndCursors, data, showOnlyForIds=None):
"""
Send data to realplexor.
Throw Dklab_Realplexor_Exception in case of error.
idsAndCursors -- Target IDs in form of: dictionary(id1 => cursor1, id2 => cursor2, ...)
of dictionary(id1, id2, id3, ...). If sending to a single ID,
you may pass it as a plain string, not dictionary.
data -- Data to be sent (any format, e.g. nested dictionaries are OK).
showOnlyForIds -- Send this message to only those who also listen any of these IDs.
This parameter may be used to limit the visibility to a closed
number of cliens: give each client an unique ID and enumerate
client IDs in $showOnlyForIds to not to send messages to others.
"""
data = json.dumps(data)
pairs = []
for id in idsAndCursors:
if type(id) == type(1):
id = cursor # this is NOT cursor, but ID!
cursor = None
if re.search('^\w+$', id) is None:
raise Dklab_Realplexor_Exception("Identifier must be alphanumeric, \"%s\" given" % id)
try:
cursor = idsAndCursors[id]
except:
cursor = None
id = (self._namespace or '') + id
if cursor is not None:
try:
i = float(cursor)
except ValueError:
raise Dklab_Realplexor_Exception("Cursor must be numeric, \"%s\" given" % cursor)
pairs.append("%s:%s" % (cursor,id))
else:
pairs.append(id)
if isinstance(showOnlyForIds, (list, tuple)):
for id in showOnlyForIds:
pairs.append("*" + (self._namespace or '') + id)
self._send(",".join(pairs), data) | [
"def",
"send",
"(",
"self",
",",
"idsAndCursors",
",",
"data",
",",
"showOnlyForIds",
"=",
"None",
")",
":",
"data",
"=",
"json",
".",
"dumps",
"(",
"data",
")",
"pairs",
"=",
"[",
"]",
"for",
"id",
"in",
"idsAndCursors",
":",
"if",
"type",
"(",
"id",
")",
"==",
"type",
"(",
"1",
")",
":",
"id",
"=",
"cursor",
"# this is NOT cursor, but ID!",
"cursor",
"=",
"None",
"if",
"re",
".",
"search",
"(",
"'^\\w+$'",
",",
"id",
")",
"is",
"None",
":",
"raise",
"Dklab_Realplexor_Exception",
"(",
"\"Identifier must be alphanumeric, \\\"%s\\\" given\"",
"%",
"id",
")",
"try",
":",
"cursor",
"=",
"idsAndCursors",
"[",
"id",
"]",
"except",
":",
"cursor",
"=",
"None",
"id",
"=",
"(",
"self",
".",
"_namespace",
"or",
"''",
")",
"+",
"id",
"if",
"cursor",
"is",
"not",
"None",
":",
"try",
":",
"i",
"=",
"float",
"(",
"cursor",
")",
"except",
"ValueError",
":",
"raise",
"Dklab_Realplexor_Exception",
"(",
"\"Cursor must be numeric, \\\"%s\\\" given\"",
"%",
"cursor",
")",
"pairs",
".",
"append",
"(",
"\"%s:%s\"",
"%",
"(",
"cursor",
",",
"id",
")",
")",
"else",
":",
"pairs",
".",
"append",
"(",
"id",
")",
"if",
"isinstance",
"(",
"showOnlyForIds",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"for",
"id",
"in",
"showOnlyForIds",
":",
"pairs",
".",
"append",
"(",
"\"*\"",
"+",
"(",
"self",
".",
"_namespace",
"or",
"''",
")",
"+",
"id",
")",
"self",
".",
"_send",
"(",
"\",\"",
".",
"join",
"(",
"pairs",
")",
",",
"data",
")"
] | https://github.com/DmitryKoterov/dklab_realplexor/blob/01281d42fddcf7b9efe763b3ab50191c4429debc/api/python/Dklab/realplexor.py#L36-L74 |
End of preview. Expand
in Dataset Viewer.
No dataset card yet
New: Create and edit this dataset card directly on the website!
Contribute a Dataset Card- Downloads last month
- 36