id_within_dataset
int64 1
55.5k
| snippet
stringlengths 19
14.2k
| tokens
sequencelengths 6
1.63k
| nl
stringlengths 6
352
| split_within_dataset
stringclasses 1
value | is_duplicated
bool 2
classes |
---|---|---|---|---|---|
1 | def install_translator(qapp):
global QT_TRANSLATOR
if (QT_TRANSLATOR is None):
qt_translator = QTranslator()
if qt_translator.load(('qt_' + QLocale.system().name()), QLibraryInfo.location(QLibraryInfo.TranslationsPath)):
QT_TRANSLATOR = qt_translator
if (QT_TRANSLATOR is not None):
qapp.installTranslator(QT_TRANSLATOR)
| [
"def",
"install_translator",
"(",
"qapp",
")",
":",
"global",
"QT_TRANSLATOR",
"if",
"(",
"QT_TRANSLATOR",
"is",
"None",
")",
":",
"qt_translator",
"=",
"QTranslator",
"(",
")",
"if",
"qt_translator",
".",
"load",
"(",
"(",
"'qt_'",
"+",
"QLocale",
".",
"system",
"(",
")",
".",
"name",
"(",
")",
")",
",",
"QLibraryInfo",
".",
"location",
"(",
"QLibraryInfo",
".",
"TranslationsPath",
")",
")",
":",
"QT_TRANSLATOR",
"=",
"qt_translator",
"if",
"(",
"QT_TRANSLATOR",
"is",
"not",
"None",
")",
":",
"qapp",
".",
"installTranslator",
"(",
"QT_TRANSLATOR",
")"
] | install qt translator to the qapplication instance . | train | true |
2 | def delete_dhcp_options(dhcp_options_id=None, dhcp_options_name=None, region=None, key=None, keyid=None, profile=None):
return _delete_resource(resource='dhcp_options', name=dhcp_options_name, resource_id=dhcp_options_id, region=region, key=key, keyid=keyid, profile=profile)
| [
"def",
"delete_dhcp_options",
"(",
"dhcp_options_id",
"=",
"None",
",",
"dhcp_options_name",
"=",
"None",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"return",
"_delete_resource",
"(",
"resource",
"=",
"'dhcp_options'",
",",
"name",
"=",
"dhcp_options_name",
",",
"resource_id",
"=",
"dhcp_options_id",
",",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")"
] | delete dhcp options by id or name . | train | true |
4 | def test_depth_first_mro():
class A(object, ):
pass
class B(A, ):
pass
class C(A, ):
pass
class D(B, C, ):
pass
class E(D, object, ):
pass
class G(object, ):
pass
class H(G, ):
pass
class I(G, ):
pass
class K(H, I, object, ):
pass
class L(K, E, ):
pass
AreEqual(L.__mro__, (L, K, H, I, G, E, D, B, C, A, object))
| [
"def",
"test_depth_first_mro",
"(",
")",
":",
"class",
"A",
"(",
"object",
",",
")",
":",
"pass",
"class",
"B",
"(",
"A",
",",
")",
":",
"pass",
"class",
"C",
"(",
"A",
",",
")",
":",
"pass",
"class",
"D",
"(",
"B",
",",
"C",
",",
")",
":",
"pass",
"class",
"E",
"(",
"D",
",",
"object",
",",
")",
":",
"pass",
"class",
"G",
"(",
"object",
",",
")",
":",
"pass",
"class",
"H",
"(",
"G",
",",
")",
":",
"pass",
"class",
"I",
"(",
"G",
",",
")",
":",
"pass",
"class",
"K",
"(",
"H",
",",
"I",
",",
"object",
",",
")",
":",
"pass",
"class",
"L",
"(",
"K",
",",
"E",
",",
")",
":",
"pass",
"AreEqual",
"(",
"L",
".",
"__mro__",
",",
"(",
"L",
",",
"K",
",",
"H",
",",
"I",
",",
"G",
",",
"E",
",",
"D",
",",
"B",
",",
"C",
",",
"A",
",",
"object",
")",
")"
] | w/o old-style . | train | false |
5 | def mapping(data_source, geom_name='geom', layer_key=0, multi_geom=False):
if isinstance(data_source, six.string_types):
data_source = DataSource(data_source)
elif isinstance(data_source, DataSource):
pass
else:
raise TypeError('Data source parameter must be a string or a DataSource object.')
_mapping = {}
for field in data_source[layer_key].fields:
mfield = field.lower()
if (mfield[(-1):] == '_'):
mfield += 'field'
_mapping[mfield] = field
gtype = data_source[layer_key].geom_type
if (multi_geom and (gtype.num in (1, 2, 3))):
prefix = 'MULTI'
else:
prefix = ''
_mapping[geom_name] = (prefix + str(gtype).upper())
return _mapping
| [
"def",
"mapping",
"(",
"data_source",
",",
"geom_name",
"=",
"'geom'",
",",
"layer_key",
"=",
"0",
",",
"multi_geom",
"=",
"False",
")",
":",
"if",
"isinstance",
"(",
"data_source",
",",
"six",
".",
"string_types",
")",
":",
"data_source",
"=",
"DataSource",
"(",
"data_source",
")",
"elif",
"isinstance",
"(",
"data_source",
",",
"DataSource",
")",
":",
"pass",
"else",
":",
"raise",
"TypeError",
"(",
"'Data source parameter must be a string or a DataSource object.'",
")",
"_mapping",
"=",
"{",
"}",
"for",
"field",
"in",
"data_source",
"[",
"layer_key",
"]",
".",
"fields",
":",
"mfield",
"=",
"field",
".",
"lower",
"(",
")",
"if",
"(",
"mfield",
"[",
"(",
"-",
"1",
")",
":",
"]",
"==",
"'_'",
")",
":",
"mfield",
"+=",
"'field'",
"_mapping",
"[",
"mfield",
"]",
"=",
"field",
"gtype",
"=",
"data_source",
"[",
"layer_key",
"]",
".",
"geom_type",
"if",
"(",
"multi_geom",
"and",
"(",
"gtype",
".",
"num",
"in",
"(",
"1",
",",
"2",
",",
"3",
")",
")",
")",
":",
"prefix",
"=",
"'MULTI'",
"else",
":",
"prefix",
"=",
"''",
"_mapping",
"[",
"geom_name",
"]",
"=",
"(",
"prefix",
"+",
"str",
"(",
"gtype",
")",
".",
"upper",
"(",
")",
")",
"return",
"_mapping"
] | given a datasource . | train | false |
7 | def test_pprint_npfloat32():
dat = np.array([1.0, 2.0], dtype=np.float32)
t = Table([dat], names=['a'])
t['a'].format = '5.2f'
assert (str(t['a']) == ' a \n-----\n 1.00\n 2.00')
| [
"def",
"test_pprint_npfloat32",
"(",
")",
":",
"dat",
"=",
"np",
".",
"array",
"(",
"[",
"1.0",
",",
"2.0",
"]",
",",
"dtype",
"=",
"np",
".",
"float32",
")",
"t",
"=",
"Table",
"(",
"[",
"dat",
"]",
",",
"names",
"=",
"[",
"'a'",
"]",
")",
"t",
"[",
"'a'",
"]",
".",
"format",
"=",
"'5.2f'",
"assert",
"(",
"str",
"(",
"t",
"[",
"'a'",
"]",
")",
"==",
"' a \\n-----\\n 1.00\\n 2.00'",
")"
] | test for #148 . | train | false |
8 | def test_interpolation():
(t0, k0) = (0, np.array([5.0]))
results = _compute_fixed_length_solns(model, t0, k0)
for (integrator, numeric_solution) in results.items():
(N, T) = (1000, numeric_solution[:, 0][(-1)])
ti = np.linspace(t0, T, N)
interp_solution = model.interpolate(numeric_solution, ti, k=3, ext=2)
analytic_solution = solow_analytic_solution(ti, k0, *valid_params)
np.testing.assert_allclose(interp_solution, analytic_solution)
| [
"def",
"test_interpolation",
"(",
")",
":",
"(",
"t0",
",",
"k0",
")",
"=",
"(",
"0",
",",
"np",
".",
"array",
"(",
"[",
"5.0",
"]",
")",
")",
"results",
"=",
"_compute_fixed_length_solns",
"(",
"model",
",",
"t0",
",",
"k0",
")",
"for",
"(",
"integrator",
",",
"numeric_solution",
")",
"in",
"results",
".",
"items",
"(",
")",
":",
"(",
"N",
",",
"T",
")",
"=",
"(",
"1000",
",",
"numeric_solution",
"[",
":",
",",
"0",
"]",
"[",
"(",
"-",
"1",
")",
"]",
")",
"ti",
"=",
"np",
".",
"linspace",
"(",
"t0",
",",
"T",
",",
"N",
")",
"interp_solution",
"=",
"model",
".",
"interpolate",
"(",
"numeric_solution",
",",
"ti",
",",
"k",
"=",
"3",
",",
"ext",
"=",
"2",
")",
"analytic_solution",
"=",
"solow_analytic_solution",
"(",
"ti",
",",
"k0",
",",
"*",
"valid_params",
")",
"np",
".",
"testing",
".",
"assert_allclose",
"(",
"interp_solution",
",",
"analytic_solution",
")"
] | test interpolation option . | train | false |
9 | def _save_and_remove_module(name, orig_modules):
if (name not in sys.modules):
__import__(name)
del sys.modules[name]
for modname in list(sys.modules):
if ((modname == name) or modname.startswith((name + '.'))):
orig_modules[modname] = sys.modules[modname]
del sys.modules[modname]
| [
"def",
"_save_and_remove_module",
"(",
"name",
",",
"orig_modules",
")",
":",
"if",
"(",
"name",
"not",
"in",
"sys",
".",
"modules",
")",
":",
"__import__",
"(",
"name",
")",
"del",
"sys",
".",
"modules",
"[",
"name",
"]",
"for",
"modname",
"in",
"list",
"(",
"sys",
".",
"modules",
")",
":",
"if",
"(",
"(",
"modname",
"==",
"name",
")",
"or",
"modname",
".",
"startswith",
"(",
"(",
"name",
"+",
"'.'",
")",
")",
")",
":",
"orig_modules",
"[",
"modname",
"]",
"=",
"sys",
".",
"modules",
"[",
"modname",
"]",
"del",
"sys",
".",
"modules",
"[",
"modname",
"]"
] | helper function to save and remove a module from sys . | train | false |
12 | def url2ip(url):
iport = urlsplit(url)[1].split(':')
if (len(iport) > 1):
return (gethostbyname(iport[0]), iport[1])
return gethostbyname(iport[0])
| [
"def",
"url2ip",
"(",
"url",
")",
":",
"iport",
"=",
"urlsplit",
"(",
"url",
")",
"[",
"1",
"]",
".",
"split",
"(",
"':'",
")",
"if",
"(",
"len",
"(",
"iport",
")",
">",
"1",
")",
":",
"return",
"(",
"gethostbyname",
"(",
"iport",
"[",
"0",
"]",
")",
",",
"iport",
"[",
"1",
"]",
")",
"return",
"gethostbyname",
"(",
"iport",
"[",
"0",
"]",
")"
] | works like turning URL => 180 . | train | false |
13 | def _wait_until_running(instance):
with start_action(action_type=u'flocker:provision:aws:wait_until_running', instance_id=instance.id) as context:
_poll_while((lambda : _node_is_booting(instance)), repeat(1, INSTANCE_TIMEOUT))
context.add_success_fields(instance_state=instance.state)
context.add_success_fields(instance_state_reason=instance.state_reason)
if (instance.state != u'running'):
raise FailedToRun(instance.state_reason)
| [
"def",
"_wait_until_running",
"(",
"instance",
")",
":",
"with",
"start_action",
"(",
"action_type",
"=",
"u'flocker:provision:aws:wait_until_running'",
",",
"instance_id",
"=",
"instance",
".",
"id",
")",
"as",
"context",
":",
"_poll_while",
"(",
"(",
"lambda",
":",
"_node_is_booting",
"(",
"instance",
")",
")",
",",
"repeat",
"(",
"1",
",",
"INSTANCE_TIMEOUT",
")",
")",
"context",
".",
"add_success_fields",
"(",
"instance_state",
"=",
"instance",
".",
"state",
")",
"context",
".",
"add_success_fields",
"(",
"instance_state_reason",
"=",
"instance",
".",
"state_reason",
")",
"if",
"(",
"instance",
".",
"state",
"!=",
"u'running'",
")",
":",
"raise",
"FailedToRun",
"(",
"instance",
".",
"state_reason",
")"
] | wait until a instance is running . | train | false |
14 | def dict_to_numpy_array(d, mapping=None):
try:
return dict_to_numpy_array2(d, mapping)
except (AttributeError, TypeError):
return dict_to_numpy_array1(d, mapping)
| [
"def",
"dict_to_numpy_array",
"(",
"d",
",",
"mapping",
"=",
"None",
")",
":",
"try",
":",
"return",
"dict_to_numpy_array2",
"(",
"d",
",",
"mapping",
")",
"except",
"(",
"AttributeError",
",",
"TypeError",
")",
":",
"return",
"dict_to_numpy_array1",
"(",
"d",
",",
"mapping",
")"
] | convert a dictionary of dictionaries to a numpy array with optional mapping . | train | false |
15 | def _is_suggestion_handled(thread_id, exploration_id):
thread = feedback_models.FeedbackThreadModel.get_by_exp_and_thread_id(exploration_id, thread_id)
return (thread.status in [feedback_models.STATUS_CHOICES_FIXED, feedback_models.STATUS_CHOICES_IGNORED])
| [
"def",
"_is_suggestion_handled",
"(",
"thread_id",
",",
"exploration_id",
")",
":",
"thread",
"=",
"feedback_models",
".",
"FeedbackThreadModel",
".",
"get_by_exp_and_thread_id",
"(",
"exploration_id",
",",
"thread_id",
")",
"return",
"(",
"thread",
".",
"status",
"in",
"[",
"feedback_models",
".",
"STATUS_CHOICES_FIXED",
",",
"feedback_models",
".",
"STATUS_CHOICES_IGNORED",
"]",
")"
] | checks if the current suggestion has already been accepted/rejected . | train | false |
17 | def get_default_site(app_name='filebrowser'):
resolver = get_resolver(get_urlconf())
name = 'filebrowser'
app_list = resolver.app_dict[app_name]
if (name not in app_list):
name = app_list[0]
return get_site_dict()[name]
| [
"def",
"get_default_site",
"(",
"app_name",
"=",
"'filebrowser'",
")",
":",
"resolver",
"=",
"get_resolver",
"(",
"get_urlconf",
"(",
")",
")",
"name",
"=",
"'filebrowser'",
"app_list",
"=",
"resolver",
".",
"app_dict",
"[",
"app_name",
"]",
"if",
"(",
"name",
"not",
"in",
"app_list",
")",
":",
"name",
"=",
"app_list",
"[",
"0",
"]",
"return",
"get_site_dict",
"(",
")",
"[",
"name",
"]"
] | returns the default site . | train | false |
18 | def ccovf(x, y, unbiased=True, demean=True):
n = len(x)
if demean:
xo = (x - x.mean())
yo = (y - y.mean())
else:
xo = x
yo = y
if unbiased:
xi = np.ones(n)
d = np.correlate(xi, xi, 'full')
else:
d = n
return (np.correlate(xo, yo, 'full') / d)[(n - 1):]
| [
"def",
"ccovf",
"(",
"x",
",",
"y",
",",
"unbiased",
"=",
"True",
",",
"demean",
"=",
"True",
")",
":",
"n",
"=",
"len",
"(",
"x",
")",
"if",
"demean",
":",
"xo",
"=",
"(",
"x",
"-",
"x",
".",
"mean",
"(",
")",
")",
"yo",
"=",
"(",
"y",
"-",
"y",
".",
"mean",
"(",
")",
")",
"else",
":",
"xo",
"=",
"x",
"yo",
"=",
"y",
"if",
"unbiased",
":",
"xi",
"=",
"np",
".",
"ones",
"(",
"n",
")",
"d",
"=",
"np",
".",
"correlate",
"(",
"xi",
",",
"xi",
",",
"'full'",
")",
"else",
":",
"d",
"=",
"n",
"return",
"(",
"np",
".",
"correlate",
"(",
"xo",
",",
"yo",
",",
"'full'",
")",
"/",
"d",
")",
"[",
"(",
"n",
"-",
"1",
")",
":",
"]"
] | crosscovariance for 1d parameters x . | train | false |
20 | def serializers(**serializers):
def decorator(func):
if (not hasattr(func, 'wsgi_serializers')):
func.wsgi_serializers = {}
func.wsgi_serializers.update(serializers)
return func
return decorator
| [
"def",
"serializers",
"(",
"**",
"serializers",
")",
":",
"def",
"decorator",
"(",
"func",
")",
":",
"if",
"(",
"not",
"hasattr",
"(",
"func",
",",
"'wsgi_serializers'",
")",
")",
":",
"func",
".",
"wsgi_serializers",
"=",
"{",
"}",
"func",
".",
"wsgi_serializers",
".",
"update",
"(",
"serializers",
")",
"return",
"func",
"return",
"decorator"
] | returns the serializers modules . | train | false |
21 | def rgb2short(r, g, b):
incs = (0, 95, 135, 175, 215, 255)
parts = [r, g, b]
res = []
for part in parts:
i = 0
while (i < (len(incs) - 1)):
(s, b) = (incs[i], incs[(i + 1)])
if (s <= part <= b):
s1 = abs((s - part))
b1 = abs((b - part))
if (s1 < b1):
closest = s
else:
closest = b
res.append(closest)
break
i += 1
return RGB2SHORT_DICT[tuple(res)]
| [
"def",
"rgb2short",
"(",
"r",
",",
"g",
",",
"b",
")",
":",
"incs",
"=",
"(",
"0",
",",
"95",
",",
"135",
",",
"175",
",",
"215",
",",
"255",
")",
"parts",
"=",
"[",
"r",
",",
"g",
",",
"b",
"]",
"res",
"=",
"[",
"]",
"for",
"part",
"in",
"parts",
":",
"i",
"=",
"0",
"while",
"(",
"i",
"<",
"(",
"len",
"(",
"incs",
")",
"-",
"1",
")",
")",
":",
"(",
"s",
",",
"b",
")",
"=",
"(",
"incs",
"[",
"i",
"]",
",",
"incs",
"[",
"(",
"i",
"+",
"1",
")",
"]",
")",
"if",
"(",
"s",
"<=",
"part",
"<=",
"b",
")",
":",
"s1",
"=",
"abs",
"(",
"(",
"s",
"-",
"part",
")",
")",
"b1",
"=",
"abs",
"(",
"(",
"b",
"-",
"part",
")",
")",
"if",
"(",
"s1",
"<",
"b1",
")",
":",
"closest",
"=",
"s",
"else",
":",
"closest",
"=",
"b",
"res",
".",
"append",
"(",
"closest",
")",
"break",
"i",
"+=",
"1",
"return",
"RGB2SHORT_DICT",
"[",
"tuple",
"(",
"res",
")",
"]"
] | rgb to short . | train | true |
22 | def output():
return s3_rest_controller()
| [
"def",
"output",
"(",
")",
":",
"return",
"s3_rest_controller",
"(",
")"
] | output -> html string either return the result of a function or a sparse htmlized error message and a message in the server log . | train | false |
24 | def gf_factor_sqf(f, p, K, method=None):
(lc, f) = gf_monic(f, p, K)
if (gf_degree(f) < 1):
return (lc, [])
method = (method or query('GF_FACTOR_METHOD'))
if (method is not None):
factors = _factor_methods[method](f, p, K)
else:
factors = gf_zassenhaus(f, p, K)
return (lc, factors)
| [
"def",
"gf_factor_sqf",
"(",
"f",
",",
"p",
",",
"K",
",",
"method",
"=",
"None",
")",
":",
"(",
"lc",
",",
"f",
")",
"=",
"gf_monic",
"(",
"f",
",",
"p",
",",
"K",
")",
"if",
"(",
"gf_degree",
"(",
"f",
")",
"<",
"1",
")",
":",
"return",
"(",
"lc",
",",
"[",
"]",
")",
"method",
"=",
"(",
"method",
"or",
"query",
"(",
"'GF_FACTOR_METHOD'",
")",
")",
"if",
"(",
"method",
"is",
"not",
"None",
")",
":",
"factors",
"=",
"_factor_methods",
"[",
"method",
"]",
"(",
"f",
",",
"p",
",",
"K",
")",
"else",
":",
"factors",
"=",
"gf_zassenhaus",
"(",
"f",
",",
"p",
",",
"K",
")",
"return",
"(",
"lc",
",",
"factors",
")"
] | factor a square-free polynomial f in gf(p)[x] . | train | false |
26 | def draw_nx(G, pos, **kwds):
draw(G, pos, **kwds)
| [
"def",
"draw_nx",
"(",
"G",
",",
"pos",
",",
"**",
"kwds",
")",
":",
"draw",
"(",
"G",
",",
"pos",
",",
"**",
"kwds",
")"
] | for backward compatibility; use draw or draw_networkx . | train | false |
27 | def start_clientbrowser(config, args):
logger.info('Start client mode (browser)')
global client
from glances.client_browser import GlancesClientBrowser
client = GlancesClientBrowser(config=config, args=args)
client.serve_forever()
client.end()
| [
"def",
"start_clientbrowser",
"(",
"config",
",",
"args",
")",
":",
"logger",
".",
"info",
"(",
"'Start client mode (browser)'",
")",
"global",
"client",
"from",
"glances",
".",
"client_browser",
"import",
"GlancesClientBrowser",
"client",
"=",
"GlancesClientBrowser",
"(",
"config",
"=",
"config",
",",
"args",
"=",
"args",
")",
"client",
".",
"serve_forever",
"(",
")",
"client",
".",
"end",
"(",
")"
] | start the browser client mode . | train | false |
28 | @raises(ValueError)
def test_bootstrap_arglength():
algo.bootstrap(np.arange(5), np.arange(10))
| [
"@",
"raises",
"(",
"ValueError",
")",
"def",
"test_bootstrap_arglength",
"(",
")",
":",
"algo",
".",
"bootstrap",
"(",
"np",
".",
"arange",
"(",
"5",
")",
",",
"np",
".",
"arange",
"(",
"10",
")",
")"
] | test that different length args raise valueerror . | train | false |
29 | def runwsgi(func):
if os.environ.has_key('SERVER_SOFTWARE'):
os.environ['FCGI_FORCE_CGI'] = 'Y'
if (os.environ.has_key('PHP_FCGI_CHILDREN') or os.environ.has_key('SERVER_SOFTWARE')):
import flup.server.fcgi
return runfcgi(func)
if ('scgi' in sys.argv):
import flup.server.scgi
return runscgi(func)
return runsimple(func, listget(sys.argv, 1, 8080))
| [
"def",
"runwsgi",
"(",
"func",
")",
":",
"if",
"os",
".",
"environ",
".",
"has_key",
"(",
"'SERVER_SOFTWARE'",
")",
":",
"os",
".",
"environ",
"[",
"'FCGI_FORCE_CGI'",
"]",
"=",
"'Y'",
"if",
"(",
"os",
".",
"environ",
".",
"has_key",
"(",
"'PHP_FCGI_CHILDREN'",
")",
"or",
"os",
".",
"environ",
".",
"has_key",
"(",
"'SERVER_SOFTWARE'",
")",
")",
":",
"import",
"flup",
".",
"server",
".",
"fcgi",
"return",
"runfcgi",
"(",
"func",
")",
"if",
"(",
"'scgi'",
"in",
"sys",
".",
"argv",
")",
":",
"import",
"flup",
".",
"server",
".",
"scgi",
"return",
"runscgi",
"(",
"func",
")",
"return",
"runsimple",
"(",
"func",
",",
"listget",
"(",
"sys",
".",
"argv",
",",
"1",
",",
"8080",
")",
")"
] | runs a wsgi-compatible function using fcgi . | train | false |
30 | def random_bucket_name(prefix='awscli-s3integ-', num_random=10):
return (prefix + random_chars(num_random))
| [
"def",
"random_bucket_name",
"(",
"prefix",
"=",
"'awscli-s3integ-'",
",",
"num_random",
"=",
"10",
")",
":",
"return",
"(",
"prefix",
"+",
"random_chars",
"(",
"num_random",
")",
")"
] | generate a random s3 bucket name . | train | false |
31 | def _proxy_process(proxyname, test):
changes_old = []
changes_new = []
if (not _is_proxy_running(proxyname)):
if (not test):
__salt__['cmd.run_all']('salt-proxy --proxyid={0} -l info -d'.format(salt.ext.six.moves.shlex_quote(proxyname)), timeout=5)
changes_new.append('Salt Proxy: Started proxy process for {0}'.format(proxyname))
else:
changes_new.append('Salt Proxy: process {0} will be started'.format(proxyname))
else:
changes_old.append('Salt Proxy: already running for {0}'.format(proxyname))
return (True, changes_new, changes_old)
| [
"def",
"_proxy_process",
"(",
"proxyname",
",",
"test",
")",
":",
"changes_old",
"=",
"[",
"]",
"changes_new",
"=",
"[",
"]",
"if",
"(",
"not",
"_is_proxy_running",
"(",
"proxyname",
")",
")",
":",
"if",
"(",
"not",
"test",
")",
":",
"__salt__",
"[",
"'cmd.run_all'",
"]",
"(",
"'salt-proxy --proxyid={0} -l info -d'",
".",
"format",
"(",
"salt",
".",
"ext",
".",
"six",
".",
"moves",
".",
"shlex_quote",
"(",
"proxyname",
")",
")",
",",
"timeout",
"=",
"5",
")",
"changes_new",
".",
"append",
"(",
"'Salt Proxy: Started proxy process for {0}'",
".",
"format",
"(",
"proxyname",
")",
")",
"else",
":",
"changes_new",
".",
"append",
"(",
"'Salt Proxy: process {0} will be started'",
".",
"format",
"(",
"proxyname",
")",
")",
"else",
":",
"changes_old",
".",
"append",
"(",
"'Salt Proxy: already running for {0}'",
".",
"format",
"(",
"proxyname",
")",
")",
"return",
"(",
"True",
",",
"changes_new",
",",
"changes_old",
")"
] | check and execute proxy process . | train | true |
32 | def _offset_or_limit_clause(element, name=None, type_=None):
if (element is None):
return None
elif hasattr(element, '__clause_element__'):
return element.__clause_element__()
elif isinstance(element, Visitable):
return element
else:
value = util.asint(element)
return _OffsetLimitParam(name, value, type_=type_, unique=True)
| [
"def",
"_offset_or_limit_clause",
"(",
"element",
",",
"name",
"=",
"None",
",",
"type_",
"=",
"None",
")",
":",
"if",
"(",
"element",
"is",
"None",
")",
":",
"return",
"None",
"elif",
"hasattr",
"(",
"element",
",",
"'__clause_element__'",
")",
":",
"return",
"element",
".",
"__clause_element__",
"(",
")",
"elif",
"isinstance",
"(",
"element",
",",
"Visitable",
")",
":",
"return",
"element",
"else",
":",
"value",
"=",
"util",
".",
"asint",
"(",
"element",
")",
"return",
"_OffsetLimitParam",
"(",
"name",
",",
"value",
",",
"type_",
"=",
"type_",
",",
"unique",
"=",
"True",
")"
] | convert the given value to an "offset or limit" clause . | train | false |
34 | def _retrieve_device_config():
return __salt__['snmp.config']()
| [
"def",
"_retrieve_device_config",
"(",
")",
":",
"return",
"__salt__",
"[",
"'snmp.config'",
"]",
"(",
")"
] | retrieves the snmp config from the device . | train | false |
35 | def normalize_formset_dict(formset, attr_list):
assert isinstance(formset, BaseSimpleFormSet)
res = []
for form in formset.forms:
res.append(normalize_form_dict(form, attr_list))
return res
| [
"def",
"normalize_formset_dict",
"(",
"formset",
",",
"attr_list",
")",
":",
"assert",
"isinstance",
"(",
"formset",
",",
"BaseSimpleFormSet",
")",
"res",
"=",
"[",
"]",
"for",
"form",
"in",
"formset",
".",
"forms",
":",
"res",
".",
"append",
"(",
"normalize_form_dict",
"(",
"form",
",",
"attr_list",
")",
")",
"return",
"res"
] | normalize_formset_dict -> a list of dictionary of . | train | false |
36 | def parse_strtime(timestr, fmt=PERFECT_TIME_FORMAT):
return datetime.datetime.strptime(timestr, fmt)
| [
"def",
"parse_strtime",
"(",
"timestr",
",",
"fmt",
"=",
"PERFECT_TIME_FORMAT",
")",
":",
"return",
"datetime",
".",
"datetime",
".",
"strptime",
"(",
"timestr",
",",
"fmt",
")"
] | turn a formatted time back into a datetime . | train | false |
37 | def hash_filehash(filename):
md4 = hashlib.new(u'md4').copy
def gen(f):
while True:
x = f.read(9728000)
if x:
(yield x)
else:
return
def md4_hash(data):
m = md4()
m.update(data)
return m
with open(filename, u'rb') as f:
a = gen(f)
hashes = [md4_hash(data).digest() for data in a]
if (len(hashes) == 1):
return to_hex(hashes[0])
else:
return md4_hash(reduce((lambda a, d: (a + d)), hashes, u'')).hexd
| [
"def",
"hash_filehash",
"(",
"filename",
")",
":",
"md4",
"=",
"hashlib",
".",
"new",
"(",
"u'md4'",
")",
".",
"copy",
"def",
"gen",
"(",
"f",
")",
":",
"while",
"True",
":",
"x",
"=",
"f",
".",
"read",
"(",
"9728000",
")",
"if",
"x",
":",
"(",
"yield",
"x",
")",
"else",
":",
"return",
"def",
"md4_hash",
"(",
"data",
")",
":",
"m",
"=",
"md4",
"(",
")",
"m",
".",
"update",
"(",
"data",
")",
"return",
"m",
"with",
"open",
"(",
"filename",
",",
"u'rb'",
")",
"as",
"f",
":",
"a",
"=",
"gen",
"(",
"f",
")",
"hashes",
"=",
"[",
"md4_hash",
"(",
"data",
")",
".",
"digest",
"(",
")",
"for",
"data",
"in",
"a",
"]",
"if",
"(",
"len",
"(",
"hashes",
")",
"==",
"1",
")",
":",
"return",
"to_hex",
"(",
"hashes",
"[",
"0",
"]",
")",
"else",
":",
"return",
"md4_hash",
"(",
"reduce",
"(",
"(",
"lambda",
"a",
",",
"d",
":",
"(",
"a",
"+",
"d",
")",
")",
",",
"hashes",
",",
"u''",
")",
")",
".",
"hexd"
] | returns the ed2k hash of a given file . | train | false |
39 | @receiver(user_logged_in)
def log_successful_login(sender, request, user, **kwargs):
if settings.FEATURES['SQUELCH_PII_IN_LOGS']:
AUDIT_LOG.info(u'Login success - user.id: {0}'.format(user.id))
else:
AUDIT_LOG.info(u'Login success - {0} ({1})'.format(user.username, user.email))
| [
"@",
"receiver",
"(",
"user_logged_in",
")",
"def",
"log_successful_login",
"(",
"sender",
",",
"request",
",",
"user",
",",
"**",
"kwargs",
")",
":",
"if",
"settings",
".",
"FEATURES",
"[",
"'SQUELCH_PII_IN_LOGS'",
"]",
":",
"AUDIT_LOG",
".",
"info",
"(",
"u'Login success - user.id: {0}'",
".",
"format",
"(",
"user",
".",
"id",
")",
")",
"else",
":",
"AUDIT_LOG",
".",
"info",
"(",
"u'Login success - {0} ({1})'",
".",
"format",
"(",
"user",
".",
"username",
",",
"user",
".",
"email",
")",
")"
] | handler to log when logins have occurred successfully . | train | false |
43 | def zoom_effect02(ax1, ax2, **kwargs):
tt = (ax1.transScale + (ax1.transLimits + ax2.transAxes))
trans = blended_transform_factory(ax2.transData, tt)
mybbox1 = ax1.bbox
mybbox2 = TransformedBbox(ax1.viewLim, trans)
prop_patches = kwargs.copy()
prop_patches['ec'] = 'none'
prop_patches['alpha'] = 0.2
(c1, c2, bbox_patch1, bbox_patch2, p) = connect_bbox(mybbox1, mybbox2, loc1a=3, loc2a=2, loc1b=4, loc2b=1, prop_lines=kwargs, prop_patches=prop_patches)
ax1.add_patch(bbox_patch1)
ax2.add_patch(bbox_patch2)
ax2.add_patch(c1)
ax2.add_patch(c2)
ax2.add_patch(p)
return (c1, c2, bbox_patch1, bbox_patch2, p)
| [
"def",
"zoom_effect02",
"(",
"ax1",
",",
"ax2",
",",
"**",
"kwargs",
")",
":",
"tt",
"=",
"(",
"ax1",
".",
"transScale",
"+",
"(",
"ax1",
".",
"transLimits",
"+",
"ax2",
".",
"transAxes",
")",
")",
"trans",
"=",
"blended_transform_factory",
"(",
"ax2",
".",
"transData",
",",
"tt",
")",
"mybbox1",
"=",
"ax1",
".",
"bbox",
"mybbox2",
"=",
"TransformedBbox",
"(",
"ax1",
".",
"viewLim",
",",
"trans",
")",
"prop_patches",
"=",
"kwargs",
".",
"copy",
"(",
")",
"prop_patches",
"[",
"'ec'",
"]",
"=",
"'none'",
"prop_patches",
"[",
"'alpha'",
"]",
"=",
"0.2",
"(",
"c1",
",",
"c2",
",",
"bbox_patch1",
",",
"bbox_patch2",
",",
"p",
")",
"=",
"connect_bbox",
"(",
"mybbox1",
",",
"mybbox2",
",",
"loc1a",
"=",
"3",
",",
"loc2a",
"=",
"2",
",",
"loc1b",
"=",
"4",
",",
"loc2b",
"=",
"1",
",",
"prop_lines",
"=",
"kwargs",
",",
"prop_patches",
"=",
"prop_patches",
")",
"ax1",
".",
"add_patch",
"(",
"bbox_patch1",
")",
"ax2",
".",
"add_patch",
"(",
"bbox_patch2",
")",
"ax2",
".",
"add_patch",
"(",
"c1",
")",
"ax2",
".",
"add_patch",
"(",
"c2",
")",
"ax2",
".",
"add_patch",
"(",
"p",
")",
"return",
"(",
"c1",
",",
"c2",
",",
"bbox_patch1",
",",
"bbox_patch2",
",",
"p",
")"
] | ax1 : the main axes ax1 : the zoomed axes similar to zoom_effect01 . | train | false |
44 | def _expand_table(table):
return np.repeat([[1, 1], [1, 0], [0, 1], [0, 0]], table.ravel(), axis=0)
| [
"def",
"_expand_table",
"(",
"table",
")",
":",
"return",
"np",
".",
"repeat",
"(",
"[",
"[",
"1",
",",
"1",
"]",
",",
"[",
"1",
",",
"0",
"]",
",",
"[",
"0",
",",
"1",
"]",
",",
"[",
"0",
",",
"0",
"]",
"]",
",",
"table",
".",
"ravel",
"(",
")",
",",
"axis",
"=",
"0",
")"
] | expand a 2 by 2 contingency table to observations . | train | false |
45 | def load_extra_data(backend, details, response, uid, user, social_user=None, *args, **kwargs):
social_user = (social_user or UserSocialAuth.get_social_auth(backend.name, uid, user))
if social_user:
extra_data = backend.extra_data(user, uid, response, details)
if (kwargs.get('original_email') and ('email' not in extra_data)):
extra_data['email'] = kwargs.get('original_email')
if (extra_data and (social_user.extra_data != extra_data)):
if social_user.extra_data:
social_user.extra_data.update(extra_data)
else:
social_user.extra_data = extra_data
social_user.save()
return {'social_user': social_user}
| [
"def",
"load_extra_data",
"(",
"backend",
",",
"details",
",",
"response",
",",
"uid",
",",
"user",
",",
"social_user",
"=",
"None",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"social_user",
"=",
"(",
"social_user",
"or",
"UserSocialAuth",
".",
"get_social_auth",
"(",
"backend",
".",
"name",
",",
"uid",
",",
"user",
")",
")",
"if",
"social_user",
":",
"extra_data",
"=",
"backend",
".",
"extra_data",
"(",
"user",
",",
"uid",
",",
"response",
",",
"details",
")",
"if",
"(",
"kwargs",
".",
"get",
"(",
"'original_email'",
")",
"and",
"(",
"'email'",
"not",
"in",
"extra_data",
")",
")",
":",
"extra_data",
"[",
"'email'",
"]",
"=",
"kwargs",
".",
"get",
"(",
"'original_email'",
")",
"if",
"(",
"extra_data",
"and",
"(",
"social_user",
".",
"extra_data",
"!=",
"extra_data",
")",
")",
":",
"if",
"social_user",
".",
"extra_data",
":",
"social_user",
".",
"extra_data",
".",
"update",
"(",
"extra_data",
")",
"else",
":",
"social_user",
".",
"extra_data",
"=",
"extra_data",
"social_user",
".",
"save",
"(",
")",
"return",
"{",
"'social_user'",
":",
"social_user",
"}"
] | load extra data from provider and store it on current usersocialauth extra_data field . | train | false |
46 | def arbitrary(module_name, func_name, args, kwargs={}):
if module_name.startswith('calibre_plugins'):
from calibre.customize.ui import find_plugin
find_plugin
module = importlib.import_module(module_name)
func = getattr(module, func_name)
return func(*args, **kwargs)
| [
"def",
"arbitrary",
"(",
"module_name",
",",
"func_name",
",",
"args",
",",
"kwargs",
"=",
"{",
"}",
")",
":",
"if",
"module_name",
".",
"startswith",
"(",
"'calibre_plugins'",
")",
":",
"from",
"calibre",
".",
"customize",
".",
"ui",
"import",
"find_plugin",
"find_plugin",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"module_name",
")",
"func",
"=",
"getattr",
"(",
"module",
",",
"func_name",
")",
"return",
"func",
"(",
"*",
"args",
",",
"**",
"kwargs",
")"
] | an entry point that allows arbitrary functions to be run in a parallel process . | train | false |
47 | def claModelControlDisableTPLearningCb(claModel):
assert isinstance(claModel, CLAModel)
claModel._getTPRegion().setParameter('learningMode', False)
return
| [
"def",
"claModelControlDisableTPLearningCb",
"(",
"claModel",
")",
":",
"assert",
"isinstance",
"(",
"claModel",
",",
"CLAModel",
")",
"claModel",
".",
"_getTPRegion",
"(",
")",
".",
"setParameter",
"(",
"'learningMode'",
",",
"False",
")",
"return"
] | disables learning in the cla models temporal pooler . | train | false |
48 | def create_api_deployment(restApiId, stageName, stageDescription='', description='', cacheClusterEnabled=False, cacheClusterSize='0.5', variables=None, region=None, key=None, keyid=None, profile=None):
try:
variables = (dict() if (variables is None) else variables)
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
deployment = conn.create_deployment(restApiId=restApiId, stageName=stageName, stageDescription=stageDescription, description=description, cacheClusterEnabled=cacheClusterEnabled, cacheClusterSize=cacheClusterSize, variables=variables)
return {'created': True, 'deployment': _convert_datetime_str(deployment)}
except ClientError as e:
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
| [
"def",
"create_api_deployment",
"(",
"restApiId",
",",
"stageName",
",",
"stageDescription",
"=",
"''",
",",
"description",
"=",
"''",
",",
"cacheClusterEnabled",
"=",
"False",
",",
"cacheClusterSize",
"=",
"'0.5'",
",",
"variables",
"=",
"None",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"try",
":",
"variables",
"=",
"(",
"dict",
"(",
")",
"if",
"(",
"variables",
"is",
"None",
")",
"else",
"variables",
")",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")",
"deployment",
"=",
"conn",
".",
"create_deployment",
"(",
"restApiId",
"=",
"restApiId",
",",
"stageName",
"=",
"stageName",
",",
"stageDescription",
"=",
"stageDescription",
",",
"description",
"=",
"description",
",",
"cacheClusterEnabled",
"=",
"cacheClusterEnabled",
",",
"cacheClusterSize",
"=",
"cacheClusterSize",
",",
"variables",
"=",
"variables",
")",
"return",
"{",
"'created'",
":",
"True",
",",
"'deployment'",
":",
"_convert_datetime_str",
"(",
"deployment",
")",
"}",
"except",
"ClientError",
"as",
"e",
":",
"return",
"{",
"'created'",
":",
"False",
",",
"'error'",
":",
"salt",
".",
"utils",
".",
"boto3",
".",
"get_error",
"(",
"e",
")",
"}"
] | creates a new api deployment . | train | true |
49 | def _point_along_a_line(x0, y0, x1, y1, d):
(dx, dy) = ((x0 - x1), (y0 - y1))
ff = (d / (((dx * dx) + (dy * dy)) ** 0.5))
(x2, y2) = ((x0 - (ff * dx)), (y0 - (ff * dy)))
return (x2, y2)
| [
"def",
"_point_along_a_line",
"(",
"x0",
",",
"y0",
",",
"x1",
",",
"y1",
",",
"d",
")",
":",
"(",
"dx",
",",
"dy",
")",
"=",
"(",
"(",
"x0",
"-",
"x1",
")",
",",
"(",
"y0",
"-",
"y1",
")",
")",
"ff",
"=",
"(",
"d",
"/",
"(",
"(",
"(",
"dx",
"*",
"dx",
")",
"+",
"(",
"dy",
"*",
"dy",
")",
")",
"**",
"0.5",
")",
")",
"(",
"x2",
",",
"y2",
")",
"=",
"(",
"(",
"x0",
"-",
"(",
"ff",
"*",
"dx",
")",
")",
",",
"(",
"y0",
"-",
"(",
"ff",
"*",
"dy",
")",
")",
")",
"return",
"(",
"x2",
",",
"y2",
")"
] | find a point along a line connecting -- whose distance from is d . | train | false |
50 | def s3_roles_permitted(name='roles_permitted', **attr):
T = current.T
represent = S3Represent(lookup='auth_group', fields=['role'])
if ('label' not in attr):
attr['label'] = T('Roles Permitted')
if ('sortby' not in attr):
attr['sortby'] = 'role'
if ('represent' not in attr):
attr['represent'] = represent
if ('requires' not in attr):
attr['requires'] = IS_EMPTY_OR(IS_ONE_OF(current.db, 'auth_group.id', represent, multiple=True))
if ('comment' not in attr):
attr['comment'] = DIV(_class='tooltip', _title=('%s|%s' % (T('Roles Permitted'), T('If this record should be restricted then select which role(s) are permitted to access the record here.'))))
if ('ondelete' not in attr):
attr['ondelete'] = 'RESTRICT'
f = S3ReusableField(name, 'list:reference auth_group', **attr)
return f()
| [
"def",
"s3_roles_permitted",
"(",
"name",
"=",
"'roles_permitted'",
",",
"**",
"attr",
")",
":",
"T",
"=",
"current",
".",
"T",
"represent",
"=",
"S3Represent",
"(",
"lookup",
"=",
"'auth_group'",
",",
"fields",
"=",
"[",
"'role'",
"]",
")",
"if",
"(",
"'label'",
"not",
"in",
"attr",
")",
":",
"attr",
"[",
"'label'",
"]",
"=",
"T",
"(",
"'Roles Permitted'",
")",
"if",
"(",
"'sortby'",
"not",
"in",
"attr",
")",
":",
"attr",
"[",
"'sortby'",
"]",
"=",
"'role'",
"if",
"(",
"'represent'",
"not",
"in",
"attr",
")",
":",
"attr",
"[",
"'represent'",
"]",
"=",
"represent",
"if",
"(",
"'requires'",
"not",
"in",
"attr",
")",
":",
"attr",
"[",
"'requires'",
"]",
"=",
"IS_EMPTY_OR",
"(",
"IS_ONE_OF",
"(",
"current",
".",
"db",
",",
"'auth_group.id'",
",",
"represent",
",",
"multiple",
"=",
"True",
")",
")",
"if",
"(",
"'comment'",
"not",
"in",
"attr",
")",
":",
"attr",
"[",
"'comment'",
"]",
"=",
"DIV",
"(",
"_class",
"=",
"'tooltip'",
",",
"_title",
"=",
"(",
"'%s|%s'",
"%",
"(",
"T",
"(",
"'Roles Permitted'",
")",
",",
"T",
"(",
"'If this record should be restricted then select which role(s) are permitted to access the record here.'",
")",
")",
")",
")",
"if",
"(",
"'ondelete'",
"not",
"in",
"attr",
")",
":",
"attr",
"[",
"'ondelete'",
"]",
"=",
"'RESTRICT'",
"f",
"=",
"S3ReusableField",
"(",
"name",
",",
"'list:reference auth_group'",
",",
"**",
"attr",
")",
"return",
"f",
"(",
")"
] | list of roles permitted to access a resource - used by cms . | train | false |
52 | @pick_context_manager_writer
def instance_group_update(context, group_uuid, values):
group = model_query(context, models.InstanceGroup).filter_by(uuid=group_uuid).first()
if (not group):
raise exception.InstanceGroupNotFound(group_uuid=group_uuid)
policies = values.get('policies')
if (policies is not None):
_instance_group_policies_add(context, group.id, values.pop('policies'), set_delete=True)
members = values.get('members')
if (members is not None):
_instance_group_members_add(context, group.id, values.pop('members'), set_delete=True)
group.update(values)
if policies:
values['policies'] = policies
if members:
values['members'] = members
| [
"@",
"pick_context_manager_writer",
"def",
"instance_group_update",
"(",
"context",
",",
"group_uuid",
",",
"values",
")",
":",
"group",
"=",
"model_query",
"(",
"context",
",",
"models",
".",
"InstanceGroup",
")",
".",
"filter_by",
"(",
"uuid",
"=",
"group_uuid",
")",
".",
"first",
"(",
")",
"if",
"(",
"not",
"group",
")",
":",
"raise",
"exception",
".",
"InstanceGroupNotFound",
"(",
"group_uuid",
"=",
"group_uuid",
")",
"policies",
"=",
"values",
".",
"get",
"(",
"'policies'",
")",
"if",
"(",
"policies",
"is",
"not",
"None",
")",
":",
"_instance_group_policies_add",
"(",
"context",
",",
"group",
".",
"id",
",",
"values",
".",
"pop",
"(",
"'policies'",
")",
",",
"set_delete",
"=",
"True",
")",
"members",
"=",
"values",
".",
"get",
"(",
"'members'",
")",
"if",
"(",
"members",
"is",
"not",
"None",
")",
":",
"_instance_group_members_add",
"(",
"context",
",",
"group",
".",
"id",
",",
"values",
".",
"pop",
"(",
"'members'",
")",
",",
"set_delete",
"=",
"True",
")",
"group",
".",
"update",
"(",
"values",
")",
"if",
"policies",
":",
"values",
"[",
"'policies'",
"]",
"=",
"policies",
"if",
"members",
":",
"values",
"[",
"'members'",
"]",
"=",
"members"
] | update the attributes of an group . | train | false |
53 | def _tree_to_bitstrs(tree):
clades_bitstrs = {}
term_names = [term.name for term in tree.find_clades(terminal=True)]
for clade in tree.find_clades(terminal=False):
bitstr = _clade_to_bitstr(clade, term_names)
clades_bitstrs[clade] = bitstr
return clades_bitstrs
| [
"def",
"_tree_to_bitstrs",
"(",
"tree",
")",
":",
"clades_bitstrs",
"=",
"{",
"}",
"term_names",
"=",
"[",
"term",
".",
"name",
"for",
"term",
"in",
"tree",
".",
"find_clades",
"(",
"terminal",
"=",
"True",
")",
"]",
"for",
"clade",
"in",
"tree",
".",
"find_clades",
"(",
"terminal",
"=",
"False",
")",
":",
"bitstr",
"=",
"_clade_to_bitstr",
"(",
"clade",
",",
"term_names",
")",
"clades_bitstrs",
"[",
"clade",
"]",
"=",
"bitstr",
"return",
"clades_bitstrs"
] | create a dict of a trees clades to corresponding bitstrings . | train | false |
54 | def timeuntil(d, now=None):
return timesince(d, now, reversed=True)
| [
"def",
"timeuntil",
"(",
"d",
",",
"now",
"=",
"None",
")",
":",
"return",
"timesince",
"(",
"d",
",",
"now",
",",
"reversed",
"=",
"True",
")"
] | formats a date as the time until that date . | train | false |
56 | def test_hashbang():
entry = tokenize('#!this is a comment\n')
assert (entry == [])
| [
"def",
"test_hashbang",
"(",
")",
":",
"entry",
"=",
"tokenize",
"(",
"'#!this is a comment\\n'",
")",
"assert",
"(",
"entry",
"==",
"[",
"]",
")"
] | ensure we can escape things . | train | false |
59 | @context.quietfunc
@with_device
def exists(path):
with AdbClient() as c:
return bool(c.stat(path))
| [
"@",
"context",
".",
"quietfunc",
"@",
"with_device",
"def",
"exists",
"(",
"path",
")",
":",
"with",
"AdbClient",
"(",
")",
"as",
"c",
":",
"return",
"bool",
"(",
"c",
".",
"stat",
"(",
"path",
")",
")"
] | check if a user exists . | train | false |
60 | def disassociate_api_key_stagekeys(apiKey, stagekeyslist, region=None, key=None, keyid=None, profile=None):
try:
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
pvlist = [('/stages', stagekey) for stagekey in stagekeyslist]
response = _api_key_patch_remove(conn, apiKey, pvlist)
return {'disassociated': True}
except ClientError as e:
return {'disassociated': False, 'error': salt.utils.boto3.get_error(e)}
| [
"def",
"disassociate_api_key_stagekeys",
"(",
"apiKey",
",",
"stagekeyslist",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"try",
":",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")",
"pvlist",
"=",
"[",
"(",
"'/stages'",
",",
"stagekey",
")",
"for",
"stagekey",
"in",
"stagekeyslist",
"]",
"response",
"=",
"_api_key_patch_remove",
"(",
"conn",
",",
"apiKey",
",",
"pvlist",
")",
"return",
"{",
"'disassociated'",
":",
"True",
"}",
"except",
"ClientError",
"as",
"e",
":",
"return",
"{",
"'disassociated'",
":",
"False",
",",
"'error'",
":",
"salt",
".",
"utils",
".",
"boto3",
".",
"get_error",
"(",
"e",
")",
"}"
] | disassociate the given stagekeyslist to the given apikey . | train | false |
61 | @then(u'we see database dropped')
def step_see_db_dropped(context):
_expect_exact(context, u'DROP DATABASE', timeout=2)
| [
"@",
"then",
"(",
"u'we see database dropped'",
")",
"def",
"step_see_db_dropped",
"(",
"context",
")",
":",
"_expect_exact",
"(",
"context",
",",
"u'DROP DATABASE'",
",",
"timeout",
"=",
"2",
")"
] | wait to see drop database output . | train | false |
62 | @bdd.when(bdd.parsers.parse('I wait for the javascript message "{message}"'))
def javascript_message_when(quteproc, message):
quteproc.wait_for_js(message)
| [
"@",
"bdd",
".",
"when",
"(",
"bdd",
".",
"parsers",
".",
"parse",
"(",
"'I wait for the javascript message \"{message}\"'",
")",
")",
"def",
"javascript_message_when",
"(",
"quteproc",
",",
"message",
")",
":",
"quteproc",
".",
"wait_for_js",
"(",
"message",
")"
] | make sure the given message was logged via javascript . | train | false |
63 | def _gitPresent():
try:
gitvers = subprocess.check_output('git --version'.split(), stderr=subprocess.PIPE)
except (CalledProcessError, OSError):
gitvers = ''
return bool(gitvers.startswith('git version'))
| [
"def",
"_gitPresent",
"(",
")",
":",
"try",
":",
"gitvers",
"=",
"subprocess",
".",
"check_output",
"(",
"'git --version'",
".",
"split",
"(",
")",
",",
"stderr",
"=",
"subprocess",
".",
"PIPE",
")",
"except",
"(",
"CalledProcessError",
",",
"OSError",
")",
":",
"gitvers",
"=",
"''",
"return",
"bool",
"(",
"gitvers",
".",
"startswith",
"(",
"'git version'",
")",
")"
] | check for git on command-line . | train | false |
64 | def create_mount_target(filesystemid, subnetid, ipaddress=None, securitygroups=None, keyid=None, key=None, profile=None, region=None, **kwargs):
client = _get_conn(key=key, keyid=keyid, profile=profile, region=region)
return client.create_mount_point(FileSystemId=filesystemid, SubnetId=subnetid, IpAddress=ipaddress, SecurityGroups=securitygroups)
| [
"def",
"create_mount_target",
"(",
"filesystemid",
",",
"subnetid",
",",
"ipaddress",
"=",
"None",
",",
"securitygroups",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"key",
"=",
"None",
",",
"profile",
"=",
"None",
",",
"region",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"client",
"=",
"_get_conn",
"(",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
",",
"region",
"=",
"region",
")",
"return",
"client",
".",
"create_mount_point",
"(",
"FileSystemId",
"=",
"filesystemid",
",",
"SubnetId",
"=",
"subnetid",
",",
"IpAddress",
"=",
"ipaddress",
",",
"SecurityGroups",
"=",
"securitygroups",
")"
] | creates a mount target for a file system . | train | false |
65 | def IsAutoGenerated(xml_str):
try:
xml_root = ElementTree.fromstring(xml_str)
return ((xml_root.tag == 'datastore-indexes') and _BooleanAttribute(xml_root.attrib.get('autoGenerate', 'false')))
except ElementTree.ParseError:
return False
| [
"def",
"IsAutoGenerated",
"(",
"xml_str",
")",
":",
"try",
":",
"xml_root",
"=",
"ElementTree",
".",
"fromstring",
"(",
"xml_str",
")",
"return",
"(",
"(",
"xml_root",
".",
"tag",
"==",
"'datastore-indexes'",
")",
"and",
"_BooleanAttribute",
"(",
"xml_root",
".",
"attrib",
".",
"get",
"(",
"'autoGenerate'",
",",
"'false'",
")",
")",
")",
"except",
"ElementTree",
".",
"ParseError",
":",
"return",
"False"
] | test if the given datastore-indexes . | train | false |
66 | def jnp_zeros(n, nt):
return jnyn_zeros(n, nt)[1]
| [
"def",
"jnp_zeros",
"(",
"n",
",",
"nt",
")",
":",
"return",
"jnyn_zeros",
"(",
"n",
",",
"nt",
")",
"[",
"1",
"]"
] | compute zeros of integer-order bessel function derivative jn(x) . | train | false |
68 | def set_time(time):
time_format = _get_date_time_format(time)
dt_obj = datetime.strptime(time, time_format)
cmd = 'systemsetup -settime {0}'.format(dt_obj.strftime('%H:%M:%S'))
return salt.utils.mac_utils.execute_return_success(cmd)
| [
"def",
"set_time",
"(",
"time",
")",
":",
"time_format",
"=",
"_get_date_time_format",
"(",
"time",
")",
"dt_obj",
"=",
"datetime",
".",
"strptime",
"(",
"time",
",",
"time_format",
")",
"cmd",
"=",
"'systemsetup -settime {0}'",
".",
"format",
"(",
"dt_obj",
".",
"strftime",
"(",
"'%H:%M:%S'",
")",
")",
"return",
"salt",
".",
"utils",
".",
"mac_utils",
".",
"execute_return_success",
"(",
"cmd",
")"
] | sets the current time . | train | true |
69 | def del_job_files(job_paths):
for path in job_paths:
if (path and clip_path(path).lower().startswith(cfg.download_dir.get_path().lower())):
remove_all(path, recursive=True)
| [
"def",
"del_job_files",
"(",
"job_paths",
")",
":",
"for",
"path",
"in",
"job_paths",
":",
"if",
"(",
"path",
"and",
"clip_path",
"(",
"path",
")",
".",
"lower",
"(",
")",
".",
"startswith",
"(",
"cfg",
".",
"download_dir",
".",
"get_path",
"(",
")",
".",
"lower",
"(",
")",
")",
")",
":",
"remove_all",
"(",
"path",
",",
"recursive",
"=",
"True",
")"
] | remove files of each path in the list . | train | false |
70 | def get_discount_modules():
return load_module_instances('SHUUP_DISCOUNT_MODULES', 'discount_module')
| [
"def",
"get_discount_modules",
"(",
")",
":",
"return",
"load_module_instances",
"(",
"'SHUUP_DISCOUNT_MODULES'",
",",
"'discount_module'",
")"
] | get a list of configured discount module instances . | train | false |
73 | def initialize_log_data(ids_bcs_added_field):
log_data = {}
for curr_key in ids_bcs_added_field.keys():
base_key = ''
if curr_key[0]:
base_key += (curr_key[0] + ',')
if curr_key[1]:
base_key += (curr_key[1] + ',')
base_key += ids_bcs_added_field[curr_key]
log_data[base_key] = 0
return log_data
| [
"def",
"initialize_log_data",
"(",
"ids_bcs_added_field",
")",
":",
"log_data",
"=",
"{",
"}",
"for",
"curr_key",
"in",
"ids_bcs_added_field",
".",
"keys",
"(",
")",
":",
"base_key",
"=",
"''",
"if",
"curr_key",
"[",
"0",
"]",
":",
"base_key",
"+=",
"(",
"curr_key",
"[",
"0",
"]",
"+",
"','",
")",
"if",
"curr_key",
"[",
"1",
"]",
":",
"base_key",
"+=",
"(",
"curr_key",
"[",
"1",
"]",
"+",
"','",
")",
"base_key",
"+=",
"ids_bcs_added_field",
"[",
"curr_key",
"]",
"log_data",
"[",
"base_key",
"]",
"=",
"0",
"return",
"log_data"
] | initializes log data . | train | false |
76 | def is_sequence_of_strings(obj):
if (not cbook.iterable(obj)):
return False
if ((not isinstance(obj, np.ndarray)) and cbook.is_string_like(obj)):
return False
for o in obj:
if (not cbook.is_string_like(o)):
return False
return True
| [
"def",
"is_sequence_of_strings",
"(",
"obj",
")",
":",
"if",
"(",
"not",
"cbook",
".",
"iterable",
"(",
"obj",
")",
")",
":",
"return",
"False",
"if",
"(",
"(",
"not",
"isinstance",
"(",
"obj",
",",
"np",
".",
"ndarray",
")",
")",
"and",
"cbook",
".",
"is_string_like",
"(",
"obj",
")",
")",
":",
"return",
"False",
"for",
"o",
"in",
"obj",
":",
"if",
"(",
"not",
"cbook",
".",
"is_string_like",
"(",
"o",
")",
")",
":",
"return",
"False",
"return",
"True"
] | returns true if *obj* is iterable and contains strings . | train | false |
77 | def highlighting(view, name, style, left, right):
if (left is not None):
left = left.move((left.begin + BEGIN_LEN), (left.end - BRACKET_LEN))
if (right is not None):
right = right.move((right.begin + END_LEN), (right.end - BRACKET_LEN))
return (left, right)
| [
"def",
"highlighting",
"(",
"view",
",",
"name",
",",
"style",
",",
"left",
",",
"right",
")",
":",
"if",
"(",
"left",
"is",
"not",
"None",
")",
":",
"left",
"=",
"left",
".",
"move",
"(",
"(",
"left",
".",
"begin",
"+",
"BEGIN_LEN",
")",
",",
"(",
"left",
".",
"end",
"-",
"BRACKET_LEN",
")",
")",
"if",
"(",
"right",
"is",
"not",
"None",
")",
":",
"right",
"=",
"right",
".",
"move",
"(",
"(",
"right",
".",
"begin",
"+",
"END_LEN",
")",
",",
"(",
"right",
".",
"end",
"-",
"BRACKET_LEN",
")",
")",
"return",
"(",
"left",
",",
"right",
")"
] | highlight only the tag name . | train | false |
78 | def _inFilesystemNamespace(path):
return (path[:1] not in ('\x00', u'\x00'))
| [
"def",
"_inFilesystemNamespace",
"(",
"path",
")",
":",
"return",
"(",
"path",
"[",
":",
"1",
"]",
"not",
"in",
"(",
"'\\x00'",
",",
"u'\\x00'",
")",
")"
] | determine whether the given unix socket path is in a filesystem namespace . | train | false |
79 | def set_policy(name, table='filter', family='ipv4', **kwargs):
ret = {'name': name, 'changes': {}, 'result': None, 'comment': ''}
for ignore in _STATE_INTERNAL_KEYWORDS:
if (ignore in kwargs):
del kwargs[ignore]
if (__salt__['iptables.get_policy'](table, kwargs['chain'], family) == kwargs['policy']):
ret['result'] = True
ret['comment'] = 'iptables default policy for chain {0} on table {1} for {2} already set to {3}'.format(kwargs['chain'], table, family, kwargs['policy'])
return ret
if __opts__['test']:
ret['comment'] = 'iptables default policy for chain {0} on table {1} for {2} needs to be set to {3}'.format(kwargs['chain'], table, family, kwargs['policy'])
return ret
if (not __salt__['iptables.set_policy'](table, kwargs['chain'], kwargs['policy'], family)):
ret['changes'] = {'locale': name}
ret['result'] = True
ret['comment'] = 'Set default policy for {0} to {1} family {2}'.format(kwargs['chain'], kwargs['policy'], family)
if ('save' in kwargs):
if kwargs['save']:
__salt__['iptables.save'](filename=None, family=family)
ret['comment'] = 'Set and saved default policy for {0} to {1} family {2}'.format(kwargs['chain'], kwargs['policy'], family)
return ret
else:
ret['result'] = False
ret['comment'] = 'Failed to set iptables default policy'
return ret
| [
"def",
"set_policy",
"(",
"name",
",",
"table",
"=",
"'filter'",
",",
"family",
"=",
"'ipv4'",
",",
"**",
"kwargs",
")",
":",
"ret",
"=",
"{",
"'name'",
":",
"name",
",",
"'changes'",
":",
"{",
"}",
",",
"'result'",
":",
"None",
",",
"'comment'",
":",
"''",
"}",
"for",
"ignore",
"in",
"_STATE_INTERNAL_KEYWORDS",
":",
"if",
"(",
"ignore",
"in",
"kwargs",
")",
":",
"del",
"kwargs",
"[",
"ignore",
"]",
"if",
"(",
"__salt__",
"[",
"'iptables.get_policy'",
"]",
"(",
"table",
",",
"kwargs",
"[",
"'chain'",
"]",
",",
"family",
")",
"==",
"kwargs",
"[",
"'policy'",
"]",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"True",
"ret",
"[",
"'comment'",
"]",
"=",
"'iptables default policy for chain {0} on table {1} for {2} already set to {3}'",
".",
"format",
"(",
"kwargs",
"[",
"'chain'",
"]",
",",
"table",
",",
"family",
",",
"kwargs",
"[",
"'policy'",
"]",
")",
"return",
"ret",
"if",
"__opts__",
"[",
"'test'",
"]",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"'iptables default policy for chain {0} on table {1} for {2} needs to be set to {3}'",
".",
"format",
"(",
"kwargs",
"[",
"'chain'",
"]",
",",
"table",
",",
"family",
",",
"kwargs",
"[",
"'policy'",
"]",
")",
"return",
"ret",
"if",
"(",
"not",
"__salt__",
"[",
"'iptables.set_policy'",
"]",
"(",
"table",
",",
"kwargs",
"[",
"'chain'",
"]",
",",
"kwargs",
"[",
"'policy'",
"]",
",",
"family",
")",
")",
":",
"ret",
"[",
"'changes'",
"]",
"=",
"{",
"'locale'",
":",
"name",
"}",
"ret",
"[",
"'result'",
"]",
"=",
"True",
"ret",
"[",
"'comment'",
"]",
"=",
"'Set default policy for {0} to {1} family {2}'",
".",
"format",
"(",
"kwargs",
"[",
"'chain'",
"]",
",",
"kwargs",
"[",
"'policy'",
"]",
",",
"family",
")",
"if",
"(",
"'save'",
"in",
"kwargs",
")",
":",
"if",
"kwargs",
"[",
"'save'",
"]",
":",
"__salt__",
"[",
"'iptables.save'",
"]",
"(",
"filename",
"=",
"None",
",",
"family",
"=",
"family",
")",
"ret",
"[",
"'comment'",
"]",
"=",
"'Set and saved default policy for {0} to {1} family {2}'",
".",
"format",
"(",
"kwargs",
"[",
"'chain'",
"]",
",",
"kwargs",
"[",
"'policy'",
"]",
",",
"family",
")",
"return",
"ret",
"else",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"ret",
"[",
"'comment'",
"]",
"=",
"'Failed to set iptables default policy'",
"return",
"ret"
] | set the current policy for the specified table/chain cli example: . | train | true |
80 | def test_formatters(Chart):
if (Chart._dual or (Chart == Box)):
return
chart = Chart(formatter=(lambda x, chart, serie: ('%s%s$' % (x, serie.title))))
chart.add('_a', [1, 2, {'value': 3, 'formatter': (lambda x: (u('%s\xc2\xa5') % x))}])
chart.add('_b', [4, 5, 6], formatter=(lambda x: (u('%s\xe2\x82\xac') % x)))
chart.x_labels = [2, 4, 6]
chart.x_labels_major = [4]
q = chart.render_pyquery()
assert (set([v.text for v in q('.value')]) == set(((u('4\xe2\x82\xac'), u('5\xe2\x82\xac'), u('6\xe2\x82\xac'), '1_a$', '2_a$', u('3\xc2\xa5')) + (('6_a$', u('15\xe2\x82\xac')) if (Chart in (Pie, SolidGauge)) else ()))))
| [
"def",
"test_formatters",
"(",
"Chart",
")",
":",
"if",
"(",
"Chart",
".",
"_dual",
"or",
"(",
"Chart",
"==",
"Box",
")",
")",
":",
"return",
"chart",
"=",
"Chart",
"(",
"formatter",
"=",
"(",
"lambda",
"x",
",",
"chart",
",",
"serie",
":",
"(",
"'%s%s$'",
"%",
"(",
"x",
",",
"serie",
".",
"title",
")",
")",
")",
")",
"chart",
".",
"add",
"(",
"'_a'",
",",
"[",
"1",
",",
"2",
",",
"{",
"'value'",
":",
"3",
",",
"'formatter'",
":",
"(",
"lambda",
"x",
":",
"(",
"u",
"(",
"'%s\\xc2\\xa5'",
")",
"%",
"x",
")",
")",
"}",
"]",
")",
"chart",
".",
"add",
"(",
"'_b'",
",",
"[",
"4",
",",
"5",
",",
"6",
"]",
",",
"formatter",
"=",
"(",
"lambda",
"x",
":",
"(",
"u",
"(",
"'%s\\xe2\\x82\\xac'",
")",
"%",
"x",
")",
")",
")",
"chart",
".",
"x_labels",
"=",
"[",
"2",
",",
"4",
",",
"6",
"]",
"chart",
".",
"x_labels_major",
"=",
"[",
"4",
"]",
"q",
"=",
"chart",
".",
"render_pyquery",
"(",
")",
"assert",
"(",
"set",
"(",
"[",
"v",
".",
"text",
"for",
"v",
"in",
"q",
"(",
"'.value'",
")",
"]",
")",
"==",
"set",
"(",
"(",
"(",
"u",
"(",
"'4\\xe2\\x82\\xac'",
")",
",",
"u",
"(",
"'5\\xe2\\x82\\xac'",
")",
",",
"u",
"(",
"'6\\xe2\\x82\\xac'",
")",
",",
"'1_a$'",
",",
"'2_a$'",
",",
"u",
"(",
"'3\\xc2\\xa5'",
")",
")",
"+",
"(",
"(",
"'6_a$'",
",",
"u",
"(",
"'15\\xe2\\x82\\xac'",
")",
")",
"if",
"(",
"Chart",
"in",
"(",
"Pie",
",",
"SolidGauge",
")",
")",
"else",
"(",
")",
")",
")",
")",
")"
] | test custom formatters . | train | false |
81 | def test_sort():
model = _create_model([[('B', '', '', 1), ('C', '', '', 2), ('A', '', '', 0)]])
filter_model = sortfilter.CompletionFilterModel(model)
filter_model.sort(0, Qt.AscendingOrder)
actual = _extract_model_data(filter_model)
assert (actual == [[('A', '', ''), ('B', '', ''), ('C', '', '')]])
filter_model.sort(0, Qt.DescendingOrder)
actual = _extract_model_data(filter_model)
assert (actual == [[('C', '', ''), ('B', '', ''), ('A', '', '')]])
| [
"def",
"test_sort",
"(",
")",
":",
"model",
"=",
"_create_model",
"(",
"[",
"[",
"(",
"'B'",
",",
"''",
",",
"''",
",",
"1",
")",
",",
"(",
"'C'",
",",
"''",
",",
"''",
",",
"2",
")",
",",
"(",
"'A'",
",",
"''",
",",
"''",
",",
"0",
")",
"]",
"]",
")",
"filter_model",
"=",
"sortfilter",
".",
"CompletionFilterModel",
"(",
"model",
")",
"filter_model",
".",
"sort",
"(",
"0",
",",
"Qt",
".",
"AscendingOrder",
")",
"actual",
"=",
"_extract_model_data",
"(",
"filter_model",
")",
"assert",
"(",
"actual",
"==",
"[",
"[",
"(",
"'A'",
",",
"''",
",",
"''",
")",
",",
"(",
"'B'",
",",
"''",
",",
"''",
")",
",",
"(",
"'C'",
",",
"''",
",",
"''",
")",
"]",
"]",
")",
"filter_model",
".",
"sort",
"(",
"0",
",",
"Qt",
".",
"DescendingOrder",
")",
"actual",
"=",
"_extract_model_data",
"(",
"filter_model",
")",
"assert",
"(",
"actual",
"==",
"[",
"[",
"(",
"'C'",
",",
"''",
",",
"''",
")",
",",
"(",
"'B'",
",",
"''",
",",
"''",
")",
",",
"(",
"'A'",
",",
"''",
",",
"''",
")",
"]",
"]",
")"
] | ensure that a sort argument passed to sort overrides dumb_sort . | train | false |
82 | def create_dendrogram(X, orientation='bottom', labels=None, colorscale=None, distfun=None, linkagefun=(lambda x: sch.linkage(x, 'complete'))):
if ((not scp) or (not scs) or (not sch)):
raise ImportError('FigureFactory.create_dendrogram requires scipy, scipy.spatial and scipy.hierarchy')
s = X.shape
if (len(s) != 2):
exceptions.PlotlyError('X should be 2-dimensional array.')
if (distfun is None):
distfun = scs.distance.pdist
dendrogram = _Dendrogram(X, orientation, labels, colorscale, distfun=distfun, linkagefun=linkagefun)
return {'layout': dendrogram.layout, 'data': dendrogram.data}
| [
"def",
"create_dendrogram",
"(",
"X",
",",
"orientation",
"=",
"'bottom'",
",",
"labels",
"=",
"None",
",",
"colorscale",
"=",
"None",
",",
"distfun",
"=",
"None",
",",
"linkagefun",
"=",
"(",
"lambda",
"x",
":",
"sch",
".",
"linkage",
"(",
"x",
",",
"'complete'",
")",
")",
")",
":",
"if",
"(",
"(",
"not",
"scp",
")",
"or",
"(",
"not",
"scs",
")",
"or",
"(",
"not",
"sch",
")",
")",
":",
"raise",
"ImportError",
"(",
"'FigureFactory.create_dendrogram requires scipy, scipy.spatial and scipy.hierarchy'",
")",
"s",
"=",
"X",
".",
"shape",
"if",
"(",
"len",
"(",
"s",
")",
"!=",
"2",
")",
":",
"exceptions",
".",
"PlotlyError",
"(",
"'X should be 2-dimensional array.'",
")",
"if",
"(",
"distfun",
"is",
"None",
")",
":",
"distfun",
"=",
"scs",
".",
"distance",
".",
"pdist",
"dendrogram",
"=",
"_Dendrogram",
"(",
"X",
",",
"orientation",
",",
"labels",
",",
"colorscale",
",",
"distfun",
"=",
"distfun",
",",
"linkagefun",
"=",
"linkagefun",
")",
"return",
"{",
"'layout'",
":",
"dendrogram",
".",
"layout",
",",
"'data'",
":",
"dendrogram",
".",
"data",
"}"
] | beta function that returns a dendrogram plotly figure object . | train | false |
83 | def uni_print(statement, out_file=None):
if (out_file is None):
out_file = sys.stdout
try:
out_file.write(statement)
except UnicodeEncodeError:
new_encoding = getattr(out_file, 'encoding', 'ascii')
if (new_encoding is None):
new_encoding = 'ascii'
new_statement = statement.encode(new_encoding, 'replace').decode(new_encoding)
out_file.write(new_statement)
out_file.flush()
| [
"def",
"uni_print",
"(",
"statement",
",",
"out_file",
"=",
"None",
")",
":",
"if",
"(",
"out_file",
"is",
"None",
")",
":",
"out_file",
"=",
"sys",
".",
"stdout",
"try",
":",
"out_file",
".",
"write",
"(",
"statement",
")",
"except",
"UnicodeEncodeError",
":",
"new_encoding",
"=",
"getattr",
"(",
"out_file",
",",
"'encoding'",
",",
"'ascii'",
")",
"if",
"(",
"new_encoding",
"is",
"None",
")",
":",
"new_encoding",
"=",
"'ascii'",
"new_statement",
"=",
"statement",
".",
"encode",
"(",
"new_encoding",
",",
"'replace'",
")",
".",
"decode",
"(",
"new_encoding",
")",
"out_file",
".",
"write",
"(",
"new_statement",
")",
"out_file",
".",
"flush",
"(",
")"
] | this function is used to properly write unicode to a file . | train | false |
84 | def pretty_name(name):
if (not name):
return ''
return name.replace('_', ' ').capitalize()
| [
"def",
"pretty_name",
"(",
"name",
")",
":",
"if",
"(",
"not",
"name",
")",
":",
"return",
"''",
"return",
"name",
".",
"replace",
"(",
"'_'",
",",
"' '",
")",
".",
"capitalize",
"(",
")"
] | converts first_name to first name . | train | false |
85 | def list_snapshots(config='root'):
try:
snapshots = snapper.ListSnapshots(config)
return [_snapshot_to_data(s) for s in snapshots]
except dbus.DBusException as exc:
raise CommandExecutionError('Error encountered while listing snapshots: {0}'.format(_dbus_exception_to_reason(exc, locals())))
| [
"def",
"list_snapshots",
"(",
"config",
"=",
"'root'",
")",
":",
"try",
":",
"snapshots",
"=",
"snapper",
".",
"ListSnapshots",
"(",
"config",
")",
"return",
"[",
"_snapshot_to_data",
"(",
"s",
")",
"for",
"s",
"in",
"snapshots",
"]",
"except",
"dbus",
".",
"DBusException",
"as",
"exc",
":",
"raise",
"CommandExecutionError",
"(",
"'Error encountered while listing snapshots: {0}'",
".",
"format",
"(",
"_dbus_exception_to_reason",
"(",
"exc",
",",
"locals",
"(",
")",
")",
")",
")"
] | list available snapshots for certain vm or for all . | train | true |
86 | def branch_list(repo):
with open_repo_closing(repo) as r:
return r.refs.keys(base='refs/heads/')
| [
"def",
"branch_list",
"(",
"repo",
")",
":",
"with",
"open_repo_closing",
"(",
"repo",
")",
"as",
"r",
":",
"return",
"r",
".",
"refs",
".",
"keys",
"(",
"base",
"=",
"'refs/heads/'",
")"
] | return a list of local or remote branches this explicitly removes head from the list of remote branches . | train | false |
87 | @depends(HAS_PYVMOMI)
def get_ntp_config(host, username, password, protocol=None, port=None, host_names=None):
service_instance = salt.utils.vmware.get_service_instance(host=host, username=username, password=password, protocol=protocol, port=port)
host_names = _check_hosts(service_instance, host, host_names)
ret = {}
for host_name in host_names:
host_ref = _get_host_ref(service_instance, host, host_name=host_name)
ntp_config = host_ref.configManager.dateTimeSystem.dateTimeInfo.ntpConfig.server
ret.update({host_name: ntp_config})
return ret
| [
"@",
"depends",
"(",
"HAS_PYVMOMI",
")",
"def",
"get_ntp_config",
"(",
"host",
",",
"username",
",",
"password",
",",
"protocol",
"=",
"None",
",",
"port",
"=",
"None",
",",
"host_names",
"=",
"None",
")",
":",
"service_instance",
"=",
"salt",
".",
"utils",
".",
"vmware",
".",
"get_service_instance",
"(",
"host",
"=",
"host",
",",
"username",
"=",
"username",
",",
"password",
"=",
"password",
",",
"protocol",
"=",
"protocol",
",",
"port",
"=",
"port",
")",
"host_names",
"=",
"_check_hosts",
"(",
"service_instance",
",",
"host",
",",
"host_names",
")",
"ret",
"=",
"{",
"}",
"for",
"host_name",
"in",
"host_names",
":",
"host_ref",
"=",
"_get_host_ref",
"(",
"service_instance",
",",
"host",
",",
"host_name",
"=",
"host_name",
")",
"ntp_config",
"=",
"host_ref",
".",
"configManager",
".",
"dateTimeSystem",
".",
"dateTimeInfo",
".",
"ntpConfig",
".",
"server",
"ret",
".",
"update",
"(",
"{",
"host_name",
":",
"ntp_config",
"}",
")",
"return",
"ret"
] | get the ntp configuration information for a given host or list of host_names . | train | true |
89 | def floating_ip_list(call=None):
if (call != 'function'):
raise SaltCloudSystemExit('The floating_ip_list action must be called with -f or --function')
conn = get_conn()
return conn.floating_ip_list()
| [
"def",
"floating_ip_list",
"(",
"call",
"=",
"None",
")",
":",
"if",
"(",
"call",
"!=",
"'function'",
")",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The floating_ip_list action must be called with -f or --function'",
")",
"conn",
"=",
"get_conn",
"(",
")",
"return",
"conn",
".",
"floating_ip_list",
"(",
")"
] | list floating ips . | train | false |
90 | def tsql_query(query, **kwargs):
try:
cur = _get_connection(**kwargs).cursor()
cur.execute(query)
return loads(_MssqlEncoder().encode({'resultset': cur.fetchall()}))['resultset']
except Exception as e:
return (('Could not run the query',), (str(e),))
| [
"def",
"tsql_query",
"(",
"query",
",",
"**",
"kwargs",
")",
":",
"try",
":",
"cur",
"=",
"_get_connection",
"(",
"**",
"kwargs",
")",
".",
"cursor",
"(",
")",
"cur",
".",
"execute",
"(",
"query",
")",
"return",
"loads",
"(",
"_MssqlEncoder",
"(",
")",
".",
"encode",
"(",
"{",
"'resultset'",
":",
"cur",
".",
"fetchall",
"(",
")",
"}",
")",
")",
"[",
"'resultset'",
"]",
"except",
"Exception",
"as",
"e",
":",
"return",
"(",
"(",
"'Could not run the query'",
",",
")",
",",
"(",
"str",
"(",
"e",
")",
",",
")",
")"
] | run a sql query and return query result as list of tuples . | train | false |
92 | def per_cpu_times():
ret = cext.per_cpu_times()
return [scputimes(*x) for x in ret]
| [
"def",
"per_cpu_times",
"(",
")",
":",
"ret",
"=",
"cext",
".",
"per_cpu_times",
"(",
")",
"return",
"[",
"scputimes",
"(",
"*",
"x",
")",
"for",
"x",
"in",
"ret",
"]"
] | return system per-cpu times as a list of named tuples . | train | false |
93 | def _run_aws(cmd, region, opts, user, **kwargs):
receipthandle = kwargs.pop('receipthandle', None)
if receipthandle:
kwargs['receipt-handle'] = receipthandle
num = kwargs.pop('num', None)
if num:
kwargs['max-number-of-messages'] = num
_formatted_args = ['--{0} "{1}"'.format(k, v) for (k, v) in six.iteritems(kwargs)]
cmd = 'aws sqs {cmd} {args} {region} {out}'.format(cmd=cmd, args=' '.join(_formatted_args), region=_region(region), out=_OUTPUT)
rtn = __salt__['cmd.run'](cmd, runas=user, python_shell=False)
return (json.loads(rtn) if rtn else '')
| [
"def",
"_run_aws",
"(",
"cmd",
",",
"region",
",",
"opts",
",",
"user",
",",
"**",
"kwargs",
")",
":",
"receipthandle",
"=",
"kwargs",
".",
"pop",
"(",
"'receipthandle'",
",",
"None",
")",
"if",
"receipthandle",
":",
"kwargs",
"[",
"'receipt-handle'",
"]",
"=",
"receipthandle",
"num",
"=",
"kwargs",
".",
"pop",
"(",
"'num'",
",",
"None",
")",
"if",
"num",
":",
"kwargs",
"[",
"'max-number-of-messages'",
"]",
"=",
"num",
"_formatted_args",
"=",
"[",
"'--{0} \"{1}\"'",
".",
"format",
"(",
"k",
",",
"v",
")",
"for",
"(",
"k",
",",
"v",
")",
"in",
"six",
".",
"iteritems",
"(",
"kwargs",
")",
"]",
"cmd",
"=",
"'aws sqs {cmd} {args} {region} {out}'",
".",
"format",
"(",
"cmd",
"=",
"cmd",
",",
"args",
"=",
"' '",
".",
"join",
"(",
"_formatted_args",
")",
",",
"region",
"=",
"_region",
"(",
"region",
")",
",",
"out",
"=",
"_OUTPUT",
")",
"rtn",
"=",
"__salt__",
"[",
"'cmd.run'",
"]",
"(",
"cmd",
",",
"runas",
"=",
"user",
",",
"python_shell",
"=",
"False",
")",
"return",
"(",
"json",
".",
"loads",
"(",
"rtn",
")",
"if",
"rtn",
"else",
"''",
")"
] | runs the given command against aws . | train | true |
94 | def TrimmedMean(t, p=0.01):
t = Trim(t, p)
return Mean(t)
| [
"def",
"TrimmedMean",
"(",
"t",
",",
"p",
"=",
"0.01",
")",
":",
"t",
"=",
"Trim",
"(",
"t",
",",
"p",
")",
"return",
"Mean",
"(",
"t",
")"
] | computes the trimmed mean of a sequence of numbers . | train | false |
95 | def ePut(Handle, IOType, Channel, Value, x1):
if (os.name == 'nt'):
staticLib = ctypes.windll.LoadLibrary('labjackud')
pv = ctypes.c_double(Value)
ec = staticLib.ePut(Handle, IOType, Channel, pv, x1)
if (ec != 0):
raise LabJackException(ec)
else:
raise LabJackException(0, 'Function only supported for Windows')
| [
"def",
"ePut",
"(",
"Handle",
",",
"IOType",
",",
"Channel",
",",
"Value",
",",
"x1",
")",
":",
"if",
"(",
"os",
".",
"name",
"==",
"'nt'",
")",
":",
"staticLib",
"=",
"ctypes",
".",
"windll",
".",
"LoadLibrary",
"(",
"'labjackud'",
")",
"pv",
"=",
"ctypes",
".",
"c_double",
"(",
"Value",
")",
"ec",
"=",
"staticLib",
".",
"ePut",
"(",
"Handle",
",",
"IOType",
",",
"Channel",
",",
"pv",
",",
"x1",
")",
"if",
"(",
"ec",
"!=",
"0",
")",
":",
"raise",
"LabJackException",
"(",
"ec",
")",
"else",
":",
"raise",
"LabJackException",
"(",
"0",
",",
"'Function only supported for Windows'",
")"
] | put one value to the labjack device eput is equivilent to an addrequest followed by a goone . | train | false |
97 | def local_binary_pattern(image, P, R, method='default'):
assert_nD(image, 2)
methods = {'default': ord('D'), 'ror': ord('R'), 'uniform': ord('U'), 'nri_uniform': ord('N'), 'var': ord('V')}
image = np.ascontiguousarray(image, dtype=np.double)
output = _local_binary_pattern(image, P, R, methods[method.lower()])
return output
| [
"def",
"local_binary_pattern",
"(",
"image",
",",
"P",
",",
"R",
",",
"method",
"=",
"'default'",
")",
":",
"assert_nD",
"(",
"image",
",",
"2",
")",
"methods",
"=",
"{",
"'default'",
":",
"ord",
"(",
"'D'",
")",
",",
"'ror'",
":",
"ord",
"(",
"'R'",
")",
",",
"'uniform'",
":",
"ord",
"(",
"'U'",
")",
",",
"'nri_uniform'",
":",
"ord",
"(",
"'N'",
")",
",",
"'var'",
":",
"ord",
"(",
"'V'",
")",
"}",
"image",
"=",
"np",
".",
"ascontiguousarray",
"(",
"image",
",",
"dtype",
"=",
"np",
".",
"double",
")",
"output",
"=",
"_local_binary_pattern",
"(",
"image",
",",
"P",
",",
"R",
",",
"methods",
"[",
"method",
".",
"lower",
"(",
")",
"]",
")",
"return",
"output"
] | gray scale and rotation invariant lbp . | train | false |
98 | def handleNewest(qry):
try:
getUserName()
except:
return _skypeError()
qry = qry.decode('utf8')
try:
if ((':' in qry) and (qry.partition(':')[0] in map((lambda s: s[0]), _readFriends()))):
return _sendMessageWait(qry)
else:
return _findNewest()
except EnvironmentError:
return PyFred('ch.xtin.skypingalfred.error', False).addItem('skypeupdate', 'skype update', 'No Skype Friends Found', 'Use skype update to cache friends!', True, 'update').toXML()
except:
return PyFred.GenericError()
| [
"def",
"handleNewest",
"(",
"qry",
")",
":",
"try",
":",
"getUserName",
"(",
")",
"except",
":",
"return",
"_skypeError",
"(",
")",
"qry",
"=",
"qry",
".",
"decode",
"(",
"'utf8'",
")",
"try",
":",
"if",
"(",
"(",
"':'",
"in",
"qry",
")",
"and",
"(",
"qry",
".",
"partition",
"(",
"':'",
")",
"[",
"0",
"]",
"in",
"map",
"(",
"(",
"lambda",
"s",
":",
"s",
"[",
"0",
"]",
")",
",",
"_readFriends",
"(",
")",
")",
")",
")",
":",
"return",
"_sendMessageWait",
"(",
"qry",
")",
"else",
":",
"return",
"_findNewest",
"(",
")",
"except",
"EnvironmentError",
":",
"return",
"PyFred",
"(",
"'ch.xtin.skypingalfred.error'",
",",
"False",
")",
".",
"addItem",
"(",
"'skypeupdate'",
",",
"'skype update'",
",",
"'No Skype Friends Found'",
",",
"'Use skype update to cache friends!'",
",",
"True",
",",
"'update'",
")",
".",
"toXML",
"(",
")",
"except",
":",
"return",
"PyFred",
".",
"GenericError",
"(",
")"
] | gets the newest 5 messages . | train | false |
99 | @validate('tree')
def valid_field_in_tree(arch):
return all(((child.tag in ('field', 'button')) for child in arch.xpath('/tree/*')))
| [
"@",
"validate",
"(",
"'tree'",
")",
"def",
"valid_field_in_tree",
"(",
"arch",
")",
":",
"return",
"all",
"(",
"(",
"(",
"child",
".",
"tag",
"in",
"(",
"'field'",
",",
"'button'",
")",
")",
"for",
"child",
"in",
"arch",
".",
"xpath",
"(",
"'/tree/*'",
")",
")",
")"
] | children of tree view must be field or button . | train | false |
100 | def cr_uid_ids(method):
method._api = 'cr_uid_ids'
return method
| [
"def",
"cr_uid_ids",
"(",
"method",
")",
":",
"method",
".",
"_api",
"=",
"'cr_uid_ids'",
"return",
"method"
] | decorate a traditional-style method that takes cr . | train | false |
101 | def issue_section(issue):
labels = issue.get('labels', [])
for label in labels:
if (not label['name'].startswith('type: ')):
continue
if (label['name'] in LOG_SECTION):
return LOG_SECTION[label['name']]
elif (label['name'] in IGNORE_ISSUE_TYPE):
return None
else:
logging.warn('unknown issue type: "{}" for: {}'.format(label['name'], issue_line(issue)))
return None
| [
"def",
"issue_section",
"(",
"issue",
")",
":",
"labels",
"=",
"issue",
".",
"get",
"(",
"'labels'",
",",
"[",
"]",
")",
"for",
"label",
"in",
"labels",
":",
"if",
"(",
"not",
"label",
"[",
"'name'",
"]",
".",
"startswith",
"(",
"'type: '",
")",
")",
":",
"continue",
"if",
"(",
"label",
"[",
"'name'",
"]",
"in",
"LOG_SECTION",
")",
":",
"return",
"LOG_SECTION",
"[",
"label",
"[",
"'name'",
"]",
"]",
"elif",
"(",
"label",
"[",
"'name'",
"]",
"in",
"IGNORE_ISSUE_TYPE",
")",
":",
"return",
"None",
"else",
":",
"logging",
".",
"warn",
"(",
"'unknown issue type: \"{}\" for: {}'",
".",
"format",
"(",
"label",
"[",
"'name'",
"]",
",",
"issue_line",
"(",
"issue",
")",
")",
")",
"return",
"None"
] | returns the section heading for the issue . | train | true |
103 | @transaction.non_atomic_requests
@ensure_csrf_cookie
@cache_control(no_cache=True, no_store=True, must_revalidate=True)
@require_global_staff
@require_http_methods(['POST', 'DELETE'])
def certificate_invalidation_view(request, course_id):
course_key = CourseKey.from_string(course_id)
try:
certificate_invalidation_data = parse_request_data(request)
certificate = validate_request_data_and_get_certificate(certificate_invalidation_data, course_key)
except ValueError as error:
return JsonResponse({'message': error.message}, status=400)
if (request.method == 'POST'):
try:
certificate_invalidation = invalidate_certificate(request, certificate, certificate_invalidation_data)
except ValueError as error:
return JsonResponse({'message': error.message}, status=400)
return JsonResponse(certificate_invalidation)
elif (request.method == 'DELETE'):
try:
re_validate_certificate(request, course_key, certificate)
except ValueError as error:
return JsonResponse({'message': error.message}, status=400)
return JsonResponse({}, status=204)
| [
"@",
"transaction",
".",
"non_atomic_requests",
"@",
"ensure_csrf_cookie",
"@",
"cache_control",
"(",
"no_cache",
"=",
"True",
",",
"no_store",
"=",
"True",
",",
"must_revalidate",
"=",
"True",
")",
"@",
"require_global_staff",
"@",
"require_http_methods",
"(",
"[",
"'POST'",
",",
"'DELETE'",
"]",
")",
"def",
"certificate_invalidation_view",
"(",
"request",
",",
"course_id",
")",
":",
"course_key",
"=",
"CourseKey",
".",
"from_string",
"(",
"course_id",
")",
"try",
":",
"certificate_invalidation_data",
"=",
"parse_request_data",
"(",
"request",
")",
"certificate",
"=",
"validate_request_data_and_get_certificate",
"(",
"certificate_invalidation_data",
",",
"course_key",
")",
"except",
"ValueError",
"as",
"error",
":",
"return",
"JsonResponse",
"(",
"{",
"'message'",
":",
"error",
".",
"message",
"}",
",",
"status",
"=",
"400",
")",
"if",
"(",
"request",
".",
"method",
"==",
"'POST'",
")",
":",
"try",
":",
"certificate_invalidation",
"=",
"invalidate_certificate",
"(",
"request",
",",
"certificate",
",",
"certificate_invalidation_data",
")",
"except",
"ValueError",
"as",
"error",
":",
"return",
"JsonResponse",
"(",
"{",
"'message'",
":",
"error",
".",
"message",
"}",
",",
"status",
"=",
"400",
")",
"return",
"JsonResponse",
"(",
"certificate_invalidation",
")",
"elif",
"(",
"request",
".",
"method",
"==",
"'DELETE'",
")",
":",
"try",
":",
"re_validate_certificate",
"(",
"request",
",",
"course_key",
",",
"certificate",
")",
"except",
"ValueError",
"as",
"error",
":",
"return",
"JsonResponse",
"(",
"{",
"'message'",
":",
"error",
".",
"message",
"}",
",",
"status",
"=",
"400",
")",
"return",
"JsonResponse",
"(",
"{",
"}",
",",
"status",
"=",
"204",
")"
] | invalidate/re-validate students to/from certificate . | train | false |
104 | def rollback():
connection._rollback()
set_clean()
| [
"def",
"rollback",
"(",
")",
":",
"connection",
".",
"_rollback",
"(",
")",
"set_clean",
"(",
")"
] | rolls back a transaction . | train | false |
105 | def _mathdefault(s):
if rcParams[u'_internal.classic_mode']:
return (u'\\mathdefault{%s}' % s)
else:
return (u'{%s}' % s)
| [
"def",
"_mathdefault",
"(",
"s",
")",
":",
"if",
"rcParams",
"[",
"u'_internal.classic_mode'",
"]",
":",
"return",
"(",
"u'\\\\mathdefault{%s}'",
"%",
"s",
")",
"else",
":",
"return",
"(",
"u'{%s}'",
"%",
"s",
")"
] | for backward compatibility . | train | false |
106 | @requires_sklearn
def test_gat_plot_nonsquared():
gat = _get_data(test_times=dict(start=0.0))
gat.plot()
ax = gat.plot_diagonal()
scores = ax.get_children()[1].get_lines()[2].get_ydata()
assert_equals(len(scores), len(gat.estimators_))
| [
"@",
"requires_sklearn",
"def",
"test_gat_plot_nonsquared",
"(",
")",
":",
"gat",
"=",
"_get_data",
"(",
"test_times",
"=",
"dict",
"(",
"start",
"=",
"0.0",
")",
")",
"gat",
".",
"plot",
"(",
")",
"ax",
"=",
"gat",
".",
"plot_diagonal",
"(",
")",
"scores",
"=",
"ax",
".",
"get_children",
"(",
")",
"[",
"1",
"]",
".",
"get_lines",
"(",
")",
"[",
"2",
"]",
".",
"get_ydata",
"(",
")",
"assert_equals",
"(",
"len",
"(",
"scores",
")",
",",
"len",
"(",
"gat",
".",
"estimators_",
")",
")"
] | test gat diagonal plot . | train | false |
107 | def _add_keys_to_request(request_field_pb, key_pbs):
for key_pb in key_pbs:
request_field_pb.add().CopyFrom(key_pb)
| [
"def",
"_add_keys_to_request",
"(",
"request_field_pb",
",",
"key_pbs",
")",
":",
"for",
"key_pb",
"in",
"key_pbs",
":",
"request_field_pb",
".",
"add",
"(",
")",
".",
"CopyFrom",
"(",
"key_pb",
")"
] | add protobuf keys to a request object . | train | false |
108 | def execute_on_completion(application, config, callback):
def inner(environ, start_response):
try:
result = application(environ, start_response)
except:
callback(environ)
raise
return generate_close_and_callback(result, callback, environ)
return inner
| [
"def",
"execute_on_completion",
"(",
"application",
",",
"config",
",",
"callback",
")",
":",
"def",
"inner",
"(",
"environ",
",",
"start_response",
")",
":",
"try",
":",
"result",
"=",
"application",
"(",
"environ",
",",
"start_response",
")",
"except",
":",
"callback",
"(",
"environ",
")",
"raise",
"return",
"generate_close_and_callback",
"(",
"result",
",",
"callback",
",",
"environ",
")",
"return",
"inner"
] | call callback once complete response is sent . | train | false |
109 | def qt5_qml_data(directory):
qmldir = qt5_qml_dir()
return (os.path.join(qmldir, directory), 'qml')
| [
"def",
"qt5_qml_data",
"(",
"directory",
")",
":",
"qmldir",
"=",
"qt5_qml_dir",
"(",
")",
"return",
"(",
"os",
".",
"path",
".",
"join",
"(",
"qmldir",
",",
"directory",
")",
",",
"'qml'",
")"
] | return qml library directory formatted for data . | train | false |
110 | def require_finance_admin(func):
def wrapped(request, course_id):
try:
course_key = CourseKey.from_string(course_id)
except InvalidKeyError:
log.error(u'Unable to find course with course key %s', course_id)
return HttpResponseNotFound()
access = auth.user_has_role(request.user, CourseFinanceAdminRole(course_key))
if access:
return func(request, course_id)
else:
return HttpResponseForbidden()
return wrapped
| [
"def",
"require_finance_admin",
"(",
"func",
")",
":",
"def",
"wrapped",
"(",
"request",
",",
"course_id",
")",
":",
"try",
":",
"course_key",
"=",
"CourseKey",
".",
"from_string",
"(",
"course_id",
")",
"except",
"InvalidKeyError",
":",
"log",
".",
"error",
"(",
"u'Unable to find course with course key %s'",
",",
"course_id",
")",
"return",
"HttpResponseNotFound",
"(",
")",
"access",
"=",
"auth",
".",
"user_has_role",
"(",
"request",
".",
"user",
",",
"CourseFinanceAdminRole",
"(",
"course_key",
")",
")",
"if",
"access",
":",
"return",
"func",
"(",
"request",
",",
"course_id",
")",
"else",
":",
"return",
"HttpResponseForbidden",
"(",
")",
"return",
"wrapped"
] | decorator for checking finance administrator access before executing an http endpoint . | train | false |
111 | def addBeginXMLTag(attributeDictionary, className, depth, output, text=''):
depthStart = (' DCTB ' * depth)
output.write(('%s<%s%s>%s\n' % (depthStart, className, getAttributeDictionaryString(attributeDictionary), text)))
| [
"def",
"addBeginXMLTag",
"(",
"attributeDictionary",
",",
"className",
",",
"depth",
",",
"output",
",",
"text",
"=",
"''",
")",
":",
"depthStart",
"=",
"(",
"' DCTB '",
"*",
"depth",
")",
"output",
".",
"write",
"(",
"(",
"'%s<%s%s>%s\\n'",
"%",
"(",
"depthStart",
",",
"className",
",",
"getAttributeDictionaryString",
"(",
"attributeDictionary",
")",
",",
"text",
")",
")",
")"
] | add the begin xml tag . | train | false |
113 | def test_cons_list():
entry = tokenize('(a . [])')[0]
assert (entry == HyList([HySymbol('a')]))
assert (type(entry) == HyList)
entry = tokenize('(a . ())')[0]
assert (entry == HyExpression([HySymbol('a')]))
assert (type(entry) == HyExpression)
entry = tokenize('(a b . {})')[0]
assert (entry == HyDict([HySymbol('a'), HySymbol('b')]))
assert (type(entry) == HyDict)
| [
"def",
"test_cons_list",
"(",
")",
":",
"entry",
"=",
"tokenize",
"(",
"'(a . [])'",
")",
"[",
"0",
"]",
"assert",
"(",
"entry",
"==",
"HyList",
"(",
"[",
"HySymbol",
"(",
"'a'",
")",
"]",
")",
")",
"assert",
"(",
"type",
"(",
"entry",
")",
"==",
"HyList",
")",
"entry",
"=",
"tokenize",
"(",
"'(a . ())'",
")",
"[",
"0",
"]",
"assert",
"(",
"entry",
"==",
"HyExpression",
"(",
"[",
"HySymbol",
"(",
"'a'",
")",
"]",
")",
")",
"assert",
"(",
"type",
"(",
"entry",
")",
"==",
"HyExpression",
")",
"entry",
"=",
"tokenize",
"(",
"'(a b . {})'",
")",
"[",
"0",
"]",
"assert",
"(",
"entry",
"==",
"HyDict",
"(",
"[",
"HySymbol",
"(",
"'a'",
")",
",",
"HySymbol",
"(",
"'b'",
")",
"]",
")",
")",
"assert",
"(",
"type",
"(",
"entry",
")",
"==",
"HyDict",
")"
] | check that cons of something and a list gets tokenized as a list . | train | false |
114 | def _wait_for_step(emr_connection, step, jobflowid, sleeptime):
sleep(180)
start = time()
step_state = get_step_state(emr_connection, jobflowid, step.name, update=True)
while (step_state in (LIVE_STATES + [PENDING])):
sleep(sleeptime)
step_state = get_step_state(emr_connection, jobflowid, step.name)
end = time()
print ('%s took %0.2fs (exit: %s)' % (step.name, (end - start), step_state))
return step_state
| [
"def",
"_wait_for_step",
"(",
"emr_connection",
",",
"step",
",",
"jobflowid",
",",
"sleeptime",
")",
":",
"sleep",
"(",
"180",
")",
"start",
"=",
"time",
"(",
")",
"step_state",
"=",
"get_step_state",
"(",
"emr_connection",
",",
"jobflowid",
",",
"step",
".",
"name",
",",
"update",
"=",
"True",
")",
"while",
"(",
"step_state",
"in",
"(",
"LIVE_STATES",
"+",
"[",
"PENDING",
"]",
")",
")",
":",
"sleep",
"(",
"sleeptime",
")",
"step_state",
"=",
"get_step_state",
"(",
"emr_connection",
",",
"jobflowid",
",",
"step",
".",
"name",
")",
"end",
"=",
"time",
"(",
")",
"print",
"(",
"'%s took %0.2fs (exit: %s)'",
"%",
"(",
"step",
".",
"name",
",",
"(",
"end",
"-",
"start",
")",
",",
"step_state",
")",
")",
"return",
"step_state"
] | poll emr and wait for a step to finish . | train | false |
116 | def get_occupied_streams(realm):
subs_filter = Subscription.objects.filter(active=True, user_profile__realm=realm, user_profile__is_active=True).values('recipient_id')
stream_ids = Recipient.objects.filter(type=Recipient.STREAM, id__in=subs_filter).values('type_id')
return Stream.objects.filter(id__in=stream_ids, realm=realm, deactivated=False)
| [
"def",
"get_occupied_streams",
"(",
"realm",
")",
":",
"subs_filter",
"=",
"Subscription",
".",
"objects",
".",
"filter",
"(",
"active",
"=",
"True",
",",
"user_profile__realm",
"=",
"realm",
",",
"user_profile__is_active",
"=",
"True",
")",
".",
"values",
"(",
"'recipient_id'",
")",
"stream_ids",
"=",
"Recipient",
".",
"objects",
".",
"filter",
"(",
"type",
"=",
"Recipient",
".",
"STREAM",
",",
"id__in",
"=",
"subs_filter",
")",
".",
"values",
"(",
"'type_id'",
")",
"return",
"Stream",
".",
"objects",
".",
"filter",
"(",
"id__in",
"=",
"stream_ids",
",",
"realm",
"=",
"realm",
",",
"deactivated",
"=",
"False",
")"
] | get streams with subscribers . | train | false |
117 | def test_find_number_6():
s = 'query1e5 not found'
r = find_number(s)
assert (s[r[0]:r[1]] == '1e5')
| [
"def",
"test_find_number_6",
"(",
")",
":",
"s",
"=",
"'query1e5 not found'",
"r",
"=",
"find_number",
"(",
"s",
")",
"assert",
"(",
"s",
"[",
"r",
"[",
"0",
"]",
":",
"r",
"[",
"1",
"]",
"]",
"==",
"'1e5'",
")"
] | tests that we find numbers with exponents . | train | false |
118 | def _build_match_rule(action, target, pluralized):
match_rule = policy.RuleCheck('rule', action)
(resource, enforce_attr_based_check) = get_resource_and_action(action, pluralized)
if enforce_attr_based_check:
res_map = attributes.RESOURCE_ATTRIBUTE_MAP
if (resource in res_map):
for attribute_name in res_map[resource]:
if _is_attribute_explicitly_set(attribute_name, res_map[resource], target, action):
attribute = res_map[resource][attribute_name]
if ('enforce_policy' in attribute):
attr_rule = policy.RuleCheck('rule', ('%s:%s' % (action, attribute_name)))
if _should_validate_sub_attributes(attribute, target[attribute_name]):
attr_rule = policy.AndCheck([attr_rule, _build_subattr_match_rule(attribute_name, attribute, action, target)])
match_rule = policy.AndCheck([match_rule, attr_rule])
return match_rule
| [
"def",
"_build_match_rule",
"(",
"action",
",",
"target",
",",
"pluralized",
")",
":",
"match_rule",
"=",
"policy",
".",
"RuleCheck",
"(",
"'rule'",
",",
"action",
")",
"(",
"resource",
",",
"enforce_attr_based_check",
")",
"=",
"get_resource_and_action",
"(",
"action",
",",
"pluralized",
")",
"if",
"enforce_attr_based_check",
":",
"res_map",
"=",
"attributes",
".",
"RESOURCE_ATTRIBUTE_MAP",
"if",
"(",
"resource",
"in",
"res_map",
")",
":",
"for",
"attribute_name",
"in",
"res_map",
"[",
"resource",
"]",
":",
"if",
"_is_attribute_explicitly_set",
"(",
"attribute_name",
",",
"res_map",
"[",
"resource",
"]",
",",
"target",
",",
"action",
")",
":",
"attribute",
"=",
"res_map",
"[",
"resource",
"]",
"[",
"attribute_name",
"]",
"if",
"(",
"'enforce_policy'",
"in",
"attribute",
")",
":",
"attr_rule",
"=",
"policy",
".",
"RuleCheck",
"(",
"'rule'",
",",
"(",
"'%s:%s'",
"%",
"(",
"action",
",",
"attribute_name",
")",
")",
")",
"if",
"_should_validate_sub_attributes",
"(",
"attribute",
",",
"target",
"[",
"attribute_name",
"]",
")",
":",
"attr_rule",
"=",
"policy",
".",
"AndCheck",
"(",
"[",
"attr_rule",
",",
"_build_subattr_match_rule",
"(",
"attribute_name",
",",
"attribute",
",",
"action",
",",
"target",
")",
"]",
")",
"match_rule",
"=",
"policy",
".",
"AndCheck",
"(",
"[",
"match_rule",
",",
"attr_rule",
"]",
")",
"return",
"match_rule"
] | create the rule to match for a given action . | train | false |
119 | def dmp_ground_LC(f, u, K):
while u:
f = dmp_LC(f, K)
u -= 1
return dup_LC(f, K)
| [
"def",
"dmp_ground_LC",
"(",
"f",
",",
"u",
",",
"K",
")",
":",
"while",
"u",
":",
"f",
"=",
"dmp_LC",
"(",
"f",
",",
"K",
")",
"u",
"-=",
"1",
"return",
"dup_LC",
"(",
"f",
",",
"K",
")"
] | return the ground leading coefficient . | train | false |
120 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get new repository . | train | false |
122 | def get_entrance_exam_score(request, course):
exam_key = UsageKey.from_string(course.entrance_exam_id)
exam_descriptor = modulestore().get_item(exam_key)
def inner_get_module(descriptor):
'\n Delegate to get_module_for_descriptor (imported here to avoid circular reference)\n '
from courseware.module_render import get_module_for_descriptor
field_data_cache = FieldDataCache([descriptor], course.id, request.user)
return get_module_for_descriptor(request.user, request, descriptor, field_data_cache, course.id, course=course)
exam_module_generators = yield_dynamic_descriptor_descendants(exam_descriptor, request.user.id, inner_get_module)
exam_modules = [module for module in exam_module_generators]
return _calculate_entrance_exam_score(request.user, course, exam_modules)
| [
"def",
"get_entrance_exam_score",
"(",
"request",
",",
"course",
")",
":",
"exam_key",
"=",
"UsageKey",
".",
"from_string",
"(",
"course",
".",
"entrance_exam_id",
")",
"exam_descriptor",
"=",
"modulestore",
"(",
")",
".",
"get_item",
"(",
"exam_key",
")",
"def",
"inner_get_module",
"(",
"descriptor",
")",
":",
"from",
"courseware",
".",
"module_render",
"import",
"get_module_for_descriptor",
"field_data_cache",
"=",
"FieldDataCache",
"(",
"[",
"descriptor",
"]",
",",
"course",
".",
"id",
",",
"request",
".",
"user",
")",
"return",
"get_module_for_descriptor",
"(",
"request",
".",
"user",
",",
"request",
",",
"descriptor",
",",
"field_data_cache",
",",
"course",
".",
"id",
",",
"course",
"=",
"course",
")",
"exam_module_generators",
"=",
"yield_dynamic_descriptor_descendants",
"(",
"exam_descriptor",
",",
"request",
".",
"user",
".",
"id",
",",
"inner_get_module",
")",
"exam_modules",
"=",
"[",
"module",
"for",
"module",
"in",
"exam_module_generators",
"]",
"return",
"_calculate_entrance_exam_score",
"(",
"request",
".",
"user",
",",
"course",
",",
"exam_modules",
")"
] | gather the set of modules which comprise the entrance exam note that request may not actually be a genuine request . | train | false |
124 | def load_passphrase_from_file():
vf_path = os.path.expanduser(kVFPassphraseFile)
assert (os.access(vf_path, os.F_OK) and os.access(vf_path, os.R_OK)), ('%s must exist and be readable' % vf_path)
with open(vf_path) as f:
user_data = f.read()
return user_data.strip('\n')
| [
"def",
"load_passphrase_from_file",
"(",
")",
":",
"vf_path",
"=",
"os",
".",
"path",
".",
"expanduser",
"(",
"kVFPassphraseFile",
")",
"assert",
"(",
"os",
".",
"access",
"(",
"vf_path",
",",
"os",
".",
"F_OK",
")",
"and",
"os",
".",
"access",
"(",
"vf_path",
",",
"os",
".",
"R_OK",
")",
")",
",",
"(",
"'%s must exist and be readable'",
"%",
"vf_path",
")",
"with",
"open",
"(",
"vf_path",
")",
"as",
"f",
":",
"user_data",
"=",
"f",
".",
"read",
"(",
")",
"return",
"user_data",
".",
"strip",
"(",
"'\\n'",
")"
] | read the viewfinder passphrase from local file . | train | false |
125 | def describe_identity_pools(IdentityPoolName, IdentityPoolId=None, region=None, key=None, keyid=None, profile=None):
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
try:
ids = _find_identity_pool_ids(IdentityPoolName, IdentityPoolId, conn)
if ids:
results = []
for pool_id in ids:
response = conn.describe_identity_pool(IdentityPoolId=pool_id)
response.pop('ResponseMetadata', None)
results.append(response)
return {'identity_pools': results}
else:
return {'identity_pools': None}
except ClientError as e:
return {'error': salt.utils.boto3.get_error(e)}
| [
"def",
"describe_identity_pools",
"(",
"IdentityPoolName",
",",
"IdentityPoolId",
"=",
"None",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")",
"try",
":",
"ids",
"=",
"_find_identity_pool_ids",
"(",
"IdentityPoolName",
",",
"IdentityPoolId",
",",
"conn",
")",
"if",
"ids",
":",
"results",
"=",
"[",
"]",
"for",
"pool_id",
"in",
"ids",
":",
"response",
"=",
"conn",
".",
"describe_identity_pool",
"(",
"IdentityPoolId",
"=",
"pool_id",
")",
"response",
".",
"pop",
"(",
"'ResponseMetadata'",
",",
"None",
")",
"results",
".",
"append",
"(",
"response",
")",
"return",
"{",
"'identity_pools'",
":",
"results",
"}",
"else",
":",
"return",
"{",
"'identity_pools'",
":",
"None",
"}",
"except",
"ClientError",
"as",
"e",
":",
"return",
"{",
"'error'",
":",
"salt",
".",
"utils",
".",
"boto3",
".",
"get_error",
"(",
"e",
")",
"}"
] | given an identity pool name . | train | false |
126 | @register.simple_tag(takes_context=True)
def zinnia_loop_template(context, default_template):
(matching, context_object) = get_context_first_matching_object(context, ['category', 'tag', 'author', 'pattern', 'year', 'month', 'week', 'day'])
context_positions = get_context_loop_positions(context)
templates = loop_template_list(context_positions, context_object, matching, default_template, ENTRY_LOOP_TEMPLATES)
return select_template(templates)
| [
"@",
"register",
".",
"simple_tag",
"(",
"takes_context",
"=",
"True",
")",
"def",
"zinnia_loop_template",
"(",
"context",
",",
"default_template",
")",
":",
"(",
"matching",
",",
"context_object",
")",
"=",
"get_context_first_matching_object",
"(",
"context",
",",
"[",
"'category'",
",",
"'tag'",
",",
"'author'",
",",
"'pattern'",
",",
"'year'",
",",
"'month'",
",",
"'week'",
",",
"'day'",
"]",
")",
"context_positions",
"=",
"get_context_loop_positions",
"(",
"context",
")",
"templates",
"=",
"loop_template_list",
"(",
"context_positions",
",",
"context_object",
",",
"matching",
",",
"default_template",
",",
"ENTRY_LOOP_TEMPLATES",
")",
"return",
"select_template",
"(",
"templates",
")"
] | return a selected template from his position within a loop and the filtering context . | train | false |
127 | def loadExperimentDescriptionScriptFromDir(experimentDir):
descriptionScriptPath = os.path.join(experimentDir, 'description.py')
module = _loadDescriptionFile(descriptionScriptPath)
return module
| [
"def",
"loadExperimentDescriptionScriptFromDir",
"(",
"experimentDir",
")",
":",
"descriptionScriptPath",
"=",
"os",
".",
"path",
".",
"join",
"(",
"experimentDir",
",",
"'description.py'",
")",
"module",
"=",
"_loadDescriptionFile",
"(",
"descriptionScriptPath",
")",
"return",
"module"
] | loads the experiment description python script from the given experiment directory . | train | true |
128 | def token_list_to_text(tokenlist):
ZeroWidthEscape = Token.ZeroWidthEscape
return u''.join((item[1] for item in tokenlist if (item[0] != ZeroWidthEscape)))
| [
"def",
"token_list_to_text",
"(",
"tokenlist",
")",
":",
"ZeroWidthEscape",
"=",
"Token",
".",
"ZeroWidthEscape",
"return",
"u''",
".",
"join",
"(",
"(",
"item",
"[",
"1",
"]",
"for",
"item",
"in",
"tokenlist",
"if",
"(",
"item",
"[",
"0",
"]",
"!=",
"ZeroWidthEscape",
")",
")",
")"
] | concatenate all the text parts again . | train | true |
129 | @profiler.trace
def transfer_list(request, detailed=True, search_opts=None):
c_client = cinderclient(request)
try:
return [VolumeTransfer(v) for v in c_client.transfers.list(detailed=detailed, search_opts=search_opts)]
except cinder_exception.Forbidden as error:
LOG.error(error)
return []
| [
"@",
"profiler",
".",
"trace",
"def",
"transfer_list",
"(",
"request",
",",
"detailed",
"=",
"True",
",",
"search_opts",
"=",
"None",
")",
":",
"c_client",
"=",
"cinderclient",
"(",
"request",
")",
"try",
":",
"return",
"[",
"VolumeTransfer",
"(",
"v",
")",
"for",
"v",
"in",
"c_client",
".",
"transfers",
".",
"list",
"(",
"detailed",
"=",
"detailed",
",",
"search_opts",
"=",
"search_opts",
")",
"]",
"except",
"cinder_exception",
".",
"Forbidden",
"as",
"error",
":",
"LOG",
".",
"error",
"(",
"error",
")",
"return",
"[",
"]"
] | to see all volumes transfers as an admin pass in a special search option: {all_tenants: 1} . | train | true |
131 | def test_discretize_callable_1d():
def f(x):
return (x ** 2)
y = discretize_model(f, ((-5), 6))
assert_allclose(y, (np.arange((-5), 6) ** 2))
| [
"def",
"test_discretize_callable_1d",
"(",
")",
":",
"def",
"f",
"(",
"x",
")",
":",
"return",
"(",
"x",
"**",
"2",
")",
"y",
"=",
"discretize_model",
"(",
"f",
",",
"(",
"(",
"-",
"5",
")",
",",
"6",
")",
")",
"assert_allclose",
"(",
"y",
",",
"(",
"np",
".",
"arange",
"(",
"(",
"-",
"5",
")",
",",
"6",
")",
"**",
"2",
")",
")"
] | test discretize when a 1d function is passed . | train | false |
End of preview. Expand
in Dataset Viewer.
- Downloads last month
- 35