name
stringlengths 1
94
| class_name
stringlengths 1
56
| class_bases
stringlengths 0
73
⌀ | is_member
bool 2
classes | args
stringlengths 0
1.04k
| class_docstr
stringlengths 0
13k
⌀ | class_docstr_tok
stringlengths 2
18.2k
⌀ | docstr
stringlengths 0
6.72k
⌀ | docstr_tok
stringlengths 13
9.01k
⌀ | returns
stringlengths 1
279
⌀ | code
stringlengths 20
27.8k
⌀ | code_tok
stringlengths 32
38.4k
⌀ | lstart
int64 1
1.03k
| lend
int64 2
1.18k
| raises
stringclasses 9
values | filename
stringlengths 5
81
| file_path
stringlengths 14
129
| imports
stringlengths 0
786
| total_objects
int64 7
10
| num_classes
float64 1
4
⌀ | num_imports
int64 0
9
| num_all_bases
float64 0
4
⌀ | num_bases
float64 1
4
⌀ | num_functions
int64 0
10
| num_methods
float64 1
8
⌀ | label_desc
stringlengths 70
1.25k
| label_desc_len
int64 70
1.25k
| label_id
stringclasses 10
values | __index_level_0__
int64 0
514k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
test_da_tokenizer_handles_dates | global | null | false | da_tokenizer,text | null | null | null | null | null | def test_da_tokenizer_handles_dates(da_tokenizer, text):
tokens = da_tokenizer(text)
assert len(tokens) == 1
| ["def","test_da_tokenizer_handles_dates","(","da_tokenizer",",","text",")",":","tokens","=","da_tokenizer","(","text",")","assert","len","(","tokens",")","==","1"] | 17 | 19 | null | test_exceptions.py | spacy/spacy/tests/lang/da/test_exceptions.py | import pytest | 7 | null | 1 | null | null | 6 | null | Use image node_id 3 for calling a global function with example usage: test_da_tokenizer_handles_dates(da_tokenizer, text) without return types | 142 | node_id 3 | 387,996 |
test_da_tokenizer_handles_exc_in_text | global | null | false | da_tokenizer | null | null | null | null | null | def test_da_tokenizer_handles_exc_in_text(da_tokenizer):
text = "Det er bl.a. ikke meningen"
tokens = da_tokenizer(text)
assert len(tokens) == 5
assert tokens[2].text == "bl.a."
| ["def","test_da_tokenizer_handles_exc_in_text","(","da_tokenizer",")",":","text","=","``","Det","er","bl.a",".","ikke","meningen","''","tokens","=","da_tokenizer","(","text",")","assert","len","(","tokens",")","==","5","assert","tokens","[","2","]",".text","==","``","bl.a",".","''"] | 22 | 26 | null | test_exceptions.py | spacy/spacy/tests/lang/da/test_exceptions.py | import pytest | 7 | null | 1 | null | null | 6 | null | Use image node_id 4 for calling a global function with example usage: test_da_tokenizer_handles_exc_in_text(da_tokenizer) without return types | 142 | node_id 4 | 387,997 |
test_da_tokenizer_handles_custom_base_exc | global | null | false | da_tokenizer | null | null | null | null | null | def test_da_tokenizer_handles_custom_base_exc(da_tokenizer):
text = "Her er noget du kan kigge i."
tokens = da_tokenizer(text)
assert len(tokens) == 8
assert tokens[6].text == "i"
assert tokens[7].text == "."
| ["def","test_da_tokenizer_handles_custom_base_exc","(","da_tokenizer",")",":","text","=","``","Her","er","noget","du","kan","kigge","i",".","''","tokens","=","da_tokenizer","(","text",")","assert","len","(","tokens",")","==","8","assert","tokens","[","6","]",".text","==","``","i","''","assert","tokens","[","7","]",".text","==","``",".","''"] | 29 | 34 | null | test_exceptions.py | spacy/spacy/tests/lang/da/test_exceptions.py | import pytest | 7 | null | 1 | null | null | 6 | null | Use image node_id 5 for calling a global function with example usage: test_da_tokenizer_handles_custom_base_exc(da_tokenizer) without return types | 146 | node_id 5 | 387,998 |
test_da_tokenizer_slash | global | null | false | da_tokenizer,text,n_tokens | null | null | null | null | null | def test_da_tokenizer_slash(da_tokenizer, text, n_tokens):
tokens = da_tokenizer(text)
assert len(tokens) == n_tokens
| ["def","test_da_tokenizer_slash","(","da_tokenizer",",","text",",","n_tokens",")",":","tokens","=","da_tokenizer","(","text",")","assert","len","(","tokens",")","==","n_tokens"] | 57 | 59 | null | test_exceptions.py | spacy/spacy/tests/lang/da/test_exceptions.py | import pytest | 7 | null | 1 | null | null | 6 | null | Use image node_id 6 for calling a global function with example usage: test_da_tokenizer_slash(da_tokenizer, text, n_tokens) without return types | 144 | node_id 6 | 387,999 |
debug | ExecMerger | Executor | true | self,docs_matrix | null | null | null | null | result | def debug(self, docs_matrix: List[DocumentArray], **kwargs):
self.logger.debug(
f"received doc matrix in exec-merger with length {len(docs_matrix)}."
)
result = DocumentArray()
for docs in zip(*docs_matrix):
traversed_executors = [
doc.tags["traversed-executors"] for doc in docs
]
shard_ids = [doc.tags["shard_id"] for doc in docs]
shards = [doc.tags["shards"] for doc in docs]
parallels = [doc.tags["parallel"] for doc in docs]
traversed_executors = list(chain(*traversed_executors))
doc = Document()
doc.tags["traversed-executors"] = traversed_executors
doc.tags["shard_id"] = shard_ids
doc.tags["shards"] = shards
doc.tags["parallel"] = parallels
doc.tags["merged"] = True
result.append(doc)
return result
| ["def","debug","(","self",",","docs_matrix",":","List","[","DocumentArray","]",",","*","*","kwargs",")",":","self.logger.debug","(","f","''","received","doc","matrix","in","exec-merger","with","length","{","len","(","docs_matrix",")","}",".","''",")","result","=","DocumentArray","(",")","for","docs","in","zip","(","*","docs_matrix",")",":","traversed_executors","=","[","doc.tags","[","``","traversed-executors","''","]","for","doc","in","docs","]","shard_ids","=","[","doc.tags","[","``","shard_id","''","]","for","doc","in","docs","]","shards","=","[","doc.tags","[","``","shards","''","]","for","doc","in","docs","]","parallels","=","[","doc.tags","[","``","parallel","''","]","for","doc","in","docs","]","traversed_executors","=","list","(","chain","(","*","traversed_executors",")",")","doc","=","Document","(",")","doc.tags","[","``","traversed-executors","''","]","=","traversed_executors","doc.tags","[","``","shard_id","''","]","=","shard_ids","doc.tags","[","``","shards","''","]","=","shards","doc.tags","[","``","parallel","''","]","=","parallels","doc.tags","[","``","merged","''","]","=","True","result.append","(","doc",")","return","result"] | 15 | 36 | null | exec_merger.py | jina/tests/docker_compose/executor-merger/exec_merger.py | from itertools import chain
from typing import Dict, List
from jina import Document, DocumentArray, Executor, requests | 7 | 1 | 3 | 1 | 1 | 0 | 2 | Use image node_id 2 for calling the ExecMerger obj's underlying member method code with example usage: obj.debug(docs_matrix) and returns: result | 145 | node_id 2 | 156,984 |
__init__ | ExecMerger | Executor | true | self | null | null | null | null | ExecMerger | def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
from jina.logging.logger import JinaLogger
self.logger = JinaLogger(self.__class__.__name__)
| ["def","__init__","(","self",",","*","args",",","*","*","kwargs",")",":","super","(",")",".__init__","(","*","args",",","*","*","kwargs",")","from","jina.logging.logger","import","JinaLogger","self.logger","=","JinaLogger","(","self.__class__.__name__",")"] | 8 | 12 | null | exec_merger.py | jina/tests/docker_compose/executor-merger/exec_merger.py | from itertools import chain
from typing import Dict, List
from jina import Document, DocumentArray, Executor, requests | 7 | 1 | 3 | 1 | 1 | 0 | 2 | Use image node_id 1 to create a new ExecMerger object from inherited base classes: Executor with example: obj = ExecMerger() | 124 | node_id 1 | 156,983 |
_qencode | global | null | false | s | null | null | null | null | enc | def _qencode(s):
enc = _encodestring(s, quotetabs=True)
# Must encode spaces, which quopri.encodestring() doesn't do
return enc.replace(b" ", b"=20")
| ["def","_qencode","(","s",")",":","enc","=","_encodestring","(","s",",","quotetabs=True",")","#","Must","encode","spaces",",","which","quopri.encodestring","(",")","does","n't","do","return","enc.replace","(","b","''","``",",","b","''","=20","''",")"] | 19 | 22 | null | encoders.py | cpython/Lib/email/encoders.py | from base64 import encodebytes
from quopri import encodestring | 7 | null | 2 | null | null | 5 | null | Use image node_id 1 for calling a global function with example usage: _qencode(s) and returns: enc | 98 | node_id 1 | 52,444 |
run | global | null | false | cmd,stdout | null | null | null | null | null | def run(cmd, stdout=True):
pipe = None if stdout else subprocess.DEVNULL
try:
subprocess.check_call(cmd, stdout=pipe, stderr=pipe)
except subprocess.CalledProcessError:
print("\n! Error executing: `%s;` aborting" % " ".join(cmd))
sys.exit(1)
| ["def","run","(","cmd",",","stdout=True",")",":","pipe","=","None","if","stdout","else","subprocess.DEVNULL","try",":","subprocess.check_call","(","cmd",",","stdout=pipe",",","stderr=pipe",")","except","subprocess.CalledProcessError",":","print","(","``","\\n","!","Error","executing",":","`","%","s",";","`","aborting","''","%","``","``",".join","(","cmd",")",")","sys.exit","(","1",")"] | 42 | 48 | null | push_docs_to_repo.py | numpy/tools/ci/push_docs_to_repo.py | import argparse
import subprocess
import tempfile
import os
import sys
import shutil | 7 | null | 6 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: run(cmd, stdout) without return types | 107 | node_id 1 | 214,130 |
intlist | global | null | false | obj | null | null | null | null | unknown | def intlist(obj):
return [int(v) for v in obj]
| ["def","intlist","(","obj",")",":","return","[","int","(","v",")","for","v","in","obj","]"] | 14 | 15 | null | test_bloom.py | redis-py/tests/test_asyncio/test_bloom.py | from math import inf
import pytest
import redis.asyncio
from redis.exceptions import ModuleError, RedisError
from redis.utils import HIREDIS_AVAILABLE
from tests.conftest import assert_resp_response, is_resp2_connection, skip_ifmodversion_lt | 7 | null | 6 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: intlist(obj) and returns: unknown | 103 | node_id 1 | 339,428 |
create_item | global | null | false | item | null | null | null | null | item | def create_item(item: Item):
return item
| ["def","create_item","(","item",":","Item",")",":","return","item"] | 16 | 17 | null | tutorial001.py | fastapi/docs_src/separate_openapi_schemas/tutorial001.py | from typing import List, Union
from fastapi import FastAPI
from pydantic import BaseModel | 7 | null | 3 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: create_item(item) and returns: item | 105 | node_id 1 | 135,814 |
parse_args | global | null | false | null | null | null | null | parser | def parse_args():
"""
Helper function parsing the command line options
@retval ArgumentParser
"""
parser = ArgumentParser(
description=(
"PyTorch TPU distributed training launch "
"helper utility that will spawn up "
"multiple distributed processes"
)
)
# Optional arguments for the launch helper
parser.add_argument(
"--num_cores",
type=int,
default=1,
help="Number of TPU cores to use (1 or 8).",
)
# positional
parser.add_argument(
"training_script",
type=str,
help=(
"The full path to the single TPU training "
"program/script to be launched in parallel, "
"followed by all the arguments for the "
"training script"
),
)
# rest from the training program
parser.add_argument("training_script_args", nargs=REMAINDER)
return parser.parse_args()
| ["def","parse_args","(",")",":","``","''","''","Helper","function","parsing","the","command","line","options","@","retval","ArgumentParser","``","''","''","parser","=","ArgumentParser","(","description=","(","``","PyTorch","TPU","distributed","training","launch","``","``","helper","utility","that","will","spawn","up","``","``","multiple","distributed","processes","''",")",")","#","Optional","arguments","for","the","launch","helper","parser.add_argument","(","``","--","num_cores","''",",","type=int",",","default=1",",","help=","''","Number","of","TPU","cores","to","use","(","1","or","8",")",".","``",",",")","#","positional","parser.add_argument","(","``","training_script","''",",","type=str",",","help=","(","``","The","full","path","to","the","single","TPU","training","``","``","program\/script","to","be","launched","in","parallel",",","``","``","followed","by","all","the","arguments","for","the","``","``","training","script","''",")",",",")","#","rest","from","the","training","program","parser.add_argument","(","``","training_script_args","''",",","nargs=REMAINDER",")","return","parser.parse_args","(",")"] | 36 | 67 | null | xla_spawn.py | accelerate/tests/xla_spawn.py | import importlib
import sys
from argparse import REMAINDER, ArgumentParser
from pathlib import Path
import torch_xla.distributed.xla_multiprocessing | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: parse_args() and returns: parser | 102 | node_id 1 | 1,215 |
|
test_suite | global | null | false | null | null | null | null | unittest | def test_suite():
return unittest.TestLoader().loadTestsFromName(__name__)
| ["def","test_suite","(",")",":","return","unittest.TestLoader","(",")",".loadTestsFromName","(","__name__",")"] | 43 | 44 | null | test_bugX000.py | psycopg2/tests/test_bugX000.py | import psycopg2
import time
import unittest | 7 | null | 3 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: test_suite() and returns: unittest | 104 | node_id 1 | 256,395 |
|
test_date_time_allocation_bug | DateTimeAllocationBugTestCase | unittest | true | self | null | null | null | null | null | def test_date_time_allocation_bug(self):
d1 = psycopg2.Date(2002, 12, 25)
d2 = psycopg2.DateFromTicks(
time.mktime((2002, 12, 25, 0, 0, 0, 0, 0, 0))
)
t1 = psycopg2.Time(13, 45, 30)
t2 = psycopg2.TimeFromTicks(
time.mktime((2001, 1, 1, 13, 45, 30, 0, 0, 0))
)
t1 = psycopg2.Timestamp(2002, 12, 25, 13, 45, 30)
t2 = psycopg2.TimestampFromTicks(
time.mktime((2002, 12, 25, 13, 45, 30, 0, 0, 0))
)
del d1, d2, t1, t2
| ["def","test_date_time_allocation_bug","(","self",")",":","d1","=","psycopg2.Date","(","2002",",","12",",","25",")","d2","=","psycopg2.DateFromTicks","(","time.mktime","(","(","2002",",","12",",","25",",","0",",","0",",","0",",","0",",","0",",","0",")",")",")","t1","=","psycopg2.Time","(","13",",","45",",","30",")","t2","=","psycopg2.TimeFromTicks","(","time.mktime","(","(","2001",",","1",",","1",",","13",",","45",",","30",",","0",",","0",",","0",")",")",")","t1","=","psycopg2.Timestamp","(","2002",",","12",",","25",",","13",",","45",",","30",")","t2","=","psycopg2.TimestampFromTicks","(","time.mktime","(","(","2002",",","12",",","25",",","13",",","45",",","30",",","0",",","0",",","0",")",")",")","del","d1",",","d2",",","t1",",","t2"] | 32 | 40 | null | test_bugX000.py | psycopg2/tests/test_bugX000.py | import psycopg2
import time
import unittest | 7 | 1 | 3 | 1 | 1 | 1 | 1 | Use image node_id 1 for calling the DateTimeAllocationBugTestCase obj's underlying member method code with example usage: obj.test_date_time_allocation_bug() without return types | 178 | node_id 1 | 256,394 |
make_ax | global | null | false | grid | null | null | null | null | ax | def make_ax(grid=False):
ax = plt.figure().add_subplot(projection="3d")
ax.grid(grid)
ax.set_axis_off()
return ax
| ["def","make_ax","(","grid=False",")",":","ax","=","plt.figure","(",")",".add_subplot","(","projection=","''","3d","''",")","ax.grid","(","grid",")","ax.set_axis_off","(",")","return","ax"] | 75 | 79 | null | plot_euler_number.py | scikit-image/doc/examples/segmentation/plot_euler_number.py | from skimage.measure import euler_number, label
import matplotlib.pyplot
import numpy | 7 | null | 3 | null | null | 4 | null | Use image node_id 1 for calling a global function with example usage: make_ax(grid) and returns: ax | 99 | node_id 1 | 344,193 |
explode | global | null | false | data | null | null | null | null | data_e | def explode(data):
"""visualization to separate voxels
Data voxels are separated by 0-valued ones so that they appear
separated in the matplotlib figure.
"""
size = np.array(data.shape) * 2
data_e = np.zeros(size - 1, dtype=data.dtype)
data_e[::2, ::2, ::2] = data
return data_e
| ["def","explode","(","data",")",":","``","''","''","visualization","to","separate","voxels","Data","voxels","are","separated","by","0-valued","ones","so","that","they","appear","separated","in","the","matplotlib","figure.","``","''","''","size","=","np.array","(","data.shape",")","*","2","data_e","=","np.zeros","(","size","-","1",",","dtype=data.dtype",")","data_e","[",":",":2",",",":",":2",",",":",":2","]","=","data","return","data_e"] | 82 | 91 | null | plot_euler_number.py | scikit-image/doc/examples/segmentation/plot_euler_number.py | from skimage.measure import euler_number, label
import matplotlib.pyplot
import numpy | 7 | null | 3 | null | null | 4 | null | Use image node_id 2 for calling a global function with example usage: explode(data) and returns: data_e | 103 | node_id 2 | 344,194 |
cli | global | null | false | null | null | null | null | null | def cli():
"""Commandline interface."""
parser = argparse.ArgumentParser(
formatter_class=argparse.ArgumentDefaultsHelpFormatter
)
parser.add_argument(
"--text", type=str, help="text to be turned into audio"
)
parser.add_argument(
"--output_filename",
type=str,
default="bark_generation.wav",
help="output audio file name",
)
parser.add_argument(
"--output_dir",
type=str,
default=".",
help="directory to save the outputs",
)
parser.add_argument(
"--history_prompt",
type=str,
default=None,
help="history choice for audio cloning, be path to the .npz file.",
)
parser.add_argument(
"--text_temp",
default=0.7,
type=float,
help="generation temperature (1.0 more diverse, 0.0 more conservative)",
)
parser.add_argument(
"--waveform_temp",
default=0.7,
type=float,
help="generation temperature (1.0 more diverse, 0.0 more conservative)",
)
parser.add_argument(
"--silent",
default=False,
type=bool,
help="disable progress bar",
)
parser.add_argument(
"--output_full",
default=False,
type=bool,
help="return full generation to be used as a history prompt",
)
args = vars(parser.parse_args())
input_text: str = args.get("text")
output_filename: str = args.get("output_filename")
output_dir: str = args.get("output_dir")
history_prompt: str = args.get("history_prompt")
text_temp: float = args.get("text_temp")
waveform_temp: float = args.get("waveform_temp")
silent: bool = args.get("silent")
output_full: bool = args.get("output_full")
try:
os.makedirs(output_dir, exist_ok=True)
generated_audio = generate_audio(
input_text,
history_prompt=history_prompt,
text_temp=text_temp,
waveform_temp=waveform_temp,
silent=silent,
output_full=output_full,
)
output_file_path = os.path.join(output_dir, output_filename)
write_wav(output_file_path, SAMPLE_RATE, generated_audio)
print(
f"Done! Output audio file is saved at: '{output_file_path}'"
)
except Exception as e:
print(f"Oops, an error occurred: {e}")
| ["def","cli","(",")",":","``","''","''","Commandline","interface",".","''","''","''","parser","=","argparse.ArgumentParser","(","formatter_class=argparse.ArgumentDefaultsHelpFormatter",")","parser.add_argument","(","``","--","text","''",",","type=str",",","help=","''","text","to","be","turned","into","audio","''",")","parser.add_argument","(","``","--","output_filename","''",",","type=str",",","default=","''","bark_generation.wav","''",",","help=","''","output","audio","file","name","''",",",")","parser.add_argument","(","``","--","output_dir","''",",","type=str",",","default=","''",".","``",",","help=","''","directory","to","save","the","outputs","''",",",")","parser.add_argument","(","``","--","history_prompt","''",",","type=str",",","default=None",",","help=","''","history","choice","for","audio","cloning",",","be","path","to","the",".npz","file",".","``",",",")","parser.add_argument","(","``","--","text_temp","''",",","default=0.7",",","type=float",",","help=","''","generation","temperature","(","1.0","more","diverse",",","0.0","more","conservative",")","''",",",")","parser.add_argument","(","``","--","waveform_temp","''",",","default=0.7",",","type=float",",","help=","''","generation","temperature","(","1.0","more","diverse",",","0.0","more","conservative",")","''",",",")","parser.add_argument","(","``","--","silent","''",",","default=False",",","type=bool",",","help=","''","disable","progress","bar","''",",",")","parser.add_argument","(","``","--","output_full","''",",","default=False",",","type=bool",",","help=","''","return","full","generation","to","be","used","as","a","history","prompt","''",",",")","args","=","vars","(","parser.parse_args","(",")",")","input_text",":","str","=","args.get","(","``","text","''",")","output_filename",":","str","=","args.get","(","``","output_filename","''",")","output_dir",":","str","=","args.get","(","``","output_dir","''",")","history_prompt",":","str","=","args.get","(","``","history_prompt","''",")","text_temp",":","float","=","args.get","(","``","text_temp","''",")","waveform_temp",":","float","=","args.get","(","``","waveform_temp","''",")","silent",":","bool","=","args.get","(","``","silent","''",")","output_full",":","bool","=","args.get","(","``","output_full","''",")","try",":","os.makedirs","(","output_dir",",","exist_ok=True",")","generated_audio","=","generate_audio","(","input_text",",","history_prompt=history_prompt",",","text_temp=text_temp",",","waveform_temp=waveform_temp",",","silent=silent",",","output_full=output_full",",",")","output_file_path","=","os.path.join","(","output_dir",",","output_filename",")","write_wav","(","output_file_path",",","SAMPLE_RATE",",","generated_audio",")","print","(","f","''","Done","!","Output","audio","file","is","saved","at",":","'","{","output_file_path","}","'","''",")","except","Exception","as","e",":","print","(","f","''","Oops",",","an","error","occurred",":","{","e","}","''",")"] | 10 | 71 | null | cli.py | ai-audio-speech/bark/cli.py | import argparse
from typing import Dict, Optional, Union
import os
from scipy.io.wavfile import write
from .api import generate_audio
from .generation import SAMPLE_RATE | 7 | null | 6 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: cli() without return types | 96 | node_id 1 | 1,222 |
|
fake_decode_token | global | null | false | token | null | null | null | null | User | def fake_decode_token(token):
return User(
username=token + "fakedecoded",
email="john@example.com",
full_name="John Doe",
)
| ["def","fake_decode_token","(","token",")",":","return","User","(","username=token","+","``","fakedecoded","''",",","email=","''","john","@","example.com","''",",","full_name=","''","John","Doe","''",",",")"] | 19 | 22 | null | tutorial002_an_py310.py | fastapi/docs_src/security/tutorial002_an_py310.py | from typing import Annotated
from fastapi import Depends, FastAPI
from fastapi.security import OAuth2PasswordBearer
from pydantic import BaseModel | 7 | null | 4 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: fake_decode_token(token) and returns: User | 112 | node_id 1 | 135,732 |
expand_coordinates | global | null | false | indices | null | null | null | null | x, y, z | def expand_coordinates(indices):
"""
This collapses together pairs of indices, so that
the gaps in the volume array will have a zero width.
"""
x, y, z = indices
x[1::2, :, :] += 1
y[:, 1::2, :] += 1
z[:, :, 1::2] += 1
return x, y, z
| ["def","expand_coordinates","(","indices",")",":","``","''","''","This","collapses","together","pairs","of","indices",",","so","that","the","gaps","in","the","volume","array","will","have","a","zero","width.","``","''","''","x",",","y",",","z","=","indices","x","[","1",":",":2",",",":",",",":","]","+=","1","y","[",":",",","1",":",":2",",",":","]","+=","1","z","[",":",",",":",",","1",":",":2","]","+=","1","return","x",",","y",",","z"] | 97 | 106 | null | plot_euler_number.py | scikit-image/doc/examples/segmentation/plot_euler_number.py | from skimage.measure import euler_number, label
import matplotlib.pyplot
import numpy | 7 | null | 3 | null | null | 4 | null | Use image node_id 3 for calling a global function with example usage: expand_coordinates(indices) and returns: x, y, z | 120 | node_id 3 | 344,195 |
prepare_tp_fused_qkvw | global | null | false | module,src,mp_size,gpu_index | null | null | null | null | _bloom_type_transpose,tp_fuseqkv_weight,split_fusedqkv,split_fusedqkv,_glm_type_transpose,torch,_bloom_type_transpose,_transpose_fused_qkvw,_codegen_type_transpose,_glm_type_transpose,_qwen_type_transpose,_bigcode_type_transpose | def prepare_tp_fused_qkvw(module, src, mp_size, gpu_index):
module_str = str(module).strip()
if src is None:
return
fused_type_dict = {
"CodeGenBlock": "codegentype",
"BloomBlock": "bloomtype",
"GLMBlock": "glmtype",
"MPTBlock": "glmtype",
"MptBlock": "glmtype",
"BaichuanLayer": "glmtype",
"DecoderLayer": "glmtype",
"QWenBlock": "qwentype",
"FalconDecoderLayer": "bloomtype",
"GPTBigCodeBlock": "bigcodetype",
}
def _codegen_type_transpose(input, mp_size, codegen_mp_num=4):
# codegen_mp_num defined in https://github.com/huggingface/transformers/blob/main/src/transformers/models/codegen/modeling_codegen.py
assert (
get_num_kv_heads() % (mp_size * codegen_mp_num) == 0
), "codgen autoTP requires num_kv_heads % (mp_size*codegen_mp_num) == 0"
# input : [3*hidden_dim, hidden_dim](weight) or [3*hidden_dim](bias)
shape = input.shape
dst_shape = get_shard_size(shape[0], mp_size)
num_mp_blocks = input.reshape(
codegen_mp_num, shape[0] // codegen_mp_num, shape[1]
)
# num_mp_blocks : [codegen_mp_num, 3*hidden_dim/codegen_mp_num, :]
src_split = list(
torch.split(
num_mp_blocks, num_mp_blocks.shape[1] // 3, dim=1
)
)
src_split = [
x.reshape(codegen_mp_num * mp_size, -1, shape[1])
for x in src_split
]
split_fusedqkv = split_by_qkvlist_and_refuse(
src_split, get_shard_size(shape[0] // 3, mp_size), 0, 1
)
tp_fuseqkv_weight = torch.cat(split_fusedqkv, dim=0).reshape(
shape[0], -1
)
return tp_fuseqkv_weight[
gpu_index * dst_shape : (gpu_index + 1) * dst_shape
]
def _glm_type_transpose(input, mp_size):
# input : [3*hidden_dim, hidden_dim](weight) or [3*hidden_dim](bias)
shape = input.shape
src_split = torch.split(input, shape[0] // 3, dim=0)
split_fusedqkv = split_by_qkvlist_and_refuse(
src_split, get_shard_size_list(shape[0] // 3, mp_size)
)
return split_fusedqkv[gpu_index]
def _bloom_type_transpose(input, mp_size):
shape = input.shape
split_fusedqkv = input.split(
get_shard_size_list(shape[0], mp_size), dim=0
)
return split_fusedqkv[gpu_index]
def _qwen_type_transpose(input, mp_size, module):
if not hasattr(module, "_ds_fusedqkv_entered"):
# Adjust splitting absolute value variables
setattr(module, "_ds_fusedqkv_entered", True)
module.attn.split_size = get_shard_size(
module.attn.split_size, mp_size
)
return _glm_type_transpose(input, mp_size)
def _bigcode_type_transpose(input, mp_size):
n_embd = get_n_embd()
q = input[:n_embd]
kv = input[n_embd:]
shape = q.shape
split_q = q.split(
get_shard_size_list(shape[0], mp_size), dim=0
)
return torch.cat((split_q[gpu_index], kv), dim=0)
def _transpose_fused_qkvw(
src, mp_size, fused_qkv_type=None, module=None
):
# suppose num_heads=n, q(n)_w means the n-th q head linear weight, the weight format are as following
# bloomtype: [q(1)_w,k(1)_w,v(1)_w,q(2)_w,k(2)_w,v(2)_w,...,q(n)_w,k(n)_w,v(n)_w]
# glmtype: [q(1)_w, q(2)_w,...,q(n)_w,k(1)_w,k(2)_w,...,k(n)_w,v(1)_w,v(2)_w,...,v(n)_w]
# codegentype: [q(1)_w,q(2)_w,...,q(n/t)_w,k(1)_w,k(2)_w,...,k(n/t)_w,v(1)_2,v(2)_w,...v(n/t)_w,q(n/t+1)_w,...], where t is a const defined in model file.
if fused_qkv_type == "bloomtype":
return _bloom_type_transpose(src, mp_size)
elif fused_qkv_type == "codegentype":
return _codegen_type_transpose(src, mp_size)
elif fused_qkv_type == "glmtype":
return _glm_type_transpose(src, mp_size)
elif fused_qkv_type == "qwentype":
return _qwen_type_transpose(src, mp_size, module)
elif fused_qkv_type == "bigcodetype":
return _bigcode_type_transpose(src, mp_size)
raise ValueError("unknown fused_qkv_type")
for module_name, fused_type in fused_type_dict.items():
if re.search(module_name, module_str):
return _transpose_fused_qkvw(
src, mp_size, fused_type, module
)
warning_once(
f"Unrecognized fusedkqv weight type, default to using bloom type,"
f"please check in prepare_tp_fused_qkvw() to avoid potential calculation errors"
)
return _bloom_type_transpose(src, mp_size)
| ["def","prepare_tp_fused_qkvw","(","module",",","src",",","mp_size",",","gpu_index",")",":","module_str","=","str","(","module",")",".strip","(",")","if","src","is","None",":","return","fused_type_dict","=","{","``","CodeGenBlock","''",":","``","codegentype","''",",","``","BloomBlock","''",":","``","bloomtype","''",",","``","GLMBlock","''",":","``","glmtype","''",",","``","MPTBlock","''",":","``","glmtype","''",",","``","MptBlock","''",":","``","glmtype","''",",","``","BaichuanLayer","''",":","``","glmtype","''",",","``","DecoderLayer","''",":","``","glmtype","''",",","``","QWenBlock","''",":","``","qwentype","''",",","``","FalconDecoderLayer","''",":","``","bloomtype","''",",","``","GPTBigCodeBlock","''",":","``","bigcodetype","''",",","}","def","_codegen_type_transpose","(","input",",","mp_size",",","codegen_mp_num=4",")",":","#","codegen_mp_num","defined","in","https",":","\/\/github.com\/huggingface\/transformers\/blob\/main\/src\/transformers\/models\/codegen\/modeling_codegen.py","assert","(","get_num_kv_heads","(",")","%","(","mp_size","*","codegen_mp_num",")","==","0",")",",","``","codgen","autoTP","requires","num_kv_heads","%","(","mp_size","*","codegen_mp_num",")","==","0","''","#","input",":","[","3","*","hidden_dim",",","hidden_dim","]","(","weight",")","or","[","3","*","hidden_dim","]","(","bias",")","shape","=","input.shape","dst_shape","=","get_shard_size","(","shape","[","0","]",",","mp_size",")","num_mp_blocks","=","input.reshape","(","codegen_mp_num",",","shape","[","0","]","\/\/","codegen_mp_num",",","shape","[","1","]",")","#","num_mp_blocks",":","[","codegen_mp_num",",","3","*","hidden_dim\/codegen_mp_num",",",":","]","src_split","=","list","(","torch.split","(","num_mp_blocks",",","num_mp_blocks.shape","[","1","]","\/\/","3",",","dim=1",")",")","src_split","=","[","x.reshape","(","codegen_mp_num","*","mp_size",",","-1",",","shape","[","1","]",")","for","x","in","src_split","]","split_fusedqkv","=","split_by_qkvlist_and_refuse","(","src_split",",","get_shard_size","(","shape","[","0","]","\/\/","3",",","mp_size",")",",","0",",","1",")","tp_fuseqkv_weight","=","torch.cat","(","split_fusedqkv",",","dim=0",")",".reshape","(","shape","[","0","]",",","-1",")","return","tp_fuseqkv_weight","[","gpu_index","*","dst_shape",":","(","gpu_index","+","1",")","*","dst_shape","]","def","_glm_type_transpose","(","input",",","mp_size",")",":","#","input",":","[","3","*","hidden_dim",",","hidden_dim","]","(","weight",")","or","[","3","*","hidden_dim","]","(","bias",")","shape","=","input.shape","src_split","=","torch.split","(","input",",","shape","[","0","]","\/\/","3",",","dim=0",")","split_fusedqkv","=","split_by_qkvlist_and_refuse","(","src_split",",","get_shard_size_list","(","shape","[","0","]","\/\/","3",",","mp_size",")",")","return","split_fusedqkv","[","gpu_index","]","def","_bloom_type_transpose","(","input",",","mp_size",")",":","shape","=","input.shape","split_fusedqkv","=","input.split","(","get_shard_size_list","(","shape","[","0","]",",","mp_size",")",",","dim=0",")","return","split_fusedqkv","[","gpu_index","]","def","_qwen_type_transpose","(","input",",","mp_size",",","module",")",":","if","not","hasattr","(","module",",","``","_ds_fusedqkv_entered","''",")",":","#","Adjust","splitting","absolute","value","variables","setattr","(","module",",","``","_ds_fusedqkv_entered","''",",","True",")","module.attn.split_size","=","get_shard_size","(","module.attn.split_size",",","mp_size",")","return","_glm_type_transpose","(","input",",","mp_size",")","def","_bigcode_type_transpose","(","input",",","mp_size",")",":","n_embd","=","get_n_embd","(",")","q","=","input","[",":","n_embd","]","kv","=","input","[","n_embd",":","]","shape","=","q.shape","split_q","=","q.split","(","get_shard_size_list","(","shape","[","0","]",",","mp_size",")",",","dim=0",")","return","torch.cat","(","(","split_q","[","gpu_index","]",",","kv",")",",","dim=0",")","def","_transpose_fused_qkvw","(","src",",","mp_size",",","fused_qkv_type=None",",","module=None",")",":","#","suppose","num_heads=n",",","q","(","n",")","_w","means","the","n-th","q","head","linear","weight",",","the","weight","format","are","as","following","#","bloomtype",":","[","q","(","1",")","_w",",","k","(","1",")","_w",",","v","(","1",")","_w",",","q","(","2",")","_w",",","k","(","2",")","_w",",","v","(","2",")","_w",",","...",",","q","(","n",")","_w",",","k","(","n",")","_w",",","v","(","n",")","_w","]","#","glmtype",":","[","q","(","1",")","_w",",","q","(","2",")","_w",",","...",",","q","(","n",")","_w",",","k","(","1",")","_w",",","k","(","2",")","_w",",","...",",","k","(","n",")","_w",",","v","(","1",")","_w",",","v","(","2",")","_w",",","...",",","v","(","n",")","_w","]","#","codegentype",":","[","q","(","1",")","_w",",","q","(","2",")","_w",",","...",",","q","(","n\/t",")","_w",",","k","(","1",")","_w",",","k","(","2",")","_w",",","...",",","k","(","n\/t",")","_w",",","v","(","1",")","_2",",","v","(","2",")","_w",",","...","v","(","n\/t",")","_w",",","q","(","n\/t+1",")","_w",",","...","]",",","where","t","is","a","const","defined","in","model","file",".","if","fused_qkv_type","==","``","bloomtype","''",":","return","_bloom_type_transpose","(","src",",","mp_size",")","elif","fused_qkv_type","==","``","codegentype","''",":","return","_codegen_type_transpose","(","src",",","mp_size",")","elif","fused_qkv_type","==","``","glmtype","''",":","return","_glm_type_transpose","(","src",",","mp_size",")","elif","fused_qkv_type","==","``","qwentype","''",":","return","_qwen_type_transpose","(","src",",","mp_size",",","module",")","elif","fused_qkv_type","==","``","bigcodetype","''",":","return","_bigcode_type_transpose","(","src",",","mp_size",")","raise","ValueError","(","``","unknown","fused_qkv_type","''",")","for","module_name",",","fused_type","in","fused_type_dict.items","(",")",":","if","re.search","(","module_name",",","module_str",")",":","return","_transpose_fused_qkvw","(","src",",","mp_size",",","fused_type",",","module",")","warning_once","(","f","''","Unrecognized","fusedkqv","weight","type",",","default","to","using","bloom","type",",","''","f","''","please","check","in","prepare_tp_fused_qkvw","(",")","to","avoid","potential","calculation","errors","''",")","return","_bloom_type_transpose","(","src",",","mp_size",")"] | 30 | 122 | null | fusedqkv_utils.py | deepspeed/deepspeed/module_inject/fusedqkv_utils.py | import torch
from deepspeed.utils.logging import warning_once
from deepspeed.module_inject.tp_shard import get_shard_size, get_shard_size_list, get_num_kv_heads, get_n_embd
import re | 7 | null | 4 | null | null | 3 | null | Use image node_id 3 for calling a global function with example usage: prepare_tp_fused_qkvw(module, src, mp_size, gpu_index) and returns: _bloom_type_transpose, tp_fuseqkv_weight, split_fusedqkv, split_fusedqkv, _glm_type_transpose, torch, _bloom_type_transpose, _transpose_fused_qkvw, _codegen_type_transpose, _glm_type_transpose, _qwen_type_transpose, _bigcode_type_transpose | 377 | node_id 3 | 102,805 |
test_en_lex_attrs_capitals | global | null | false | word | null | null | null | null | null | def test_en_lex_attrs_capitals(word):
assert like_num(word)
assert like_num(word.upper())
| ["def","test_en_lex_attrs_capitals","(","word",")",":","assert","like_num","(","word",")","assert","like_num","(","word.upper","(",")",")"] | 68 | 70 | null | test_text.py | spacy/spacy/tests/lang/en/test_text.py | import pytest
from spacy.lang.en.lex_attrs import like_num | 7 | null | 2 | null | null | 5 | null | Use image node_id 5 for calling a global function with example usage: test_en_lex_attrs_capitals(word) without return types | 123 | node_id 5 | 388,124 |
test_en_lex_attrs_like_number_for_ordinal | global | null | false | word | null | null | null | null | null | def test_en_lex_attrs_like_number_for_ordinal(word):
assert like_num(word)
| ["def","test_en_lex_attrs_like_number_for_ordinal","(","word",")",":","assert","like_num","(","word",")"] | 63 | 64 | null | test_text.py | spacy/spacy/tests/lang/en/test_text.py | import pytest
from spacy.lang.en.lex_attrs import like_num | 7 | null | 2 | null | null | 5 | null | Use image node_id 4 for calling a global function with example usage: test_en_lex_attrs_like_number_for_ordinal(word) without return types | 138 | node_id 4 | 388,123 |
test_lex_attrs_like_number | global | null | false | en_tokenizer,text,match | null | null | null | null | null | def test_lex_attrs_like_number(en_tokenizer, text, match):
tokens = en_tokenizer(text)
assert len(tokens) == 1
assert tokens[0].like_num == match
| ["def","test_lex_attrs_like_number","(","en_tokenizer",",","text",",","match",")",":","tokens","=","en_tokenizer","(","text",")","assert","len","(","tokens",")","==","1","assert","tokens","[","0","]",".like_num","==","match"] | 54 | 57 | null | test_text.py | spacy/spacy/tests/lang/en/test_text.py | import pytest
from spacy.lang.en.lex_attrs import like_num | 7 | null | 2 | null | null | 5 | null | Use image node_id 3 for calling a global function with example usage: test_lex_attrs_like_number(en_tokenizer, text, match) without return types | 144 | node_id 3 | 388,122 |
test_en_tokenizer_handles_cnts | global | null | false | en_tokenizer,text,length | null | null | null | null | null | def test_en_tokenizer_handles_cnts(en_tokenizer, text, length):
tokens = en_tokenizer(text)
assert len(tokens) == length
| ["def","test_en_tokenizer_handles_cnts","(","en_tokenizer",",","text",",","length",")",":","tokens","=","en_tokenizer","(","text",")","assert","len","(","tokens",")","==","length"] | 33 | 35 | null | test_text.py | spacy/spacy/tests/lang/en/test_text.py | import pytest
from spacy.lang.en.lex_attrs import like_num | 7 | null | 2 | null | null | 5 | null | Use image node_id 2 for calling a global function with example usage: test_en_tokenizer_handles_cnts(en_tokenizer, text, length) without return types | 149 | node_id 2 | 388,121 |
test_en_tokenizer_handles_long_text | global | null | false | en_tokenizer | null | null | null | null | null | def test_en_tokenizer_handles_long_text(en_tokenizer):
text = """Tributes pour in for late British Labour Party leader
Tributes poured in from around the world Thursday
to the late Labour Party leader John Smith, who died earlier from a massive
heart attack aged 55.
In Washington, the US State Department issued a statement regretting "the
untimely death" of the rapier-tongued Scottish barrister and parliamentarian.
"Mr. Smith, throughout his distinguished"""
tokens = en_tokenizer(text)
assert len(tokens) == 76
| ["def","test_en_tokenizer_handles_long_text","(","en_tokenizer",")",":","text","=","``","''","''","Tributes","pour","in","for","late","British","Labour","Party","leader","Tributes","poured","in","from","around","the","world","Thursday","to","the","late","Labour","Party","leader","John","Smith",",","who","died","earlier","from","a","massive","heart","attack","aged","55",".","In","Washington",",","the","US","State","Department","issued","a","statement","regretting","``","the","untimely","death","''","of","the","rapier-tongued","Scottish","barrister","and","parliamentarian",".","``","Mr.","Smith",",","throughout","his","distinguished","''","''","''","tokens","=","en_tokenizer","(","text",")","assert","len","(","tokens",")","==","76"] | 6 | 18 | null | test_text.py | spacy/spacy/tests/lang/en/test_text.py | import pytest
from spacy.lang.en.lex_attrs import like_num | 7 | null | 2 | null | null | 5 | null | Use image node_id 1 for calling a global function with example usage: test_en_tokenizer_handles_long_text(en_tokenizer) without return types | 140 | node_id 1 | 388,120 |
display_voxels | global | null | false | volume | null | null | null | null | null | def display_voxels(volume):
"""
volume: (N,M,P) array
Represents a binary set of pixels: objects are marked with 1,
complementary (porosities) with 0.
The voxels are actually represented with blue transparent surfaces.
Inner porosities are represented in red.
"""
# define colors
red = "#ff0000ff"
blue = "#1f77b410"
# upscale the above voxel image, leaving gaps
filled = explode(np.ones(volume.shape))
fcolors = explode(np.where(volume, blue, red))
# Shrink the gaps
x, y, z = expand_coordinates(
np.indices(np.array(filled.shape) + 1)
)
# Define 3D figure and place voxels
ax = make_ax()
ax.voxels(x, y, z, filled, facecolors=fcolors)
# Compute Euler number in 6 and 26 neighborhood configuration, that
# correspond to 1 and 3 connectivity, respectively
e26 = euler_number(volume, connectivity=3)
e6 = euler_number(volume, connectivity=1)
plt.title(f"Euler number for N26: {e26}, for N6: {e6}")
plt.show()
| ["def","display_voxels","(","volume",")",":","``","''","''","volume",":","(","N",",","M",",","P",")","array","Represents","a","binary","set","of","pixels",":","objects","are","marked","with","1",",","complementary","(","porosities",")","with","0",".","The","voxels","are","actually","represented","with","blue","transparent","surfaces",".","Inner","porosities","are","represented","in","red.","``","''","''","#","define","colors","red","=","``","#","ff0000ff","''","blue","=","``","#","1f77b410","''","#","upscale","the","above","voxel","image",",","leaving","gaps","filled","=","explode","(","np.ones","(","volume.shape",")",")","fcolors","=","explode","(","np.where","(","volume",",","blue",",","red",")",")","#","Shrink","the","gaps","x",",","y",",","z","=","expand_coordinates","(","np.indices","(","np.array","(","filled.shape",")","+","1",")",")","#","Define","3D","figure","and","place","voxels","ax","=","make_ax","(",")","ax.voxels","(","x",",","y",",","z",",","filled",",","facecolors=fcolors",")","#","Compute","Euler","number","in","6","and","26","neighborhood","configuration",",","that","#","correspond","to","1","and","3","connectivity",",","respectively","e26","=","euler_number","(","volume",",","connectivity=3",")","e6","=","euler_number","(","volume",",","connectivity=1",")","plt.title","(","f","''","Euler","number","for","N26",":","{","e26","}",",","for","N6",":","{","e6","}","''",")","plt.show","(",")"] | 109 | 139 | null | plot_euler_number.py | scikit-image/doc/examples/segmentation/plot_euler_number.py | from skimage.measure import euler_number, label
import matplotlib.pyplot
import numpy | 7 | null | 3 | null | null | 4 | null | Use image node_id 4 for calling a global function with example usage: display_voxels(volume) without return types | 113 | node_id 4 | 344,196 |
__init__ | C2SimpleNet | null | true | self,op_name,num_inputs,debug | This module constructs a net with 'op_name' operator. The net consist
a series of such operator.
It initializes the workspace with input blob equal to the number of parameters
needed for the op.
Provides forward method to run the net niter times. | ["This","module","constructs","a","net","with","'op_name","'","operator",".","The","net","consist","a","series","of","such","operator",".","It","initializes","the","workspace","with","input","blob","equal","to","the","number","of","parameters","needed","for","the","op",".","Provides","forward","method","to","run","the","net","niter","times","."] | null | null | C2SimpleNet | def __init__(self, op_name, num_inputs=1, debug=False):
self.input_names = []
self.net = core.Net("framework_benchmark_net")
self.input_names = [f"in_{i}" for i in range(num_inputs)]
for i in range(num_inputs):
add_blob(workspace, self.input_names[i], [1])
self.net.AddExternalInputs(self.input_names)
op_constructor = getattr(self.net, op_name)
op_constructor(self.input_names)
self.output_name = self.net._net.op[-1].output
print(f"Benchmarking op {op_name}:")
for _ in range(NUM_LOOP_ITERS):
output_name = self.net._net.op[-1].output
self.input_names[-1] = output_name[0]
assert len(self.input_names) == num_inputs
op_constructor(self.input_names)
workspace.CreateNet(self.net)
if debug:
print(self.net._net)
| ["def","__init__","(","self",",","op_name",",","num_inputs=1",",","debug=False",")",":","self.input_names","=","[","]","self.net","=","core.Net","(","``","framework_benchmark_net","''",")","self.input_names","=","[","f","''","in_","{","i","}","''","for","i","in","range","(","num_inputs",")","]","for","i","in","range","(","num_inputs",")",":","add_blob","(","workspace",",","self.input_names","[","i","]",",","[","1","]",")","self.net.AddExternalInputs","(","self.input_names",")","op_constructor","=","getattr","(","self.net",",","op_name",")","op_constructor","(","self.input_names",")","self.output_name","=","self.net._net.op","[","-1","]",".output","print","(","f","''","Benchmarking","op","{","op_name","}",":","''",")","for","_","in","range","(","NUM_LOOP_ITERS",")",":","output_name","=","self.net._net.op","[","-1","]",".output","self.input_names","[","-1","]","=","output_name","[","0","]","assert","len","(","self.input_names",")","==","num_inputs","op_constructor","(","self.input_names",")","workspace.CreateNet","(","self.net",")","if","debug",":","print","(","self.net._net",")"] | 23 | 41 | null | C2Module.py | pytorch/benchmarks/framework_overhead_benchmark/C2Module.py | import numpy
from caffe2.python import core, workspace
from utils import NUM_LOOP_ITERS | 7 | 1 | 3 | 0 | null | 1 | 2 | Use image node_id 1 to create a new C2SimpleNet object with example: obj = C2SimpleNet(op_name, num_inputs, debug) | 115 | node_id 1 | 271,494 |
__init__ | Retry | null | true | self,backoff,retries,supported_errors | Retry a specific number of times after a failure | ["Retry","a","specific","number","of","times","after","a","failure"] | Initialize a `Retry` object with a `Backoff` object
that retries a maximum of `retries` times.
`retries` can be negative to retry forever.
You can specify the types of supported errors which trigger
a retry with the `supported_errors` parameter. | ["Initialize","a","`","Retry","`","object","with","a","`","Backoff","`","object","that","retries","a","maximum","of","`","retries","`","times",".","`","retries","`","can","be","negative","to","retry","forever",".","You","can","specify","the","types","of","supported","errors","which","trigger","a","retry","with","the","`","supported_errors","`","parameter","."] | Retry | def __init__(
self,
backoff,
retries,
supported_errors=(ConnectionError, TimeoutError, socket.timeout),
):
"""
Initialize a `Retry` object with a `Backoff` object
that retries a maximum of `retries` times.
`retries` can be negative to retry forever.
You can specify the types of supported errors which trigger
a retry with the `supported_errors` parameter.
"""
self._backoff = backoff
self._retries = retries
self._supported_errors = supported_errors
| ["def","__init__","(","self",",","backoff",",","retries",",","supported_errors=","(","ConnectionError",",","TimeoutError",",","socket.timeout",")",",",")",":","``","''","''","Initialize","a","`","Retry","`","object","with","a","`","Backoff","`","object","that","retries","a","maximum","of","`","retries","`","times",".","`","retries","`","can","be","negative","to","retry","forever",".","You","can","specify","the","types","of","supported","errors","which","trigger","a","retry","with","the","`","supported_errors","`","parameter.","``","''","''","self._backoff","=","backoff","self._retries","=","retries","self._supported_errors","=","supported_errors"] | 10 | 25 | null | retry.py | redis-py/redis/retry.py | import socket
from time import sleep
from redis.exceptions import ConnectionError, TimeoutError | 7 | 1 | 3 | 0 | null | 0 | 3 | Use image node_id 1 to create a new Retry object with example: obj = Retry(backoff, retries, supported_errors) | 111 | node_id 1 | 339,330 |
test_polygon_area | global | null | false | null | null | null | null | null | def test_polygon_area():
x = [0, 0, 1, 1]
y = [0, 1, 1, 0]
assert_almost_equal(polygon_area(y, x), 1)
x = [0, 0, 1]
y = [0, 1, 1]
assert_almost_equal(polygon_area(y, x), 0.5)
x = [0, 0, 0.5, 1, 1, 0.5]
y = [0, 1, 0.5, 1, 0, 0.5]
assert_almost_equal(polygon_area(y, x), 0.5)
| ["def","test_polygon_area","(",")",":","x","=","[","0",",","0",",","1",",","1","]","y","=","[","0",",","1",",","1",",","0","]","assert_almost_equal","(","polygon_area","(","y",",","x",")",",","1",")","x","=","[","0",",","0",",","1","]","y","=","[","0",",","1",",","1","]","assert_almost_equal","(","polygon_area","(","y",",","x",")",",","0.5",")","x","=","[","0",",","0",",","0.5",",","1",",","1",",","0.5","]","y","=","[","0",",","1",",","0.5",",","1",",","0",",","0.5","]","assert_almost_equal","(","polygon_area","(","y",",","x",")",",","0.5",")"] | 38 | 52 | null | test_geometry.py | scikit-image/skimage/_shared/tests/test_geometry.py | import pytest
from skimage._shared._geometry import polygon_clip, polygon_area
import numpy
from numpy.testing import assert_equal, assert_almost_equal | 7 | null | 4 | null | null | 3 | null | Use image node_id 1 for calling a global function with example usage: test_polygon_area() without return types | 110 | node_id 1 | 344,274 |
|
test_return_sequences | Seq2seqTrainerTester | TestCasePlus | true | self | null | null | null | null | model_inputs | def test_return_sequences(self):
# Tests that the number of generated sequences is correct when num_return_sequences > 1
# and essentially ensuring that `accelerator.gather()` is used instead of `gather_for_metrics`
INPUT_COLUMN = "question"
TARGET_COLUMN = "answer"
MAX_INPUT_LENGTH = 256
MAX_TARGET_LENGTH = 256
dataset = datasets.load_dataset(
"gsm8k", "main", split="train[:38]"
)
model = AutoModelForSeq2SeqLM.from_pretrained("t5-small")
tokenizer = T5Tokenizer.from_pretrained("t5-small")
data_collator = DataCollatorForSeq2Seq(
tokenizer, model=model, return_tensors="pt", padding="longest"
)
gen_config = GenerationConfig.from_pretrained(
"t5-small",
max_length=None,
min_length=None,
max_new_tokens=256,
min_new_tokens=1,
num_beams=5,
)
training_args = Seq2SeqTrainingArguments(
".", predict_with_generate=True
)
trainer = Seq2SeqTrainer(
model=model,
args=training_args,
tokenizer=tokenizer,
data_collator=data_collator,
compute_metrics=lambda x: {"samples": x[0].shape[0]},
)
def prepare_data(examples):
# Remove pairs where at least one record is none
inputs = examples[INPUT_COLUMN]
targets = examples[TARGET_COLUMN]
model_inputs = tokenizer(
inputs, max_length=MAX_INPUT_LENGTH, truncation=True
)
labels = tokenizer(
text_target=targets,
max_length=MAX_TARGET_LENGTH,
truncation=True,
)
model_inputs["labels"] = labels["input_ids"]
return model_inputs
prepared_dataset = dataset.map(
prepare_data,
batched=True,
remove_columns=[INPUT_COLUMN, TARGET_COLUMN],
)
dataset_len = len(prepared_dataset) # 38
for num_return_sequences in range(3, 0, -1):
gen_config.num_return_sequences = num_return_sequences
metrics = trainer.evaluate(
eval_dataset=prepared_dataset,
generation_config=gen_config,
)
assert (
metrics["eval_samples"]
== dataset_len * num_return_sequences
), f"Got {metrics['eval_samples']}, expected: {dataset_len * num_return_sequences}"
| ["def","test_return_sequences","(","self",")",":","#","Tests","that","the","number","of","generated","sequences","is","correct","when","num_return_sequences",">","1","#","and","essentially","ensuring","that","`","accelerator.gather","(",")","`","is","used","instead","of","`","gather_for_metrics","`","INPUT_COLUMN","=","``","question","''","TARGET_COLUMN","=","``","answer","''","MAX_INPUT_LENGTH","=","256","MAX_TARGET_LENGTH","=","256","dataset","=","datasets.load_dataset","(","``","gsm8k","''",",","``","main","''",",","split=","''","train","[",":38","]","''",")","model","=","AutoModelForSeq2SeqLM.from_pretrained","(","``","t5-small","''",")","tokenizer","=","T5Tokenizer.from_pretrained","(","``","t5-small","''",")","data_collator","=","DataCollatorForSeq2Seq","(","tokenizer",",","model=model",",","return_tensors=","''","pt","''",",","padding=","''","longest","''",")","gen_config","=","GenerationConfig.from_pretrained","(","``","t5-small","''",",","max_length=None",",","min_length=None",",","max_new_tokens=256",",","min_new_tokens=1",",","num_beams=5",",",")","training_args","=","Seq2SeqTrainingArguments","(","``",".","``",",","predict_with_generate=True",")","trainer","=","Seq2SeqTrainer","(","model=model",",","args=training_args",",","tokenizer=tokenizer",",","data_collator=data_collator",",","compute_metrics=lambda","x",":","{","``","samples","''",":","x","[","0","]",".shape","[","0","]","}",",",")","def","prepare_data","(","examples",")",":","#","Remove","pairs","where","at","least","one","record","is","none","inputs","=","examples","[","INPUT_COLUMN","]","targets","=","examples","[","TARGET_COLUMN","]","model_inputs","=","tokenizer","(","inputs",",","max_length=MAX_INPUT_LENGTH",",","truncation=True",")","labels","=","tokenizer","(","text_target=targets",",","max_length=MAX_TARGET_LENGTH",",","truncation=True",",",")","model_inputs","[","``","labels","''","]","=","labels","[","``","input_ids","''","]","return","model_inputs","prepared_dataset","=","dataset.map","(","prepare_data",",","batched=True",",","remove_columns=","[","INPUT_COLUMN",",","TARGET_COLUMN","]",",",")","dataset_len","=","len","(","prepared_dataset",")","#","38","for","num_return_sequences","in","range","(","3",",","0",",","-1",")",":","gen_config.num_return_sequences","=","num_return_sequences","metrics","=","trainer.evaluate","(","eval_dataset=prepared_dataset",",","generation_config=gen_config",",",")","assert","(","metrics","[","``","eval_samples","''","]","==","dataset_len","*","num_return_sequences",")",",","f","''","Got","{","metrics","[","'eval_samples","'","]","}",",","expected",":","{","dataset_len","*","num_return_sequences","}","''"] | 138 | 183 | null | test_trainer_seq2seq.py | transformers/tests/trainer/test_trainer_seq2seq.py | from transformers import AutoModelForSeq2SeqLM, BertTokenizer, DataCollatorForSeq2Seq, EncoderDecoderModel, GenerationConfig, Seq2SeqTrainer, Seq2SeqTrainingArguments, T5Tokenizer
from transformers.testing_utils import TestCasePlus, require_torch, slow
from transformers.utils import is_datasets_available | 7 | 1 | 3 | 1 | 1 | 0 | 2 | Use image node_id 2 for calling the Seq2seqTrainerTester obj's underlying member method code with example usage: obj.test_return_sequences() and returns: model_inputs | 166 | node_id 2 | 486,561 |
main | global | null | false | null | null | null | null | null | def main():
bench("eager ", add1)
bench("compiled", torch.compile(add1))
| ["def","main","(",")",":","bench","(","``","eager","``",",","add1",")","bench","(","``","compiled","''",",","torch.compile","(","add1",")",")"] | 24 | 26 | null | overheads.py | pytorch/benchmarks/dynamo/microbenchmarks/overheads.py | import time
import timeit
import numpy
import torch | 7 | null | 4 | null | null | 3 | null | Use image node_id 3 for calling a global function with example usage: main() without return types | 97 | node_id 3 | 271,273 |
|
test_hooks_unpickle | global | null | false | null | null | null | null | str | def test_hooks_unpickle():
def inner_func(d1, d2):
return "hello!"
nlp = Language()
doc = nlp("Hello")
doc.user_hooks["similarity"] = inner_func
b = pickle.dumps(doc)
doc2 = pickle.loads(b)
assert doc2.similarity(None) == "hello!"
| ["def","test_hooks_unpickle","(",")",":","def","inner_func","(","d1",",","d2",")",":","return","``","hello","!","''","nlp","=","Language","(",")","doc","=","nlp","(","``","Hello","''",")","doc.user_hooks","[","``","similarity","''","]","=","inner_func","b","=","pickle.dumps","(","doc",")","doc2","=","pickle.loads","(","b",")","assert","doc2.similarity","(","None",")","==","``","hello","!","''"] | 45 | 54 | null | test_pickle_doc.py | spacy/spacy/tests/doc/test_pickle_doc.py | from spacy.compat import pickle
from spacy.language import Language | 7 | null | 2 | null | null | 5 | null | Use image node_id 5 for calling a global function with example usage: test_hooks_unpickle() and returns: str | 108 | node_id 5 | 387,845 |
|
tir_call | global | null | false | ib,extern,name | null | null | null | null | buf,tvm,tvm | def tir_call(ib: tvm.tir.ir_builder, extern: bool, name: str, *args):
"""
ib: ir_builder
extern: bool
True --> tvm.tir.call_extern
False --> tvm.tir.call_packed
name: str
function name
*args:
arguments for function call
"""
def buf_from_array(ib, arr, dtype):
# Allocate enough memory to store the whole array
var = ib.allocate("int32", (len(arr),), scope="global")
for i, v in enumerate(arr):
var[i] = v
# Declare a buffer, which is basically a view on the chunk of memory that we allocated
buf = tvm.tir.decl_buffer(
(len(arr),), dtype, data=var, scope="global"
)
return buf
if extern:
args = [
i.data if isinstance(i, tvm.tir.Buffer) else i
for i in args
]
return tvm.tir.call_extern("int32", name, *args)
else:
args = [
buf_from_array(ib, i, "int32")
if isinstance(i, (tuple, list, tvm.ir.container.Array))
else i
for i in args
]
return tvm.tir.call_packed(name, *args)
| ["def","tir_call","(","ib",":","tvm.tir.ir_builder",",","extern",":","bool",",","name",":","str",",","*","args",")",":","``","''","''","ib",":","ir_builder","extern",":","bool","True","--",">","tvm.tir.call_extern","False","--",">","tvm.tir.call_packed","name",":","str","function","name","*","args",":","arguments","for","function","call","``","''","''","def","buf_from_array","(","ib",",","arr",",","dtype",")",":","#","Allocate","enough","memory","to","store","the","whole","array","var","=","ib.allocate","(","``","int32","''",",","(","len","(","arr",")",",",")",",","scope=","''","global","''",")","for","i",",","v","in","enumerate","(","arr",")",":","var","[","i","]","=","v","#","Declare","a","buffer",",","which","is","basically","a","view","on","the","chunk","of","memory","that","we","allocated","buf","=","tvm.tir.decl_buffer","(","(","len","(","arr",")",",",")",",","dtype",",","data=var",",","scope=","''","global","''",")","return","buf","if","extern",":","args","=","[","i.data","if","isinstance","(","i",",","tvm.tir.Buffer",")","else","i","for","i","in","args","]","return","tvm.tir.call_extern","(","``","int32","''",",","name",",","*","args",")","else",":","args","=","[","buf_from_array","(","ib",",","i",",","``","int32","''",")","if","isinstance","(","i",",","(","tuple",",","list",",","tvm.ir.container.Array",")",")","else","i","for","i","in","args","]","return","tvm.tir.call_packed","(","name",",","*","args",")"] | 105 | 136 | null | passes.py | tvm/apps/uma/_template/passes.py | import tvm
from tvm import tir
from tvm.relay.backend.contrib.uma.api.utils import add_llvm_to_block | 7 | null | 3 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: tir_call(ib, extern, name) and returns: buf, tvm, tvm | 123 | node_id 1 | 487,914 |
_my_ai_hw_conv2d_pass | MyAiHwConv2dPass | null | true | cls,func,mod,ctx | null | null | null | null | r,unknown,op,func,func,irb_result,op | def _my_ai_hw_conv2d_pass(cls, func, mod, ctx):
_loops = dict()
_handles = []
_entry_node = None
def _has_block(name: str, func: tvm.tir.PrimFunc) -> bool:
"""
Determine of a tir.block with `name` exists in `func`
"""
def _hb(op):
if isinstance(op, tvm.tir.Block):
_found_blocks.append(op.name_hint)
_found_blocks = []
tvm.tir.stmt_functor.post_order_visit(func.body, _hb)
return name in _found_blocks
def _detect_and_replace_conv2d(
func: tvm.tir.PrimFunc,
mod: tvm.ir.IRModule,
ctx: tvm.ir.transform.PassContext,
) -> tvm.tir.PrimFunc:
def _replace_conv2d(op):
if op == _entry_node:
irb = tvm.tir.ir_builder.create()
# Collection of buffer address
buffers = [b[1].data for b in _handles]
# extraction of loop offsets
for k, v in _loops.items():
assert v.min.value == 0
offset_order = ["co", "w", "h", "ci", "kh", "kw"]
offsets = [
_loops[i].extent.value for i in offset_order
]
args = buffers + offsets
irb.emit(
tir_call(
irb, True, cls._EXTERNAL_FUNCTION_NAME, *args
)
)
irb_result = irb.get()
return irb_result
elif isinstance(op, tvm.tir.SeqStmt):
# Remove that pad block of TOPI's conv2DNCHW by only returning the 2nd statement
return op.seq[1]
return op
sch = tir.Schedule(func)
if _has_block(cls._TVM_BLOCK_MATCH_NAME, func):
conv2d_block = sch.get_block(cls._TVM_BLOCK_MATCH_NAME)
rv_loops = sch.get_loops(conv2d_block)
assert len(rv_loops) == 7
loops = dict(
n=rv_loops[0],
co=rv_loops[1],
h=rv_loops[2],
w=rv_loops[3],
ci=rv_loops[4],
kh=rv_loops[5],
kw=rv_loops[6],
)
_entry_node = sch.get(rv_loops[1])
_loops = {k: sch.get(v) for k, v in loops.items()}
_handles = func.buffer_map.items()
x = tvm.tir.stmt_functor.ir_transform(
func.body,
None,
_replace_conv2d,
["tir.For", "tir.SeqStmt"],
)
return func.with_body(x)
else:
return func
r = _detect_and_replace_conv2d(func, mod, ctx)
return r
| ["def","_my_ai_hw_conv2d_pass","(","cls",",","func",",","mod",",","ctx",")",":","_loops","=","dict","(",")","_handles","=","[","]","_entry_node","=","None","def","_has_block","(","name",":","str",",","func",":","tvm.tir.PrimFunc",")","-",">","bool",":","``","''","''","Determine","of","a","tir.block","with","`","name","`","exists","in","`","func","`","``","''","''","def","_hb","(","op",")",":","if","isinstance","(","op",",","tvm.tir.Block",")",":","_found_blocks.append","(","op.name_hint",")","_found_blocks","=","[","]","tvm.tir.stmt_functor.post_order_visit","(","func.body",",","_hb",")","return","name","in","_found_blocks","def","_detect_and_replace_conv2d","(","func",":","tvm.tir.PrimFunc",",","mod",":","tvm.ir.IRModule",",","ctx",":","tvm.ir.transform.PassContext",",",")","-",">","tvm.tir.PrimFunc",":","def","_replace_conv2d","(","op",")",":","if","op","==","_entry_node",":","irb","=","tvm.tir.ir_builder.create","(",")","#","Collection","of","buffer","address","buffers","=","[","b","[","1","]",".data","for","b","in","_handles","]","#","extraction","of","loop","offsets","for","k",",","v","in","_loops.items","(",")",":","assert","v.min.value","==","0","offset_order","=","[","``","co","''",",","``","w","''",",","``","h","''",",","``","ci","''",",","``","kh","''",",","``","kw","''","]","offsets","=","[","_loops","[","i","]",".extent.value","for","i","in","offset_order","]","args","=","buffers","+","offsets","irb.emit","(","tir_call","(","irb",",","True",",","cls._EXTERNAL_FUNCTION_NAME",",","*","args",")",")","irb_result","=","irb.get","(",")","return","irb_result","elif","isinstance","(","op",",","tvm.tir.SeqStmt",")",":","#","Remove","that","pad","block","of","TOPI","'s","conv2DNCHW","by","only","returning","the","2nd","statement","return","op.seq","[","1","]","return","op","sch","=","tir.Schedule","(","func",")","if","_has_block","(","cls._TVM_BLOCK_MATCH_NAME",",","func",")",":","conv2d_block","=","sch.get_block","(","cls._TVM_BLOCK_MATCH_NAME",")","rv_loops","=","sch.get_loops","(","conv2d_block",")","assert","len","(","rv_loops",")","==","7","loops","=","dict","(","n=rv_loops","[","0","]",",","co=rv_loops","[","1","]",",","h=rv_loops","[","2","]",",","w=rv_loops","[","3","]",",","ci=rv_loops","[","4","]",",","kh=rv_loops","[","5","]",",","kw=rv_loops","[","6","]",",",")","_entry_node","=","sch.get","(","rv_loops","[","1","]",")","_loops","=","{","k",":","sch.get","(","v",")","for","k",",","v","in","loops.items","(",")","}","_handles","=","func.buffer_map.items","(",")","x","=","tvm.tir.stmt_functor.ir_transform","(","func.body",",","None",",","_replace_conv2d",",","[","``","tir.For","''",",","``","tir.SeqStmt","''","]",",",")","return","func.with_body","(","x",")","else",":","return","func","r","=","_detect_and_replace_conv2d","(","func",",","mod",",","ctx",")","return","r"] | 35 | 102 | null | passes.py | tvm/apps/uma/_template/passes.py | import tvm
from tvm import tir
from tvm.relay.backend.contrib.uma.api.utils import add_llvm_to_block | 7 | 1 | 3 | 0 | null | 1 | 2 | Use image node_id 2 for calling the MyAiHwConv2dPass obj's underlying member method code with example usage: obj._my_ai_hw_conv2d_pass(cls, func, mod, ctx) and returns: r, unknown, op, func, func, irb_result, op | 211 | node_id 2 | 487,913 |
transform_function | MyAiHwConv2dPass | null | true | self,func,mod,ctx | null | null | null | null | self | def transform_function(
self,
func: tvm.tir.PrimFunc,
mod: tvm.ir.IRModule,
ctx: tvm.ir.transform.PassContext,
) -> tvm.tir.PrimFunc:
return self._my_ai_hw_conv2d_pass(func, mod, ctx)
| ["def","transform_function","(","self",",","func",":","tvm.tir.PrimFunc",",","mod",":","tvm.ir.IRModule",",","ctx",":","tvm.ir.transform.PassContext",",",")","-",">","tvm.tir.PrimFunc",":","return","self._my_ai_hw_conv2d_pass","(","func",",","mod",",","ctx",")"] | 29 | 32 | null | passes.py | tvm/apps/uma/_template/passes.py | import tvm
from tvm import tir
from tvm.relay.backend.contrib.uma.api.utils import add_llvm_to_block | 7 | 1 | 3 | 0 | null | 1 | 2 | Use image node_id 1 for calling the MyAiHwConv2dPass obj's underlying member method code with example usage: obj.transform_function(func, mod, ctx) and returns: self | 165 | node_id 1 | 487,912 |
pytest_cmdline_preparse | global | null | false | config,args | null | null | null | null | null | def pytest_cmdline_preparse(config, args):
if "--unskip-missing-tokenizer" not in args:
return
global unskip_missing_tokenizer
unskip_missing_tokenizer = True
| ["def","pytest_cmdline_preparse","(","config",",","args",")",":","if","``","--","unskip-missing-tokenizer","''","not","in","args",":","return","global","unskip_missing_tokenizer","unskip_missing_tokenizer","=","True"] | 46 | 50 | null | conftest.py | llama-recipes/tests/conftest.py | import pytest
from transformers import LlamaTokenizer | 7 | null | 2 | null | null | 5 | null | Use image node_id 5 for calling a global function with example usage: pytest_cmdline_preparse(config, args) without return types | 128 | node_id 5 | 166,997 |
downgrade | global | null | false | null | null | null | null | null | def downgrade():
bind = op.get_bind()
session = Session(bind=bind)
# Add the old permissions on the migration itself
add_pvms(session, get_reversed_new_pvms(PVM_MAP))
migrate_roles(session, get_reversed_pvm_map(PVM_MAP))
try:
session.commit()
except SQLAlchemyError as ex:
print(
f"An error occurred while downgrading permissions: {ex}"
)
session.rollback()
pass
| ["def","downgrade","(",")",":","bind","=","op.get_bind","(",")","session","=","Session","(","bind=bind",")","#","Add","the","old","permissions","on","the","migration","itself","add_pvms","(","session",",","get_reversed_new_pvms","(","PVM_MAP",")",")","migrate_roles","(","session",",","get_reversed_pvm_map","(","PVM_MAP",")",")","try",":","session.commit","(",")","except","SQLAlchemyError","as","ex",":","print","(","f","''","An","error","occurred","while","downgrading","permissions",":","{","ex","}","''",")","session.rollback","(",")","pass"] | 96 | 108 | null | 2020-12-14_10-49_42b4c9e01447_security_converge_databases.py | superset/superset/migrations/versions/2020-12-14_10-49_42b4c9e01447_security_converge_databases.py | import sqlalchemy
from alembic import op
from sqlalchemy.exc import SQLAlchemyError
from sqlalchemy.orm import Session
from superset.migrations.shared.security_converge import add_pvms, get_reversed_new_pvms, get_reversed_pvm_map, migrate_roles, Pvm | 7 | null | 5 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: downgrade() without return types | 102 | node_id 2 | 437,212 |
|
test_dtype | global | null | false | null | null | null | null | null | def test_dtype():
dtype = tvm.DataType("handle")
assert dtype.type_code == tvm.DataTypeCode.HANDLE
| ["def","test_dtype","(",")",":","dtype","=","tvm.DataType","(","``","handle","''",")","assert","dtype.type_code","==","tvm.DataTypeCode.HANDLE"] | 63 | 65 | null | test_runtime_ndarray.py | tvm/tests/python/all-platform-minimal-test/test_runtime_ndarray.py | import tvm
from tvm import te
import numpy
import tvm.testing | 7 | null | 4 | null | null | 3 | null | Use image node_id 3 for calling a global function with example usage: test_dtype() without return types | 103 | node_id 3 | 498,113 |
|
upgrade | global | null | false | null | null | null | null | null | def upgrade():
bind = op.get_bind()
session = Session(bind=bind)
# Add the new permissions on the migration itself
add_pvms(session, NEW_PVMS)
migrate_roles(session, PVM_MAP)
try:
session.commit()
except SQLAlchemyError as ex:
print(f"An error occurred while upgrading permissions: {ex}")
session.rollback()
| ["def","upgrade","(",")",":","bind","=","op.get_bind","(",")","session","=","Session","(","bind=bind",")","#","Add","the","new","permissions","on","the","migration","itself","add_pvms","(","session",",","NEW_PVMS",")","migrate_roles","(","session",",","PVM_MAP",")","try",":","session.commit","(",")","except","SQLAlchemyError","as","ex",":","print","(","f","''","An","error","occurred","while","upgrading","permissions",":","{","ex","}","''",")","session.rollback","(",")"] | 82 | 93 | null | 2020-12-14_10-49_42b4c9e01447_security_converge_databases.py | superset/superset/migrations/versions/2020-12-14_10-49_42b4c9e01447_security_converge_databases.py | import sqlalchemy
from alembic import op
from sqlalchemy.exc import SQLAlchemyError
from sqlalchemy.orm import Session
from superset.migrations.shared.security_converge import add_pvms, get_reversed_new_pvms, get_reversed_pvm_map, migrate_roles, Pvm | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: upgrade() without return types | 100 | node_id 1 | 437,211 |
|
test_pod_log_reader_rolebinding | TestPodReader | null | true | self,triggerer,webserver,expected | Tests RBAC Pod Reader. | ["Tests","RBAC","Pod","Reader","."] | null | null | null | def test_pod_log_reader_rolebinding(
self, triggerer, webserver, expected
):
docs = render_chart(
values={
"triggerer": {"enabled": triggerer},
"webserver": {"allowPodLogReading": webserver},
},
show_only=["templates/rbac/pod-log-reader-rolebinding.yaml"],
)
actual = (
jmespath.search("subjects[*].name", docs[0]) if docs else []
)
assert actual == expected
| ["def","test_pod_log_reader_rolebinding","(","self",",","triggerer",",","webserver",",","expected",")",":","docs","=","render_chart","(","values=","{","``","triggerer","''",":","{","``","enabled","''",":","triggerer","}",",","``","webserver","''",":","{","``","allowPodLogReading","''",":","webserver","}",",","}",",","show_only=","[","``","templates\/rbac\/pod-log-reader-rolebinding.yaml","''","]",",",")","actual","=","(","jmespath.search","(","``","subjects","[","*","]",".name","''",",","docs","[","0","]",")","if","docs","else","[","]",")","assert","actual","==","expected"] | 37 | 46 | null | test_rbac_pod_log_reader.py | airflow/helm_tests/security/test_rbac_pod_log_reader.py | from __future__ import annotations
import jmespath
import pytest
from tests.charts.helm_template_generator import render_chart | 7 | 1 | 4 | 0 | null | 0 | 2 | Use image node_id 1 for calling the TestPodReader obj's underlying member method code with example usage: obj.test_pod_log_reader_rolebinding(triggerer, webserver, expected) without return types | 194 | node_id 1 | 13,850 |
test_pod_log_reader_role | TestPodReader | null | true | self,triggerer,webserver,expected | Tests RBAC Pod Reader. | ["Tests","RBAC","Pod","Reader","."] | null | null | null | def test_pod_log_reader_role(self, triggerer, webserver, expected):
docs = render_chart(
values={
"triggerer": {"enabled": triggerer},
"webserver": {"allowPodLogReading": webserver},
},
show_only=["templates/rbac/pod-log-reader-role.yaml"],
)
actual = (
jmespath.search("metadata.name", docs[0]) if docs else None
)
assert actual == expected
| ["def","test_pod_log_reader_role","(","self",",","triggerer",",","webserver",",","expected",")",":","docs","=","render_chart","(","values=","{","``","triggerer","''",":","{","``","enabled","''",":","triggerer","}",",","``","webserver","''",":","{","``","allowPodLogReading","''",":","webserver","}",",","}",",","show_only=","[","``","templates\/rbac\/pod-log-reader-role.yaml","''","]",",",")","actual","=","(","jmespath.search","(","``","metadata.name","''",",","docs","[","0","]",")","if","docs","else","None",")","assert","actual","==","expected"] | 57 | 66 | null | test_rbac_pod_log_reader.py | airflow/helm_tests/security/test_rbac_pod_log_reader.py | from __future__ import annotations
import jmespath
import pytest
from tests.charts.helm_template_generator import render_chart | 7 | 1 | 4 | 0 | null | 0 | 2 | Use image node_id 2 for calling the TestPodReader obj's underlying member method code with example usage: obj.test_pod_log_reader_role(triggerer, webserver, expected) without return types | 187 | node_id 2 | 13,851 |
test_env_after_learn | global | null | false | algo | null | null | null | null | env | def test_env_after_learn(algo):
def make_env():
# acktr requires too much RAM, fails on travis
env = gym.make(
"CartPole-v1" if algo == "acktr" else "PongNoFrameskip-v4"
)
return env
make_session(make_default=True, graph=tf.Graph())
env = SubprocVecEnv([make_env])
learn = get_learn_function(algo)
# Commenting out the following line resolves the issue, though crash happens at env.reset().
learn(
network="mlp",
env=env,
total_timesteps=0,
load_path=None,
seed=None,
)
env.reset()
env.close()
| ["def","test_env_after_learn","(","algo",")",":","def","make_env","(",")",":","#","acktr","requires","too","much","RAM",",","fails","on","travis","env","=","gym.make","(","``","CartPole-v1","''","if","algo","==","``","acktr","''","else","``","PongNoFrameskip-v4","''",")","return","env","make_session","(","make_default=True",",","graph=tf.Graph","(",")",")","env","=","SubprocVecEnv","(","[","make_env","]",")","learn","=","get_learn_function","(","algo",")","#","Commenting","out","the","following","line","resolves","the","issue",",","though","crash","happens","at","env.reset","(",")",".","learn","(","network=","''","mlp","''",",","env=env",",","total_timesteps=0",",","load_path=None",",","seed=None",",",")","env.reset","(",")","env.close","(",")"] | 12 | 27 | null | test_env_after_learn.py | baselines/baselines/common/tests/test_env_after_learn.py | import pytest
import gym
import tensorflow
from baselines.common.vec_env.subproc_vec_env import SubprocVecEnv
from baselines.run import get_learn_function
from baselines.common.tf_util import make_session | 7 | null | 6 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: test_env_after_learn(algo) and returns: env | 113 | node_id 1 | 35,143 |
find_constraint_name | global | null | false | upgrade | null | null | null | null | generic_find_constraint_name | def find_constraint_name(upgrade=True):
cols = {"column_name"} if upgrade else {"datasource_name"}
return generic_find_constraint_name(
table="columns",
columns=cols,
referenced="datasources",
database=db,
)
| ["def","find_constraint_name","(","upgrade=True",")",":","cols","=","{","``","column_name","''","}","if","upgrade","else","{","``","datasource_name","''","}","return","generic_find_constraint_name","(","table=","''","columns","''",",","columns=cols",",","referenced=","''","datasources","''",",","database=db",",",")"] | 41 | 45 | null | 2016-05-27_15-03_1226819ee0e3_fix_wrong_constraint_on_table_columns.py | superset/superset/migrations/versions/2016-05-27_15-03_1226819ee0e3_fix_wrong_constraint_on_table_columns.py | import logging
from alembic import op
from superset import db
from superset.utils.core import generic_find_constraint_name | 7 | null | 4 | null | null | 3 | null | Use image node_id 1 for calling a global function with example usage: find_constraint_name(upgrade) and returns: generic_find_constraint_name | 141 | node_id 1 | 436,850 |
test_strict_warning_false | global | null | false | setup,strictness | null | null | null | null | null | def test_strict_warning_false(setup, strictness):
# If the user doesn't wish to be strict about warnings
# the following shouldn't raise any error
os.environ["SKIMAGE_TEST_STRICT_WARNINGS"] = strictness
with expected_warnings(["some warnings"]):
pass
| ["def","test_strict_warning_false","(","setup",",","strictness",")",":","#","If","the","user","does","n't","wish","to","be","strict","about","warnings","#","the","following","should","n't","raise","any","error","os.environ","[","``","SKIMAGE_TEST_STRICT_WARNINGS","''","]","=","strictness","with","expected_warnings","(","[","``","some","warnings","''","]",")",":","pass"] | 32 | 37 | null | test_warnings.py | scikit-image/skimage/_shared/tests/test_warnings.py | import os
from skimage._shared._warnings import expected_warnings
import pytest | 7 | null | 3 | null | null | 4 | null | Use image node_id 4 for calling a global function with example usage: test_strict_warning_false(setup, strictness) without return types | 135 | node_id 4 | 344,315 |
test_strict_warning_true | global | null | false | setup,strictness | null | null | null | null | null | def test_strict_warning_true(setup, strictness):
os.environ["SKIMAGE_TEST_STRICT_WARNINGS"] = strictness
with pytest.raises(ValueError):
with expected_warnings(["some warnings"]):
pass
| ["def","test_strict_warning_true","(","setup",",","strictness",")",":","os.environ","[","``","SKIMAGE_TEST_STRICT_WARNINGS","''","]","=","strictness","with","pytest.raises","(","ValueError",")",":","with","expected_warnings","(","[","``","some","warnings","''","]",")",":","pass"] | 24 | 28 | null | test_warnings.py | scikit-image/skimage/_shared/tests/test_warnings.py | import os
from skimage._shared._warnings import expected_warnings
import pytest | 7 | null | 3 | null | null | 4 | null | Use image node_id 3 for calling a global function with example usage: test_strict_warning_true(setup, strictness) without return types | 134 | node_id 3 | 344,314 |
test_strict_warnigns_default | global | null | false | setup | null | null | null | null | null | def test_strict_warnigns_default(setup):
# By default we should fail on missing expected warnings
with pytest.raises(ValueError):
with expected_warnings(["some warnings"]):
pass
| ["def","test_strict_warnigns_default","(","setup",")",":","#","By","default","we","should","fail","on","missing","expected","warnings","with","pytest.raises","(","ValueError",")",":","with","expected_warnings","(","[","``","some","warnings","''","]",")",":","pass"] | 16 | 20 | null | test_warnings.py | scikit-image/skimage/_shared/tests/test_warnings.py | import os
from skimage._shared._warnings import expected_warnings
import pytest | 7 | null | 3 | null | null | 4 | null | Use image node_id 2 for calling a global function with example usage: test_strict_warnigns_default(setup) without return types | 126 | node_id 2 | 344,313 |
test_poly_clip | global | null | false | null | null | null | null | null | def test_poly_clip():
x = [0, 1, 2, 1]
y = [0, -1, 0, 1]
yc, xc = polygon_clip(y, x, 0, 0, 1, 1)
assert_equal(polygon_area(yc, xc), 0.5)
x = [-1, 1.5, 1.5, -1]
y = [0.5, 0.5, 1.5, 1.5]
yc, xc = polygon_clip(y, x, 0, 0, 1, 1)
assert_equal(polygon_area(yc, xc), 0.5)
| ["def","test_poly_clip","(",")",":","x","=","[","0",",","1",",","2",",","1","]","y","=","[","0",",","-1",",","0",",","1","]","yc",",","xc","=","polygon_clip","(","y",",","x",",","0",",","0",",","1",",","1",")","assert_equal","(","polygon_area","(","yc",",","xc",")",",","0.5",")","x","=","[","-1",",","1.5",",","1.5",",","-1","]","y","=","[","0.5",",","0.5",",","1.5",",","1.5","]","yc",",","xc","=","polygon_clip","(","y",",","x",",","0",",","0",",","1",",","1",")","assert_equal","(","polygon_area","(","yc",",","xc",")",",","0.5",")"] | 55 | 65 | null | test_geometry.py | scikit-image/skimage/_shared/tests/test_geometry.py | import pytest
from skimage._shared._geometry import polygon_clip, polygon_area
import numpy
from numpy.testing import assert_equal, assert_almost_equal | 7 | null | 4 | null | null | 3 | null | Use image node_id 2 for calling a global function with example usage: test_poly_clip() without return types | 107 | node_id 2 | 344,275 |
|
test_hand_clip | global | null | false | null | null | null | null | null | def test_hand_clip():
(r0, c0, r1, c1) = (1.0, 1.5, 2.1, 2.5)
clip_r, clip_c = polygon_clip(
hand[:, 1], hand[:, 0], r0, c0, r1, c1
)
assert_equal(clip_r.size, 19)
assert_equal(clip_r[0], clip_r[-1])
assert_equal(clip_c[0], clip_c[-1])
(r0, c0, r1, c1) = (1.0, 1.5, 1.7, 2.5)
clip_r, clip_c = polygon_clip(
hand[:, 1], hand[:, 0], r0, c0, r1, c1
)
assert_equal(clip_r.size, 6)
(r0, c0, r1, c1) = (1.0, 1.5, 1.5, 2.5)
clip_r, clip_c = polygon_clip(
hand[:, 1], hand[:, 0], r0, c0, r1, c1
)
assert_equal(clip_r.size, 5)
| ["def","test_hand_clip","(",")",":","(","r0",",","c0",",","r1",",","c1",")","=","(","1.0",",","1.5",",","2.1",",","2.5",")","clip_r",",","clip_c","=","polygon_clip","(","hand","[",":",",","1","]",",","hand","[",":",",","0","]",",","r0",",","c0",",","r1",",","c1",")","assert_equal","(","clip_r.size",",","19",")","assert_equal","(","clip_r","[","0","]",",","clip_r","[","-1","]",")","assert_equal","(","clip_c","[","0","]",",","clip_c","[","-1","]",")","(","r0",",","c0",",","r1",",","c1",")","=","(","1.0",",","1.5",",","1.7",",","2.5",")","clip_r",",","clip_c","=","polygon_clip","(","hand","[",":",",","1","]",",","hand","[",":",",","0","]",",","r0",",","c0",",","r1",",","c1",")","assert_equal","(","clip_r.size",",","6",")","(","r0",",","c0",",","r1",",","c1",")","=","(","1.0",",","1.5",",","1.5",",","2.5",")","clip_r",",","clip_c","=","polygon_clip","(","hand","[",":",",","1","]",",","hand","[",":",",","0","]",",","r0",",","c0",",","r1",",","c1",")","assert_equal","(","clip_r.size",",","5",")"] | 68 | 81 | null | test_geometry.py | scikit-image/skimage/_shared/tests/test_geometry.py | import pytest
from skimage._shared._geometry import polygon_clip, polygon_area
import numpy
from numpy.testing import assert_equal, assert_almost_equal | 7 | null | 4 | null | null | 3 | null | Use image node_id 3 for calling a global function with example usage: test_hand_clip() without return types | 107 | node_id 3 | 344,276 |
|
test_call_works_when_ngram_size_is_1 | TestNGramRepeatBlockProcessor | null | true | self | null | null | null | null | null | def test_call_works_when_ngram_size_is_1(self) -> None:
seq1 = torch.tensor([1, 3, 0], device=device)
seq2 = torch.tensor([2, 1, 1], device=device)
seqs = torch.stack([seq1, seq2])
probs1 = torch.tensor([0.1, 0.1, 0.1, 0.1], device=device)
probs2 = torch.tensor([0.1, 0.1, 0.1, 0.1], device=device)
probs = torch.stack([probs1, probs2])
processor = NGramRepeatBlockProcessor(ngram_size=1)
processor(seqs, probs)
assert_close(probs[0], [0.0, 0.0, 0.1, 0.0])
assert_close(probs[1], [0.1, 0.0, 0.0, 0.1])
| ["def","test_call_works_when_ngram_size_is_1","(","self",")","-",">","None",":","seq1","=","torch.tensor","(","[","1",",","3",",","0","]",",","device=device",")","seq2","=","torch.tensor","(","[","2",",","1",",","1","]",",","device=device",")","seqs","=","torch.stack","(","[","seq1",",","seq2","]",")","probs1","=","torch.tensor","(","[","0.1",",","0.1",",","0.1",",","0.1","]",",","device=device",")","probs2","=","torch.tensor","(","[","0.1",",","0.1",",","0.1",",","0.1","]",",","device=device",")","probs","=","torch.stack","(","[","probs1",",","probs2","]",")","processor","=","NGramRepeatBlockProcessor","(","ngram_size=1",")","processor","(","seqs",",","probs",")","assert_close","(","probs","[","0","]",",","[","0.0",",","0.0",",","0.1",",","0.0","]",")","assert_close","(","probs","[","1","]",",","[","0.1",",","0.0",",","0.0",",","0.1","]",")"] | 35 | 51 | null | test_step_processor.py | fairseq2/tests/unit/generation/test_step_processor.py | import torch
from fairseq2.generation import NGramRepeatBlockProcessor
from tests.common import assert_close, device | 7 | 1 | 3 | 0 | null | 0 | 3 | Use image node_id 2 for calling the TestNGramRepeatBlockProcessor obj's underlying member method code with example usage: obj.test_call_works_when_ngram_size_is_1() without return types | 185 | node_id 2 | 133,804 |
get_credentials | UsernamePasswordCredentialProvider | CredentialProvider | true | self | Simple implementation of CredentialProvider that just wraps static
username and password. | ["Simple","implementation","of","CredentialProvider","that","just","wraps","static","username","and","password","."] | null | null | self,self, self | def get_credentials(self):
if self.username:
return self.username, self.password
return (self.password,)
| ["def","get_credentials","(","self",")",":","if","self.username",":","return","self.username",",","self.password","return","(","self.password",",",")"] | 23 | 26 | null | credentials.py | redis-py/redis/credentials.py | from typing import Optional, Tuple, Union | 7 | 2 | 1 | 1 | 1 | 0 | 2 | Use image node_id 2 for calling the UsernamePasswordCredentialProvider obj's underlying member method code with example usage: obj.get_credentials() and returns: self, self, self | 179 | node_id 2 | 339,298 |
__init__ | UsernamePasswordCredentialProvider | CredentialProvider | true | self,username,password | Simple implementation of CredentialProvider that just wraps static
username and password. | ["Simple","implementation","of","CredentialProvider","that","just","wraps","static","username","and","password","."] | null | null | UsernamePasswordCredentialProvider | def __init__(
self,
username: Optional[str] = None,
password: Optional[str] = None,
):
self.username = username or ""
self.password = password or ""
| ["def","__init__","(","self",",","username",":","Optional","[","str","]","=","None",",","password",":","Optional","[","str","]","=","None",",",")",":","self.username","=","username","or","``","''","self.password","=","password","or","``","''"] | 19 | 21 | null | credentials.py | redis-py/redis/credentials.py | from typing import Optional, Tuple, Union | 7 | 2 | 1 | 1 | 1 | 0 | 2 | Use image node_id 1 to create a new UsernamePasswordCredentialProvider object from inherited base classes: CredentialProvider with example: obj = UsernamePasswordCredentialProvider(username, password) | 200 | node_id 1 | 339,297 |
get_credentials | CredentialProvider | null | true | self | Credentials Provider. | ["Credentials","Provider","."] | null | null | null | def get_credentials(self) -> Union[Tuple[str], Tuple[str, str]]:
raise NotImplementedError("get_credentials must be implemented")
| ["def","get_credentials","(","self",")","-",">","Union","[","Tuple","[","str","]",",","Tuple","[","str",",","str","]","]",":","raise","NotImplementedError","(","``","get_credentials","must","be","implemented","''",")"] | 9 | 10 | credentials.py | redis-py/redis/credentials.py | from typing import Optional, Tuple, Union | 7 | 2 | 1 | 1 | null | 0 | 1 | Use image node_id 1 for calling the CredentialProvider obj's underlying member method code with example usage: obj.get_credentials() without return types | 153 | node_id 1 | 339,296 |
|
fake_decode_token | global | null | false | token | null | null | null | null | User | def fake_decode_token(token):
return User(
username=token + "fakedecoded",
email="john@example.com",
full_name="John Doe",
)
| ["def","fake_decode_token","(","token",")",":","return","User","(","username=token","+","``","fakedecoded","''",",","email=","''","john","@","example.com","''",",","full_name=","''","John","Doe","''",",",")"] | 19 | 22 | null | tutorial002.py | fastapi/docs_src/security/tutorial002.py | from typing import Union
from fastapi import Depends, FastAPI
from fastapi.security import OAuth2PasswordBearer
from pydantic import BaseModel | 7 | null | 4 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: fake_decode_token(token) and returns: User | 112 | node_id 1 | 135,730 |
downgrade | global | null | false | null | null | null | null | null | def downgrade():
constraint = (
find_constraint_name(False)
or "fk_columns_datasource_name_datasources"
)
with op.batch_alter_table(
"columns", naming_convention=naming_convention
) as batch_op:
batch_op.drop_constraint(constraint, type_="foreignkey")
batch_op.create_foreign_key(
"fk_columns_column_name_datasources",
"datasources",
["column_name"],
["datasource_name"],
)
| ["def","downgrade","(",")",":","constraint","=","(","find_constraint_name","(","False",")","or","``","fk_columns_datasource_name_datasources","''",")","with","op.batch_alter_table","(","``","columns","''",",","naming_convention=naming_convention",")","as","batch_op",":","batch_op.drop_constraint","(","constraint",",","type_=","''","foreignkey","''",")","batch_op.create_foreign_key","(","``","fk_columns_column_name_datasources","''",",","``","datasources","''",",","[","``","column_name","''","]",",","[","``","datasource_name","''","]",",",")"] | 66 | 77 | null | 2016-05-27_15-03_1226819ee0e3_fix_wrong_constraint_on_table_columns.py | superset/superset/migrations/versions/2016-05-27_15-03_1226819ee0e3_fix_wrong_constraint_on_table_columns.py | import logging
from alembic import op
from superset import db
from superset.utils.core import generic_find_constraint_name | 7 | null | 4 | null | null | 3 | null | Use image node_id 3 for calling a global function with example usage: downgrade() without return types | 102 | node_id 3 | 436,852 |
|
llama_tokenizer | global | null | false | null | null | null | null | None,LlamaTokenizer | def llama_tokenizer():
try:
return LlamaTokenizer.from_pretrained(
"meta-llama/Llama-2-7b-hf"
)
except OSError as e:
if unskip_missing_tokenizer:
raise e
return None
| ["def","llama_tokenizer","(",")",":","try",":","return","LlamaTokenizer.from_pretrained","(","``","meta-llama\/Llama-2-7b-hf","''",")","except","OSError","as","e",":","if","unskip_missing_tokenizer",":","raise","e","return","None"] | 13 | 19 | null | conftest.py | llama-recipes/tests/conftest.py | import pytest
from transformers import LlamaTokenizer | 7 | null | 2 | null | null | 5 | null | Use image node_id 1 for calling a global function with example usage: llama_tokenizer() and returns: None, LlamaTokenizer | 121 | node_id 1 | 166,993 |
|
setup_tokenizer | global | null | false | llama_tokenizer | null | null | null | null | _helper | def setup_tokenizer(llama_tokenizer):
def _helper(tokenizer_mock):
# Align with Llama 2 tokenizer
tokenizer_mock.from_pretrained.return_value = llama_tokenizer
return _helper
| ["def","setup_tokenizer","(","llama_tokenizer",")",":","def","_helper","(","tokenizer_mock",")",":","#","Align","with","Llama","2","tokenizer","tokenizer_mock.from_pretrained.return_value","=","llama_tokenizer","return","_helper"] | 23 | 28 | null | conftest.py | llama-recipes/tests/conftest.py | import pytest
from transformers import LlamaTokenizer | 7 | null | 2 | null | null | 5 | null | Use image node_id 2 for calling a global function with example usage: setup_tokenizer(llama_tokenizer) and returns: _helper | 123 | node_id 2 | 166,994 |
skip_if_tokenizer_is_missing | global | null | false | request,llama_tokenizer | null | null | null | null | null | def skip_if_tokenizer_is_missing(request, llama_tokenizer):
if (
request.node.get_closest_marker("skip_missing_tokenizer")
and not unskip_missing_tokenizer
):
if llama_tokenizer is None:
pytest.skip(ACCESS_ERROR_MSG)
| ["def","skip_if_tokenizer_is_missing","(","request",",","llama_tokenizer",")",":","if","(","request.node.get_closest_marker","(","``","skip_missing_tokenizer","''",")","and","not","unskip_missing_tokenizer",")",":","if","llama_tokenizer","is","None",":","pytest.skip","(","ACCESS_ERROR_MSG",")"] | 32 | 35 | null | conftest.py | llama-recipes/tests/conftest.py | import pytest
from transformers import LlamaTokenizer | 7 | null | 2 | null | null | 5 | null | Use image node_id 3 for calling a global function with example usage: skip_if_tokenizer_is_missing(request, llama_tokenizer) without return types | 145 | node_id 3 | 166,995 |
pytest_addoption | global | null | false | parser | null | null | null | null | null | def pytest_addoption(parser):
parser.addoption(
"--unskip-missing-tokenizer",
action="store_true",
default=False,
help="disable skip missing tokenizer",
)
| ["def","pytest_addoption","(","parser",")",":","parser.addoption","(","``","--","unskip-missing-tokenizer","''",",","action=","''","store_true","''",",","default=False",",","help=","''","disable","skip","missing","tokenizer","''",",",")"] | 38 | 42 | null | conftest.py | llama-recipes/tests/conftest.py | import pytest
from transformers import LlamaTokenizer | 7 | null | 2 | null | null | 5 | null | Use image node_id 4 for calling a global function with example usage: pytest_addoption(parser) without return types | 115 | node_id 4 | 166,996 |
upgrade | global | null | false | null | null | null | null | null | def upgrade():
try:
constraint = find_constraint_name()
with op.batch_alter_table(
"columns", naming_convention=naming_convention
) as batch_op:
if constraint:
batch_op.drop_constraint(
constraint, type_="foreignkey"
)
batch_op.create_foreign_key(
"fk_columns_datasource_name_datasources",
"datasources",
["datasource_name"],
["datasource_name"],
)
except:
logging.warning(
"Could not find or drop constraint on `columns`"
)
| ["def","upgrade","(",")",":","try",":","constraint","=","find_constraint_name","(",")","with","op.batch_alter_table","(","``","columns","''",",","naming_convention=naming_convention",")","as","batch_op",":","if","constraint",":","batch_op.drop_constraint","(","constraint",",","type_=","''","foreignkey","''",")","batch_op.create_foreign_key","(","``","fk_columns_datasource_name_datasources","''",",","``","datasources","''",",","[","``","datasource_name","''","]",",","[","``","datasource_name","''","]",",",")","except",":","logging.warning","(","``","Could","not","find","or","drop","constraint","on","`","columns","`","``",")"] | 48 | 63 | null | 2016-05-27_15-03_1226819ee0e3_fix_wrong_constraint_on_table_columns.py | superset/superset/migrations/versions/2016-05-27_15-03_1226819ee0e3_fix_wrong_constraint_on_table_columns.py | import logging
from alembic import op
from superset import db
from superset.utils.core import generic_find_constraint_name | 7 | null | 4 | null | null | 3 | null | Use image node_id 2 for calling a global function with example usage: upgrade() without return types | 100 | node_id 2 | 436,851 |
|
setup | global | null | false | null | null | null | null | null | def setup():
# Remove any environment variable if it exists
old_strictness = os.environ.pop(
"SKIMAGE_TEST_STRICT_WARNINGS", None
)
yield
# Add the user's desired strictness
if old_strictness is not None:
os.environ["SKIMAGE_TEST_STRICT_WARNINGS"] = old_strictness
| ["def","setup","(",")",":","#","Remove","any","environment","variable","if","it","exists","old_strictness","=","os.environ.pop","(","``","SKIMAGE_TEST_STRICT_WARNINGS","''",",","None",")","yield","#","Add","the","user","'s","desired","strictness","if","old_strictness","is","not","None",":","os.environ","[","``","SKIMAGE_TEST_STRICT_WARNINGS","''","]","=","old_strictness"] | 7 | 13 | null | test_warnings.py | scikit-image/skimage/_shared/tests/test_warnings.py | import os
from skimage._shared._warnings import expected_warnings
import pytest | 7 | null | 3 | null | null | 4 | null | Use image node_id 1 for calling a global function with example usage: setup() without return types | 98 | node_id 1 | 344,312 |
|
test_benchmark_infra_runs | TestDynamoBenchmark | unittest | true | self | null | null | Basic smoke test that TorchBench runs.
This test is mainly meant to check that our setup in fbcode
doesn't break.
If you see a failure here related to missing CPP headers, then
you likely need to update the resources list in:
//caffe2:inductor | ["Basic","smoke","test","that","TorchBench","runs",".","This","test","is","mainly","meant","to","check","that","our","setup","in","fbcode","does","n't","break",".","If","you","see","a","failure","here","related","to","missing","CPP","headers",",","then","you","likely","need","to","update","the","resources","list","in",":","\/\/caffe2",":","inductor"] | null | def test_benchmark_infra_runs(self) -> None:
"""
Basic smoke test that TorchBench runs.
This test is mainly meant to check that our setup in fbcode
doesn't break.
If you see a failure here related to missing CPP headers, then
you likely need to update the resources list in:
//caffe2:inductor
"""
original_dir = setup_torchbench_cwd()
try:
args = parse_args(
[
"-dcpu",
"--inductor",
"--performance",
"--only=BERT_pytorch",
"-n1",
"--batch-size=1",
]
)
run(TorchBenchmarkRunner(), args, original_dir)
finally:
os.chdir(original_dir)
| ["def","test_benchmark_infra_runs","(","self",")","-",">","None",":","``","''","''","Basic","smoke","test","that","TorchBench","runs",".","This","test","is","mainly","meant","to","check","that","our","setup","in","fbcode","does","n't","break",".","If","you","see","a","failure","here","related","to","missing","CPP","headers",",","then","you","likely","need","to","update","the","resources","list","in",":","\/\/caffe2",":","inductor","``","''","''","original_dir","=","setup_torchbench_cwd","(",")","try",":","args","=","parse_args","(","[","``","-dcpu","''",",","``","--","inductor","''",",","``","--","performance","''",",","``","--","only=BERT_pytorch","''",",","``","-n1","''",",","``","--","batch-size=1","''",",","]",")","run","(","TorchBenchmarkRunner","(",")",",","args",",","original_dir",")","finally",":","os.chdir","(","original_dir",")"] | 19 | 44 | null | test.py | pytorch/benchmarks/dynamo/test.py | import os
import unittest
from .common import parse_args, run
from .torchbench import setup_torchbench_cwd, TorchBenchmarkRunner | 7 | 1 | 4 | 1 | 1 | 0 | 1 | Use image node_id 1 for calling the TestDynamoBenchmark obj's underlying member method code with example usage: obj.test_benchmark_infra_runs() without return types | 164 | node_id 1 | 271,340 |
forward | C2SimpleNet | null | true | self,niters | This module constructs a net with 'op_name' operator. The net consist
a series of such operator.
It initializes the workspace with input blob equal to the number of parameters
needed for the op.
Provides forward method to run the net niter times. | ["This","module","constructs","a","net","with","'op_name","'","operator",".","The","net","consist","a","series","of","such","operator",".","It","initializes","the","workspace","with","input","blob","equal","to","the","number","of","parameters","needed","for","the","op",".","Provides","forward","method","to","run","the","net","niter","times","."] | null | null | null | def forward(self, niters):
workspace.RunNet(self.net, niters, False)
| ["def","forward","(","self",",","niters",")",":","workspace.RunNet","(","self.net",",","niters",",","False",")"] | 43 | 44 | null | C2Module.py | pytorch/benchmarks/framework_overhead_benchmark/C2Module.py | import numpy
from caffe2.python import core, workspace
from utils import NUM_LOOP_ITERS | 7 | 1 | 3 | 0 | null | 1 | 2 | Use image node_id 2 for calling the C2SimpleNet obj's underlying member method code with example usage: obj.forward(niters) without return types | 144 | node_id 2 | 271,495 |
read_items | global | null | false | null | null | null | null | list | def read_items() -> List[Item]:
return [
Item(
name="Portal Gun",
description="Device to travel through the multi-rick-verse",
),
Item(name="Plumbus"),
]
| ["def","read_items","(",")","-",">","List","[","Item","]",":","return","[","Item","(","name=","''","Portal","Gun","''",",","description=","''","Device","to","travel","through","the","multi-rick-verse","''",",",")",",","Item","(","name=","''","Plumbus","''",")",",","]"] | 21 | 28 | null | tutorial001.py | fastapi/docs_src/separate_openapi_schemas/tutorial001.py | from typing import List, Union
from fastapi import FastAPI
from pydantic import BaseModel | 7 | null | 3 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: read_items() and returns: list | 100 | node_id 2 | 135,815 |
|
upgrade | global | null | false | null | null | null | null | null | def upgrade():
"""Apply Add ``RenderedTaskInstanceFields`` table"""
json_type = sa.JSON
conn = op.get_bind()
if conn.dialect.name != "postgresql":
# Mysql 5.7+/MariaDB 10.2.3 has JSON support. Rather than checking for
# versions, check for the function existing.
try:
conn.execute(text("SELECT JSON_VALID(1)")).fetchone()
except (sa.exc.OperationalError, sa.exc.ProgrammingError):
json_type = sa.Text
op.create_table(
TABLE_NAME,
sa.Column("dag_id", StringID(), nullable=False),
sa.Column("task_id", StringID(), nullable=False),
sa.Column(
"execution_date",
sa.TIMESTAMP(timezone=True),
nullable=False,
),
sa.Column("rendered_fields", json_type(), nullable=False),
sa.PrimaryKeyConstraint(
"dag_id", "task_id", "execution_date"
),
)
| ["def","upgrade","(",")",":","``","''","''","Apply","Add","``","RenderedTaskInstanceFields","``","table","''","''","''","json_type","=","sa.JSON","conn","=","op.get_bind","(",")","if","conn.dialect.name","!","=","``","postgresql","''",":","#","Mysql","5.7+\/MariaDB","10.2.3","has","JSON","support",".","Rather","than","checking","for","#","versions",",","check","for","the","function","existing",".","try",":","conn.execute","(","text","(","``","SELECT","JSON_VALID","(","1",")","''",")",")",".fetchone","(",")","except","(","sa.exc.OperationalError",",","sa.exc.ProgrammingError",")",":","json_type","=","sa.Text","op.create_table","(","TABLE_NAME",",","sa.Column","(","``","dag_id","''",",","StringID","(",")",",","nullable=False",")",",","sa.Column","(","``","task_id","''",",","StringID","(",")",",","nullable=False",")",",","sa.Column","(","``","execution_date","''",",","sa.TIMESTAMP","(","timezone=True",")",",","nullable=False",",",")",",","sa.Column","(","``","rendered_fields","''",",","json_type","(",")",",","nullable=False",")",",","sa.PrimaryKeyConstraint","(","``","dag_id","''",",","``","task_id","''",",","``","execution_date","''",")",",",")"] | 43 | 63 | null | 0053_1_10_10_add_rendered_task_instance_fields_table.py | airflow/airflow/migrations/versions/0053_1_10_10_add_rendered_task_instance_fields_table.py | from __future__ import annotations
import sqlalchemy
from alembic import op
from sqlalchemy import text
from airflow.migrations.db_types import StringID | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: upgrade() without return types | 100 | node_id 1 | 2,595 |
|
has_unique_constraint | global | null | false | constraint_name,table_name | null | null | null | null | unknown | def has_unique_constraint(
constraint_name: str, table_name: str
) -> bool:
bind = op.get_bind()
inspector = Inspector.from_engine(bind)
unique_constraints = inspector.get_unique_constraints(table_name)
return constraint_name in {
constraint["name"] for constraint in unique_constraints
}
| ["def","has_unique_constraint","(","constraint_name",":","str",",","table_name",":","str",")","-",">","bool",":","bind","=","op.get_bind","(",")","inspector","=","Inspector.from_engine","(","bind",")","unique_constraints","=","inspector.get_unique_constraints","(","table_name",")","return","constraint_name","in","{","constraint","[","``","name","''","]","for","constraint","in","unique_constraints","}"] | 129 | 133 | null | 2020-11-04_11-06_49b5a32daba5_add_report_schedules.py | superset/superset/migrations/versions/2020-11-04_11-06_49b5a32daba5_add_report_schedules.py | import sqlalchemy
from alembic import op
from sqlalchemy.engine.reflection import Inspector
from sqlalchemy.exc import OperationalError | 7 | null | 4 | null | null | 3 | null | Use image node_id 2 for calling a global function with example usage: has_unique_constraint(constraint_name, table_name) and returns: unknown | 141 | node_id 2 | 437,189 |
upgrade | global | null | false | null | null | null | null | null | def upgrade():
op.create_table(
"report_schedule",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("type", sa.String(length=50), nullable=False),
sa.Column(
"name", sa.String(length=150), nullable=False, unique=True
),
sa.Column("description", sa.Text(), nullable=True),
sa.Column("context_markdown", sa.Text(), nullable=True),
sa.Column(
"active", sa.Boolean(), default=True, nullable=True
),
sa.Column("crontab", sa.String(length=50), nullable=False),
sa.Column("sql", sa.Text(), nullable=True),
sa.Column("chart_id", sa.Integer(), nullable=True),
sa.Column("dashboard_id", sa.Integer(), nullable=True),
sa.Column("database_id", sa.Integer(), nullable=True),
sa.Column("last_eval_dttm", sa.DateTime(), nullable=True),
sa.Column("last_state", sa.String(length=50), nullable=True),
sa.Column("last_value", sa.Float(), nullable=True),
sa.Column("last_value_row_json", sa.Text(), nullable=True),
sa.Column(
"validator_type", sa.String(length=100), nullable=True
),
sa.Column(
"validator_config_json",
sa.Text(),
default="{}",
nullable=True,
),
sa.Column(
"log_retention", sa.Integer(), nullable=True, default=90
),
sa.Column(
"grace_period",
sa.Integer(),
nullable=True,
default=60 * 60 * 4,
),
# Audit Mixin
sa.Column("created_on", sa.DateTime(), nullable=True),
sa.Column("changed_on", sa.DateTime(), nullable=True),
sa.Column("created_by_fk", sa.Integer(), nullable=True),
sa.Column("changed_by_fk", sa.Integer(), nullable=True),
sa.ForeignKeyConstraint(["chart_id"], ["slices.id"]),
sa.ForeignKeyConstraint(["dashboard_id"], ["dashboards.id"]),
sa.ForeignKeyConstraint(["database_id"], ["dbs.id"]),
sa.ForeignKeyConstraint(["changed_by_fk"], ["ab_user.id"]),
sa.ForeignKeyConstraint(["created_by_fk"], ["ab_user.id"]),
sa.PrimaryKeyConstraint("id"),
)
try:
op.create_unique_constraint(
"uq_report_schedule_name", "report_schedule", ["name"]
)
except Exception:
# Expected to fail on SQLite
pass
op.create_index(
op.f("ix_report_schedule_active"),
"report_schedule",
["active"],
unique=False,
)
op.create_table(
"report_execution_log",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("scheduled_dttm", sa.DateTime(), nullable=False),
sa.Column("start_dttm", sa.DateTime(), nullable=True),
sa.Column("end_dttm", sa.DateTime(), nullable=True),
sa.Column("value", sa.Float(), nullable=True),
sa.Column("value_row_json", sa.Text(), nullable=True),
sa.Column("state", sa.String(length=50), nullable=False),
sa.Column("error_message", sa.Text(), nullable=True),
sa.Column("report_schedule_id", sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(
["report_schedule_id"], ["report_schedule.id"]
),
sa.PrimaryKeyConstraint("id"),
)
op.create_table(
"report_recipient",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("type", sa.String(length=50), nullable=False),
sa.Column(
"recipient_config_json",
sa.Text(),
default="{}",
nullable=True,
),
sa.Column("report_schedule_id", sa.Integer(), nullable=False),
# Audit Mixin
sa.Column("created_on", sa.DateTime(), nullable=True),
sa.Column("changed_on", sa.DateTime(), nullable=True),
sa.Column("created_by_fk", sa.Integer(), nullable=True),
sa.Column("changed_by_fk", sa.Integer(), nullable=True),
sa.ForeignKeyConstraint(
["report_schedule_id"], ["report_schedule.id"]
),
sa.ForeignKeyConstraint(["changed_by_fk"], ["ab_user.id"]),
sa.ForeignKeyConstraint(["created_by_fk"], ["ab_user.id"]),
sa.PrimaryKeyConstraint("id"),
)
op.create_table(
"report_schedule_user",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("user_id", sa.Integer(), nullable=False),
sa.Column("report_schedule_id", sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(
["report_schedule_id"],
["report_schedule.id"],
),
sa.ForeignKeyConstraint(
["user_id"],
["ab_user.id"],
),
sa.PrimaryKeyConstraint("id"),
)
| ["def","upgrade","(",")",":","op.create_table","(","``","report_schedule","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","type","''",",","sa.String","(","length=50",")",",","nullable=False",")",",","sa.Column","(","``","name","''",",","sa.String","(","length=150",")",",","nullable=False",",","unique=True",")",",","sa.Column","(","``","description","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","context_markdown","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","active","''",",","sa.Boolean","(",")",",","default=True",",","nullable=True",")",",","sa.Column","(","``","crontab","''",",","sa.String","(","length=50",")",",","nullable=False",")",",","sa.Column","(","``","sql","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","chart_id","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","dashboard_id","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","database_id","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","last_eval_dttm","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","last_state","''",",","sa.String","(","length=50",")",",","nullable=True",")",",","sa.Column","(","``","last_value","''",",","sa.Float","(",")",",","nullable=True",")",",","sa.Column","(","``","last_value_row_json","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","validator_type","''",",","sa.String","(","length=100",")",",","nullable=True",")",",","sa.Column","(","``","validator_config_json","''",",","sa.Text","(",")",",","default=","''","{","}","''",",","nullable=True",",",")",",","sa.Column","(","``","log_retention","''",",","sa.Integer","(",")",",","nullable=True",",","default=90",")",",","sa.Column","(","``","grace_period","''",",","sa.Integer","(",")",",","nullable=True",",","default=60","*","60","*","4",",",")",",","#","Audit","Mixin","sa.Column","(","``","created_on","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","changed_on","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","created_by_fk","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","changed_by_fk","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.ForeignKeyConstraint","(","[","``","chart_id","''","]",",","[","``","slices.id","''","]",")",",","sa.ForeignKeyConstraint","(","[","``","dashboard_id","''","]",",","[","``","dashboards.id","''","]",")",",","sa.ForeignKeyConstraint","(","[","``","database_id","''","]",",","[","``","dbs.id","''","]",")",",","sa.ForeignKeyConstraint","(","[","``","changed_by_fk","''","]",",","[","``","ab_user.id","''","]",")",",","sa.ForeignKeyConstraint","(","[","``","created_by_fk","''","]",",","[","``","ab_user.id","''","]",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",",")","try",":","op.create_unique_constraint","(","``","uq_report_schedule_name","''",",","``","report_schedule","''",",","[","``","name","''","]",")","except","Exception",":","#","Expected","to","fail","on","SQLite","pass","op.create_index","(","op.f","(","``","ix_report_schedule_active","''",")",",","``","report_schedule","''",",","[","``","active","''","]",",","unique=False",",",")","op.create_table","(","``","report_execution_log","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","scheduled_dttm","''",",","sa.DateTime","(",")",",","nullable=False",")",",","sa.Column","(","``","start_dttm","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","end_dttm","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","value","''",",","sa.Float","(",")",",","nullable=True",")",",","sa.Column","(","``","value_row_json","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","state","''",",","sa.String","(","length=50",")",",","nullable=False",")",",","sa.Column","(","``","error_message","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","report_schedule_id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.ForeignKeyConstraint","(","[","``","report_schedule_id","''","]",",","[","``","report_schedule.id","''","]",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",",")","op.create_table","(","``","report_recipient","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","type","''",",","sa.String","(","length=50",")",",","nullable=False",")",",","sa.Column","(","``","recipient_config_json","''",",","sa.Text","(",")",",","default=","''","{","}","''",",","nullable=True",",",")",",","sa.Column","(","``","report_schedule_id","''",",","sa.Integer","(",")",",","nullable=False",")",",","#","Audit","Mixin","sa.Column","(","``","created_on","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","changed_on","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.Column","(","``","created_by_fk","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","changed_by_fk","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.ForeignKeyConstraint","(","[","``","report_schedule_id","''","]",",","[","``","report_schedule.id","''","]",")",",","sa.ForeignKeyConstraint","(","[","``","changed_by_fk","''","]",",","[","``","ab_user.id","''","]",")",",","sa.ForeignKeyConstraint","(","[","``","created_by_fk","''","]",",","[","``","ab_user.id","''","]",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",",")","op.create_table","(","``","report_schedule_user","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","user_id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","report_schedule_id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.ForeignKeyConstraint","(","[","``","report_schedule_id","''","]",",","[","``","report_schedule.id","''","]",",",")",",","sa.ForeignKeyConstraint","(","[","``","user_id","''","]",",","[","``","ab_user.id","''","]",",",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",",")"] | 35 | 126 | null | 2020-11-04_11-06_49b5a32daba5_add_report_schedules.py | superset/superset/migrations/versions/2020-11-04_11-06_49b5a32daba5_add_report_schedules.py | import sqlalchemy
from alembic import op
from sqlalchemy.engine.reflection import Inspector
from sqlalchemy.exc import OperationalError | 7 | null | 4 | null | null | 3 | null | Use image node_id 1 for calling a global function with example usage: upgrade() without return types | 100 | node_id 1 | 437,188 |
|
upgrade | global | null | false | null | null | null | null | null | def upgrade():
# We previously had a KnownEvent's table, but we deleted the table without
# a down migration to remove it (so we didn't delete anyone's data if they
# were happening to use the feature.
#
# But before we can delete the users table we need to drop the FK
conn = op.get_bind()
inspector = inspect(conn)
tables = inspector.get_table_names()
if "known_event" in tables:
for fkey in inspector.get_foreign_keys(
table_name="known_event", referred_table="users"
):
if fkey["name"]:
with op.batch_alter_table(
table_name="known_event"
) as bop:
bop.drop_constraint(
fkey["name"], type_="foreignkey"
)
if "chart" in tables:
op.drop_table(
"chart",
)
if "users" in tables:
op.drop_table("users")
| ["def","upgrade","(",")",":","#","We","previously","had","a","KnownEvent","'s","table",",","but","we","deleted","the","table","without","#","a","down","migration","to","remove","it","(","so","we","did","n't","delete","anyone","'s","data","if","they","#","were","happening","to","use","the","feature",".","#","#","But","before","we","can","delete","the","users","table","we","need","to","drop","the","FK","conn","=","op.get_bind","(",")","inspector","=","inspect","(","conn",")","tables","=","inspector.get_table_names","(",")","if","``","known_event","''","in","tables",":","for","fkey","in","inspector.get_foreign_keys","(","table_name=","''","known_event","''",",","referred_table=","''","users","''",")",":","if","fkey","[","``","name","''","]",":","with","op.batch_alter_table","(","table_name=","''","known_event","''",")","as","bop",":","bop.drop_constraint","(","fkey","[","``","name","''","]",",","type_=","''","foreignkey","''",")","if","``","chart","''","in","tables",":","op.drop_table","(","``","chart","''",",",")","if","``","users","''","in","tables",":","op.drop_table","(","``","users","''",")"] | 39 | 62 | null | 0059_2_0_0_drop_user_and_chart.py | airflow/airflow/migrations/versions/0059_2_0_0_drop_user_and_chart.py | from __future__ import annotations
import sqlalchemy
from alembic import op
from sqlalchemy import inspect, text
from sqlalchemy.dialects import mysql | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: upgrade() without return types | 100 | node_id 1 | 2,607 |
|
downgrade | global | null | false | null | null | null | null | null | def downgrade():
conn = op.get_bind()
op.create_table(
"users",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("username", sa.String(length=250), nullable=True),
sa.Column("email", sa.String(length=500), nullable=True),
sa.Column("password", sa.String(255)),
sa.Column("superuser", sa.Boolean(), default=False),
sa.PrimaryKeyConstraint("id"),
sa.UniqueConstraint("username"),
)
op.create_table(
"chart",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("label", sa.String(length=200), nullable=True),
sa.Column("conn_id", sa.String(length=250), nullable=False),
sa.Column("user_id", sa.Integer(), nullable=True),
sa.Column("chart_type", sa.String(length=100), nullable=True),
sa.Column("sql_layout", sa.String(length=50), nullable=True),
sa.Column("sql", sa.Text(), nullable=True),
sa.Column("y_log_scale", sa.Boolean(), nullable=True),
sa.Column("show_datatable", sa.Boolean(), nullable=True),
sa.Column("show_sql", sa.Boolean(), nullable=True),
sa.Column("height", sa.Integer(), nullable=True),
sa.Column(
"default_params", sa.String(length=5000), nullable=True
),
sa.Column("x_is_date", sa.Boolean(), nullable=True),
sa.Column("iteration_no", sa.Integer(), nullable=True),
sa.Column("last_modified", sa.DateTime(), nullable=True),
sa.ForeignKeyConstraint(
["user_id"],
["users.id"],
),
sa.PrimaryKeyConstraint("id"),
)
if conn.dialect.name == "mysql":
conn.execute(text("SET time_zone = '+00:00'"))
op.alter_column(
table_name="chart",
column_name="last_modified",
type_=mysql.TIMESTAMP(fsp=6),
)
else:
if conn.dialect.name in ("sqlite", "mssql"):
return
if conn.dialect.name == "postgresql":
conn.execute(text("set timezone=UTC"))
op.alter_column(
table_name="chart",
column_name="last_modified",
type_=sa.TIMESTAMP(timezone=True),
)
| ["def","downgrade","(",")",":","conn","=","op.get_bind","(",")","op.create_table","(","``","users","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","username","''",",","sa.String","(","length=250",")",",","nullable=True",")",",","sa.Column","(","``","email","''",",","sa.String","(","length=500",")",",","nullable=True",")",",","sa.Column","(","``","password","''",",","sa.String","(","255",")",")",",","sa.Column","(","``","superuser","''",",","sa.Boolean","(",")",",","default=False",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",","sa.UniqueConstraint","(","``","username","''",")",",",")","op.create_table","(","``","chart","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","label","''",",","sa.String","(","length=200",")",",","nullable=True",")",",","sa.Column","(","``","conn_id","''",",","sa.String","(","length=250",")",",","nullable=False",")",",","sa.Column","(","``","user_id","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","chart_type","''",",","sa.String","(","length=100",")",",","nullable=True",")",",","sa.Column","(","``","sql_layout","''",",","sa.String","(","length=50",")",",","nullable=True",")",",","sa.Column","(","``","sql","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","y_log_scale","''",",","sa.Boolean","(",")",",","nullable=True",")",",","sa.Column","(","``","show_datatable","''",",","sa.Boolean","(",")",",","nullable=True",")",",","sa.Column","(","``","show_sql","''",",","sa.Boolean","(",")",",","nullable=True",")",",","sa.Column","(","``","height","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","default_params","''",",","sa.String","(","length=5000",")",",","nullable=True",")",",","sa.Column","(","``","x_is_date","''",",","sa.Boolean","(",")",",","nullable=True",")",",","sa.Column","(","``","iteration_no","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","last_modified","''",",","sa.DateTime","(",")",",","nullable=True",")",",","sa.ForeignKeyConstraint","(","[","``","user_id","''","]",",","[","``","users.id","''","]",",",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",",")","if","conn.dialect.name","==","``","mysql","''",":","conn.execute","(","text","(","``","SET","time_zone","=","'+00:00","'","''",")",")","op.alter_column","(","table_name=","''","chart","''",",","column_name=","''","last_modified","''",",","type_=mysql.TIMESTAMP","(","fsp=6",")",",",")","else",":","if","conn.dialect.name","in","(","``","sqlite","''",",","``","mssql","''",")",":","return","if","conn.dialect.name","==","``","postgresql","''",":","conn.execute","(","text","(","``","set","timezone=UTC","''",")",")","op.alter_column","(","table_name=","''","chart","''",",","column_name=","''","last_modified","''",",","type_=sa.TIMESTAMP","(","timezone=True",")",",",")"] | 65 | 113 | null | 0059_2_0_0_drop_user_and_chart.py | airflow/airflow/migrations/versions/0059_2_0_0_drop_user_and_chart.py | from __future__ import annotations
import sqlalchemy
from alembic import op
from sqlalchemy import inspect, text
from sqlalchemy.dialects import mysql | 7 | null | 5 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: downgrade() without return types | 102 | node_id 2 | 2,608 |
|
evaluation | global | null | false | X_selected,n_clusters,y | null | null | null | null | nmi, acc | def evaluation(X_selected, n_clusters, y):
"""
This function calculates ARI, ACC and NMI of clustering results
Input
-----
X_selected: {numpy array}, shape (n_samples, n_selected_features}
input data on the selected features
n_clusters: {int}
number of clusters
y: {numpy array}, shape (n_samples,)
true labels
Output
------
nmi: {float}
Normalized Mutual Information
acc: {float}
Accuracy
"""
k_means = KMeans(
n_clusters=n_clusters,
init="k-means++",
n_init=10,
max_iter=300,
tol=0.0001,
precompute_distances=True,
verbose=0,
random_state=None,
copy_x=True,
n_jobs=1,
)
k_means.fit(X_selected)
y_predict = k_means.labels_
# calculate NMI
nmi = normalized_mutual_info_score(y, y_predict)
# calculate ACC
y_permuted_predict = best_map(y, y_predict)
acc = accuracy_score(y, y_permuted_predict)
return nmi, acc
| ["def","evaluation","(","X_selected",",","n_clusters",",","y",")",":","``","''","''","This","function","calculates","ARI",",","ACC","and","NMI","of","clustering","results","Input","--","--","-","X_selected",":","{","numpy","array","}",",","shape","(","n_samples",",","n_selected_features","}","input","data","on","the","selected","features","n_clusters",":","{","int","}","number","of","clusters","y",":","{","numpy","array","}",",","shape","(","n_samples",",",")","true","labels","Output","--","--","--","nmi",":","{","float","}","Normalized","Mutual","Information","acc",":","{","float","}","Accuracy","``","''","''","k_means","=","KMeans","(","n_clusters=n_clusters",",","init=","''","k-means++","''",",","n_init=10",",","max_iter=300",",","tol=0.0001",",","precompute_distances=True",",","verbose=0",",","random_state=None",",","copy_x=True",",","n_jobs=1",",",")","k_means.fit","(","X_selected",")","y_predict","=","k_means.labels_","#","calculate","NMI","nmi","=","normalized_mutual_info_score","(","y",",","y_predict",")","#","calculate","ACC","y_permuted_predict","=","best_map","(","y",",","y_predict",")","acc","=","accuracy_score","(","y",",","y_permuted_predict",")","return","nmi",",","acc"] | 39 | 73 | null | unsupervised_evaluation.py | scikit-feature/skfeature/utility/unsupervised_evaluation.py | import numpy
import sklearn.utils.linear_assignment_
from sklearn.metrics import accuracy_score
from sklearn.metrics.cluster import normalized_mutual_info_score
from sklearn.cluster import KMeans | 7 | null | 5 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: evaluation(X_selected, n_clusters, y) and returns: nmi, acc | 130 | node_id 2 | 343,991 |
best_map | global | null | false | l1,l2 | null | null | null | null | new_l2 | def best_map(l1, l2):
"""
Permute labels of l2 to match l1 as much as possible
"""
if len(l1) != len(l2):
print("L1.shape must == L2.shape")
exit(0)
label1 = np.unique(l1)
n_class1 = len(label1)
label2 = np.unique(l2)
n_class2 = len(label2)
n_class = max(n_class1, n_class2)
G = np.zeros((n_class, n_class))
for i in range(0, n_class1):
for j in range(0, n_class2):
ss = l1 == label1[i]
tt = l2 == label2[j]
G[i, j] = np.count_nonzero(ss & tt)
A = la.linear_assignment(-G)
new_l2 = np.zeros(l2.shape)
for i in range(0, n_class2):
new_l2[l2 == label2[A[i][1]]] = label1[A[i][0]]
return new_l2.astype(int)
| ["def","best_map","(","l1",",","l2",")",":","``","''","''","Permute","labels","of","l2","to","match","l1","as","much","as","possible","``","''","''","if","len","(","l1",")","!","=","len","(","l2",")",":","print","(","``","L1.shape","must","==","L2.shape","''",")","exit","(","0",")","label1","=","np.unique","(","l1",")","n_class1","=","len","(","label1",")","label2","=","np.unique","(","l2",")","n_class2","=","len","(","label2",")","n_class","=","max","(","n_class1",",","n_class2",")","G","=","np.zeros","(","(","n_class",",","n_class",")",")","for","i","in","range","(","0",",","n_class1",")",":","for","j","in","range","(","0",",","n_class2",")",":","ss","=","l1","==","label1","[","i","]","tt","=","l2","==","label2","[","j","]","G","[","i",",","j","]","=","np.count_nonzero","(","ss","&","tt",")","A","=","la.linear_assignment","(","-G",")","new_l2","=","np.zeros","(","l2.shape",")","for","i","in","range","(","0",",","n_class2",")",":","new_l2","[","l2","==","label2","[","A","[","i","]","[","1","]","]","]","=","label1","[","A","[","i","]","[","0","]","]","return","new_l2.astype","(","int",")"] | 8 | 36 | null | unsupervised_evaluation.py | scikit-feature/skfeature/utility/unsupervised_evaluation.py | import numpy
import sklearn.utils.linear_assignment_
from sklearn.metrics import accuracy_score
from sklearn.metrics.cluster import normalized_mutual_info_score
from sklearn.cluster import KMeans | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: best_map(l1, l2) and returns: new_l2 | 106 | node_id 1 | 343,990 |
encode_base64 | global | null | false | msg | null | null | null | null | null | def encode_base64(msg):
"""Encode the message's payload in Base64.
Also, add an appropriate Content-Transfer-Encoding header.
"""
orig = msg.get_payload(decode=True)
encdata = str(_bencode(orig), "ascii")
msg.set_payload(encdata)
msg["Content-Transfer-Encoding"] = "base64"
| ["def","encode_base64","(","msg",")",":","``","''","''","Encode","the","message","'s","payload","in","Base64",".","Also",",","add","an","appropriate","Content-Transfer-Encoding","header.","``","''","''","orig","=","msg.get_payload","(","decode=True",")","encdata","=","str","(","_bencode","(","orig",")",",","``","ascii","''",")","msg.set_payload","(","encdata",")","msg","[","``","Content-Transfer-Encoding","''","]","=","``","base64","''"] | 25 | 33 | null | encoders.py | cpython/Lib/email/encoders.py | from base64 import encodebytes
from quopri import encodestring | 7 | null | 2 | null | null | 5 | null | Use image node_id 2 for calling a global function with example usage: encode_base64(msg) without return types | 109 | node_id 2 | 52,445 |
encode_quopri | global | null | false | msg | null | null | null | null | null | def encode_quopri(msg):
"""Encode the message's payload in quoted-printable.
Also, add an appropriate Content-Transfer-Encoding header.
"""
orig = msg.get_payload(decode=True)
encdata = _qencode(orig)
msg.set_payload(encdata)
msg["Content-Transfer-Encoding"] = "quoted-printable"
| ["def","encode_quopri","(","msg",")",":","``","''","''","Encode","the","message","'s","payload","in","quoted-printable",".","Also",",","add","an","appropriate","Content-Transfer-Encoding","header.","``","''","''","orig","=","msg.get_payload","(","decode=True",")","encdata","=","_qencode","(","orig",")","msg.set_payload","(","encdata",")","msg","[","``","Content-Transfer-Encoding","''","]","=","``","quoted-printable","''"] | 36 | 44 | null | encoders.py | cpython/Lib/email/encoders.py | from base64 import encodebytes
from quopri import encodestring | 7 | null | 2 | null | null | 5 | null | Use image node_id 3 for calling a global function with example usage: encode_quopri(msg) without return types | 109 | node_id 3 | 52,446 |
encode_7or8bit | global | null | false | msg | null | null | null | null | null | def encode_7or8bit(msg):
"""Set the Content-Transfer-Encoding header to 7bit or 8bit."""
orig = msg.get_payload(decode=True)
if orig is None:
# There's no payload. For backwards compatibility we use 7bit
msg["Content-Transfer-Encoding"] = "7bit"
return
# We play a trick to make this go fast. If decoding from ASCII succeeds,
# we know the data must be 7bit, otherwise treat it as 8bit.
try:
orig.decode("ascii")
except UnicodeError:
msg["Content-Transfer-Encoding"] = "8bit"
else:
msg["Content-Transfer-Encoding"] = "7bit"
| ["def","encode_7or8bit","(","msg",")",":","``","''","''","Set","the","Content-Transfer-Encoding","header","to","7bit","or","8bit",".","''","''","''","orig","=","msg.get_payload","(","decode=True",")","if","orig","is","None",":","#","There","'s","no","payload",".","For","backwards","compatibility","we","use","7bit","msg","[","``","Content-Transfer-Encoding","''","]","=","``","7bit","''","return","#","We","play","a","trick","to","make","this","go","fast",".","If","decoding","from","ASCII","succeeds",",","#","we","know","the","data","must","be","7bit",",","otherwise","treat","it","as","8bit",".","try",":","orig.decode","(","``","ascii","''",")","except","UnicodeError",":","msg","[","``","Content-Transfer-Encoding","''","]","=","``","8bit","''","else",":","msg","[","``","Content-Transfer-Encoding","''","]","=","``","7bit","''"] | 47 | 61 | null | encoders.py | cpython/Lib/email/encoders.py | from base64 import encodebytes
from quopri import encodestring | 7 | null | 2 | null | null | 5 | null | Use image node_id 4 for calling a global function with example usage: encode_7or8bit(msg) without return types | 110 | node_id 4 | 52,447 |
encode_noop | global | null | false | msg | null | null | null | null | null | def encode_noop(msg):
"""Do nothing."""
| ["def","encode_noop","(","msg",")",":","``","''","''","Do","nothing",".","''","''","''"] | 64 | 65 | null | encoders.py | cpython/Lib/email/encoders.py | from base64 import encodebytes
from quopri import encodestring | 7 | null | 2 | null | null | 5 | null | Use image node_id 5 for calling a global function with example usage: encode_noop(msg) without return types | 107 | node_id 5 | 52,448 |
__init__ | Foo | null | true | self | Docstring for class Foo. Example from Sphinx docs. | ["Docstring","for","class","Foo",".","Example","from","Sphinx","docs","."] | null | null | Foo | def __init__(self):
#: Doc comment for instance attribute qux.
self.qux = 3
self.spam = 4
"""Docstring for instance attribute spam."""
| ["def","__init__","(","self",")",":","#",":","Doc","comment","for","instance","attribute","qux",".","self.qux","=","3","self.spam","=","4","``","''","''","Docstring","for","instance","attribute","spam",".","''","''","''"] | 70 | 75 | null | comments.py | black/tests/data/cases/comments.py | import os.path
import sys
import a
from b.c import X | 7 | 1 | 4 | 0 | null | 1 | 1 | Use image node_id 1 to create a new Foo object with example: obj = Foo() | 73 | node_id 1 | 37,107 |
torch_nn_func_batchnorm | global | null | false | input,running_mean,running_var,weight,bias,training,momentum,eps | null | null | null | null | flops, macs | def torch_nn_func_batchnorm(
input: torch.Tensor,
running_mean: Optional[torch.Tensor],
running_var: Optional[torch.Tensor],
weight: Optional[torch.Tensor] = None,
bias: Optional[torch.Tensor] = None,
training: bool = False,
momentum: float = 0.1,
eps: float = 1e-5,
) -> Tuple[int, int]:
has_affine = weight is not None
if training:
flops = input.numel() * (2 if has_affine else 1)
else:
flops = input.numel() * (5 if has_affine else 4)
macs = 0
return flops, macs
| ["def","torch_nn_func_batchnorm","(","input",":","torch.Tensor",",","running_mean",":","Optional","[","torch.Tensor","]",",","running_var",":","Optional","[","torch.Tensor","]",",","weight",":","Optional","[","torch.Tensor","]","=","None",",","bias",":","Optional","[","torch.Tensor","]","=","None",",","training",":","bool","=","False",",","momentum",":","float","=","0.1",",","eps",":","float","=","1e-5",",",")","-",">","Tuple","[","int",",","int","]",":","has_affine","=","weight","is","not","None","if","training",":","flops","=","input.numel","(",")","*","(","2","if","has_affine","else","1",")","else",":","flops","=","input.numel","(",")","*","(","5","if","has_affine","else","4",")","macs","=","0","return","flops",",","macs"] | 54 | 70 | null | normalization.py | colossalai/colossalai/fx/profiler/experimental/profiler_function/normalization.py | from typing import List, Optional, Tuple
import torch
from ..registry import meta_profiler_function | 7 | null | 3 | null | null | 4 | null | Use image node_id 4 for calling a global function with example usage: torch_nn_func_batchnorm(input, running_mean, running_var, weight, bias, training, momentum, eps) and returns: flops, macs | 192 | node_id 4 | 41,339 |
torch_nn_func_layernorm | global | null | false | input,normalized_shape,weight,bias,eps | null | null | null | null | flops, macs | def torch_nn_func_layernorm(
input: torch.Tensor,
normalized_shape: List[int],
weight: Optional[torch.Tensor] = None,
bias: Optional[torch.Tensor] = None,
eps: float = 1e-5,
) -> Tuple[int, int]:
has_affine = weight is not None
flops = input.numel() * (5 if has_affine else 4)
macs = 0
return flops, macs
| ["def","torch_nn_func_layernorm","(","input",":","torch.Tensor",",","normalized_shape",":","List","[","int","]",",","weight",":","Optional","[","torch.Tensor","]","=","None",",","bias",":","Optional","[","torch.Tensor","]","=","None",",","eps",":","float","=","1e-5",",",")","-",">","Tuple","[","int",",","int","]",":","has_affine","=","weight","is","not","None","flops","=","input.numel","(",")","*","(","5","if","has_affine","else","4",")","macs","=","0","return","flops",",","macs"] | 40 | 50 | null | normalization.py | colossalai/colossalai/fx/profiler/experimental/profiler_function/normalization.py | from typing import List, Optional, Tuple
import torch
from ..registry import meta_profiler_function | 7 | null | 3 | null | null | 4 | null | Use image node_id 3 for calling a global function with example usage: torch_nn_func_layernorm(input, normalized_shape, weight, bias, eps) and returns: flops, macs | 163 | node_id 3 | 41,338 |
test_model_dump_json | global | null | false | null | null | null | null | null | def test_model_dump_json() -> None:
class A(BaseModel):
a: str
b: int = 2
assert (
model_dump_json(A(a="aaa")).replace(" ", "")
== '{"a":"aaa","b":2}'
)
| ["def","test_model_dump_json","(",")","-",">","None",":","class","A","(","BaseModel",")",":","a",":","str","b",":","int","=","2","assert","(","model_dump_json","(","A","(","a=","''","aaa","''",")",")",".replace","(","``","``",",","``","''",")","==","'","{","``","a","''",":","''","aaa","''",",","''","b","''",":2","}","'",")"] | 36 | 41 | null | test_pydantic.py | autogen/test/test_pydantic.py | from typing import Dict, List, Optional, Tuple, Union
from pydantic import BaseModel, Field
from typing_extensions import Annotated
from autogen._pydantic import model_dump, model_dump_json, type2schema | 7 | null | 4 | null | null | 3 | null | Use image node_id 3 for calling a global function with example usage: test_model_dump_json() without return types | 113 | node_id 3 | 33,747 |
|
test_complex_file | global | null | false | test_case | null | null | null | null | null | def test_complex_file(test_case):
"""Run domain-specific tests from test.py file."""
case_module = __import__(
f"test_cases.{test_case}",
globals=globals(),
locals=locals(),
fromlist=["expectation"],
level=1,
)
test_case_dir = os.path.normcase(os.path.dirname(__file__))
test_case_file = os.path.join(
test_case_dir, "test_cases", test_case + ".py"
)
results = list(
check(
[test_case_file],
select=set(ErrorRegistry.get_error_codes()),
ignore_decorators=re.compile("wraps|ignored_decorator"),
property_decorators=DEFAULT_PROPERTY_DECORATORS,
)
)
for error in results:
assert isinstance(error, Error)
results = {(e.definition.name, e.message) for e in results}
assert case_module.expectation.expected == results
| ["def","test_complex_file","(","test_case",")",":","``","''","''","Run","domain-specific","tests","from","test.py","file",".","''","''","''","case_module","=","__import__","(","f","''","test_cases",".","{","test_case","}","''",",","globals=globals","(",")",",","locals=locals","(",")",",","fromlist=","[","``","expectation","''","]",",","level=1",",",")","test_case_dir","=","os.path.normcase","(","os.path.dirname","(","__file__",")",")","test_case_file","=","os.path.join","(","test_case_dir",",","``","test_cases","''",",","test_case","+","``",".py","''",")","results","=","list","(","check","(","[","test_case_file","]",",","select=set","(","ErrorRegistry.get_error_codes","(",")",")",",","ignore_decorators=re.compile","(","``","wraps|ignored_decorator","''",")",",","property_decorators=DEFAULT_PROPERTY_DECORATORS",",",")",")","for","error","in","results",":","assert","isinstance","(","error",",","Error",")","results","=","{","(","e.definition.name",",","e.message",")","for","e","in","results","}","assert","case_module.expectation.expected","==","results"] | 30 | 52 | null | test_definitions.py | pydocstyle/src/tests/test_definitions.py | import os
import re
import pytest
from pydocstyle.violations import Error, ErrorRegistry
from pydocstyle.checker import check
from pydocstyle.config import ConfigurationParser | 7 | null | 6 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: test_complex_file(test_case) without return types | 119 | node_id 1 | 258,436 |
upgrade | global | null | false | null | null | null | null | null | def upgrade():
conn = op.get_bind()
inspector = inspect(conn)
tables = inspector.get_table_names()
if "sensor_instance" in tables:
return
op.create_table(
"sensor_instance",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("task_id", StringID(), nullable=False),
sa.Column("dag_id", StringID(), nullable=False),
sa.Column("execution_date", TIMESTAMP, nullable=False),
sa.Column("state", sa.String(length=20), nullable=True),
sa.Column("try_number", sa.Integer(), nullable=True),
sa.Column("start_date", TIMESTAMP, nullable=True),
sa.Column("operator", sa.String(length=1000), nullable=False),
sa.Column(
"op_classpath", sa.String(length=1000), nullable=False
),
sa.Column("hashcode", sa.BigInteger(), nullable=False),
sa.Column("shardcode", sa.Integer(), nullable=False),
sa.Column("poke_context", sa.Text(), nullable=False),
sa.Column("execution_context", sa.Text(), nullable=True),
sa.Column(
"created_at", TIMESTAMP, default=func.now, nullable=False
),
sa.Column(
"updated_at", TIMESTAMP, default=func.now, nullable=False
),
sa.PrimaryKeyConstraint("id"),
)
op.create_index(
"ti_primary_key",
"sensor_instance",
["dag_id", "task_id", "execution_date"],
unique=True,
)
op.create_index(
"si_hashcode", "sensor_instance", ["hashcode"], unique=False
)
op.create_index(
"si_shardcode", "sensor_instance", ["shardcode"], unique=False
)
op.create_index(
"si_state_shard",
"sensor_instance",
["state", "shardcode"],
unique=False,
)
op.create_index(
"si_updated_at",
"sensor_instance",
["updated_at"],
unique=False,
)
| ["def","upgrade","(",")",":","conn","=","op.get_bind","(",")","inspector","=","inspect","(","conn",")","tables","=","inspector.get_table_names","(",")","if","``","sensor_instance","''","in","tables",":","return","op.create_table","(","``","sensor_instance","''",",","sa.Column","(","``","id","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","task_id","''",",","StringID","(",")",",","nullable=False",")",",","sa.Column","(","``","dag_id","''",",","StringID","(",")",",","nullable=False",")",",","sa.Column","(","``","execution_date","''",",","TIMESTAMP",",","nullable=False",")",",","sa.Column","(","``","state","''",",","sa.String","(","length=20",")",",","nullable=True",")",",","sa.Column","(","``","try_number","''",",","sa.Integer","(",")",",","nullable=True",")",",","sa.Column","(","``","start_date","''",",","TIMESTAMP",",","nullable=True",")",",","sa.Column","(","``","operator","''",",","sa.String","(","length=1000",")",",","nullable=False",")",",","sa.Column","(","``","op_classpath","''",",","sa.String","(","length=1000",")",",","nullable=False",")",",","sa.Column","(","``","hashcode","''",",","sa.BigInteger","(",")",",","nullable=False",")",",","sa.Column","(","``","shardcode","''",",","sa.Integer","(",")",",","nullable=False",")",",","sa.Column","(","``","poke_context","''",",","sa.Text","(",")",",","nullable=False",")",",","sa.Column","(","``","execution_context","''",",","sa.Text","(",")",",","nullable=True",")",",","sa.Column","(","``","created_at","''",",","TIMESTAMP",",","default=func.now",",","nullable=False",")",",","sa.Column","(","``","updated_at","''",",","TIMESTAMP",",","default=func.now",",","nullable=False",")",",","sa.PrimaryKeyConstraint","(","``","id","''",")",",",")","op.create_index","(","``","ti_primary_key","''",",","``","sensor_instance","''",",","[","``","dag_id","''",",","``","task_id","''",",","``","execution_date","''","]",",","unique=True",",",")","op.create_index","(","``","si_hashcode","''",",","``","sensor_instance","''",",","[","``","hashcode","''","]",",","unique=False",")","op.create_index","(","``","si_shardcode","''",",","``","sensor_instance","''",",","[","``","shardcode","''","]",",","unique=False",")","op.create_index","(","``","si_state_shard","''",",","``","sensor_instance","''",",","[","``","state","''",",","``","shardcode","''","]",",","unique=False",",",")","op.create_index","(","``","si_updated_at","''",",","``","sensor_instance","''",",","[","``","updated_at","''","]",",","unique=False",",",")"] | 41 | 71 | null | 0065_2_0_0_update_schema_for_smart_sensor.py | airflow/airflow/migrations/versions/0065_2_0_0_update_schema_for_smart_sensor.py | from __future__ import annotations
import sqlalchemy
from alembic import op
from sqlalchemy import func, inspect
from airflow.migrations.db_types import TIMESTAMP, StringID | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: upgrade() without return types | 100 | node_id 1 | 2,622 |
|
downgrade | global | null | false | null | null | null | null | null | def downgrade():
conn = op.get_bind()
inspector = inspect(conn)
tables = inspector.get_table_names()
if "sensor_instance" in tables:
op.drop_table("sensor_instance")
| ["def","downgrade","(",")",":","conn","=","op.get_bind","(",")","inspector","=","inspect","(","conn",")","tables","=","inspector.get_table_names","(",")","if","``","sensor_instance","''","in","tables",":","op.drop_table","(","``","sensor_instance","''",")"] | 74 | 79 | null | 0065_2_0_0_update_schema_for_smart_sensor.py | airflow/airflow/migrations/versions/0065_2_0_0_update_schema_for_smart_sensor.py | from __future__ import annotations
import sqlalchemy
from alembic import op
from sqlalchemy import func, inspect
from airflow.migrations.db_types import TIMESTAMP, StringID | 7 | null | 5 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: downgrade() without return types | 102 | node_id 2 | 2,623 |
|
__init__ | Dialog | null | true | self,master | null | null | null | null | Dialog | def __init__(self, master=None, **options):
if master is None:
master = options.get("parent")
self.master = master
self.options = options
| ["def","__init__","(","self",",","master=None",",","*","*","options",")",":","if","master","is","None",":","master","=","options.get","(","``","parent","''",")","self.master","=","master","self.options","=","options"] | 20 | 24 | null | commondialog.py | cpython/Lib/tkinter/commondialog.py | from tkinter import _get_temp_root, _destroy_temp_root | 7 | 1 | 1 | 0 | null | 0 | 5 | Use image node_id 1 to create a new Dialog object with example: obj = Dialog(master) | 85 | node_id 1 | 89,691 |
function | global | null | false | default | null | null | null | null | default,x | def function(default=None):
"""Docstring comes first.
Possibly many lines.
"""
# FIXME: Some comment about why this function is crap but still in production.
import inner_imports
if inner_imports.are_evil():
# Explains why we have this if.
# In great detail indeed.
x = X()
return x.method1() # type: ignore
# This return is also commented for some reason.
return default
| ["def","function","(","default=None",")",":","``","''","''","Docstring","comes","first",".","Possibly","many","lines.","``","''","''","#","FIXME",":","Some","comment","about","why","this","function","is","crap","but","still","in","production",".","import","inner_imports","if","inner_imports.are_evil","(",")",":","#","Explains","why","we","have","this","if",".","#","In","great","detail","indeed",".","x","=","X","(",")","return","x.method1","(",")","#","type",":","ignore","#","This","return","is","also","commented","for","some","reason",".","return","default"] | 32 | 47 | null | comments.py | black/tests/data/cases/comments.py | import os.path
import sys
import a
from b.c import X | 7 | null | 4 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: function(default) and returns: default, x | 111 | node_id 1 | 37,108 |
test_compare | global | null | false | null | null | null | null | null | def test_compare():
node_1 = node.Node(node_id=1)
node_2 = node.Node(node_id=2)
edge_1 = edge.Edge(node_1, None, node_2)
assert path.Path.new_empty_path() == path.Path.new_empty_path()
assert path.Path(
nodes=[node_1, node_2], edges=[edge_1]
) == path.Path(nodes=[node_1, node_2], edges=[edge_1])
assert path.Path(nodes=[node_1], edges=[]) != path.Path(
nodes=[], edges=[]
)
assert path.Path(nodes=[node_1], edges=[]) != path.Path(
nodes=[], edges=[]
)
assert path.Path(nodes=[node_1], edges=[]) != path.Path(
nodes=[node_2], edges=[]
)
assert path.Path(nodes=[node_1], edges=[edge_1]) != path.Path(
nodes=[node_1], edges=[]
)
assert path.Path(nodes=[node_1], edges=[edge_1]) != path.Path(
nodes=[node_2], edges=[edge_1]
)
| ["def","test_compare","(",")",":","node_1","=","node.Node","(","node_id=1",")","node_2","=","node.Node","(","node_id=2",")","edge_1","=","edge.Edge","(","node_1",",","None",",","node_2",")","assert","path.Path.new_empty_path","(",")","==","path.Path.new_empty_path","(",")","assert","path.Path","(","nodes=","[","node_1",",","node_2","]",",","edges=","[","edge_1","]",")","==","path.Path","(","nodes=","[","node_1",",","node_2","]",",","edges=","[","edge_1","]",")","assert","path.Path","(","nodes=","[","node_1","]",",","edges=","[","]",")","!","=","path.Path","(","nodes=","[","]",",","edges=","[","]",")","assert","path.Path","(","nodes=","[","node_1","]",",","edges=","[","]",")","!","=","path.Path","(","nodes=","[","]",",","edges=","[","]",")","assert","path.Path","(","nodes=","[","node_1","]",",","edges=","[","]",")","!","=","path.Path","(","nodes=","[","node_2","]",",","edges=","[","]",")","assert","path.Path","(","nodes=","[","node_1","]",",","edges=","[","edge_1","]",")","!","=","path.Path","(","nodes=","[","node_1","]",",","edges=","[","]",")","assert","path.Path","(","nodes=","[","node_1","]",",","edges=","[","edge_1","]",")","!","=","path.Path","(","nodes=","[","node_2","]",",","edges=","[","edge_1","]",")"] | 73 | 90 | null | test_path.py | redis-py/tests/test_graph_utils/test_path.py | import pytest
from redis.commands.graph import edge, node, path | 7 | null | 2 | null | null | 5 | null | Use image node_id 5 for calling a global function with example usage: test_compare() without return types | 105 | node_id 5 | 340,401 |
|
test_nodes_and_edges | global | null | false | null | null | null | null | null | def test_nodes_and_edges():
node_1 = node.Node(node_id=1)
node_2 = node.Node(node_id=2)
edge_1 = edge.Edge(node_1, None, node_2)
p = path.Path.new_empty_path()
assert p.nodes() == []
p.add_node(node_1)
assert [] == p.edges()
assert 0 == p.edge_count()
assert [node_1] == p.nodes()
assert node_1 == p.get_node(0)
assert node_1 == p.first_node()
assert node_1 == p.last_node()
assert 1 == p.nodes_count()
p.add_edge(edge_1)
assert [edge_1] == p.edges()
assert 1 == p.edge_count()
assert edge_1 == p.get_relationship(0)
p.add_node(node_2)
assert [node_1, node_2] == p.nodes()
assert node_1 == p.first_node()
assert node_2 == p.last_node()
assert 2 == p.nodes_count()
| ["def","test_nodes_and_edges","(",")",":","node_1","=","node.Node","(","node_id=1",")","node_2","=","node.Node","(","node_id=2",")","edge_1","=","edge.Edge","(","node_1",",","None",",","node_2",")","p","=","path.Path.new_empty_path","(",")","assert","p.nodes","(",")","==","[","]","p.add_node","(","node_1",")","assert","[","]","==","p.edges","(",")","assert","0","==","p.edge_count","(",")","assert","[","node_1","]","==","p.nodes","(",")","assert","node_1","==","p.get_node","(","0",")","assert","node_1","==","p.first_node","(",")","assert","node_1","==","p.last_node","(",")","assert","1","==","p.nodes_count","(",")","p.add_edge","(","edge_1",")","assert","[","edge_1","]","==","p.edges","(",")","assert","1","==","p.edge_count","(",")","assert","edge_1","==","p.get_relationship","(","0",")","p.add_node","(","node_2",")","assert","[","node_1",",","node_2","]","==","p.nodes","(",")","assert","node_1","==","p.first_node","(",")","assert","node_2","==","p.last_node","(",")","assert","2","==","p.nodes_count","(",")"] | 46 | 69 | null | test_path.py | redis-py/tests/test_graph_utils/test_path.py | import pytest
from redis.commands.graph import edge, node, path | 7 | null | 2 | null | null | 5 | null | Use image node_id 4 for calling a global function with example usage: test_nodes_and_edges() without return types | 113 | node_id 4 | 340,400 |
|
main | global | null | false | null | null | null | null | null | def main() -> None:
"""Compare two llama checkpoint directories"""
one_files = sorted(
glob.glob(os.path.join(sys.argv[1], "consolidated.*.pth"))
)
two_files = sorted(
glob.glob(os.path.join(sys.argv[2], "consolidated.*.pth"))
)
assert len(one_files) == len(
two_files
), "One directory has {} files while another has {} files.".format(
len(one_files), len(two_files)
)
deltas = []
for i in tqdm.trange(len(one_files), desc="Comparing shards"):
one = torch.load(one_files[i])
two = torch.load(two_files[i])
assert len(one) == len(
two
), "shard should have the same length: {} != {}".format(
len(one), len(two)
)
for _, (v, w) in enumerate(zip(one.items(), two.items())):
assert v[0] == w[0], "{} != {}".format(v[0], w[0])
assert (
v[1].shape == w[1].shape
), "tensor {} shape {} != {}".format(
v[0], v[1].shape, w[1].shape
)
delta = (v[1] - w[1]).abs().max().item()
deltas.append((i, v[0], delta))
del one
del two
gc.collect()
deltas = sorted(deltas, key=lambda x: x[-1], reverse=True)
print("Top 10 largest deltas:")
for i, k, v in deltas[:10]:
print(f" shard {i} {k}: {v}")
| ["def","main","(",")","-",">","None",":","``","''","''","Compare","two","llama","checkpoint","directories","''","''","''","one_files","=","sorted","(","glob.glob","(","os.path.join","(","sys.argv","[","1","]",",","``","consolidated",".","*",".pth","''",")",")",")","two_files","=","sorted","(","glob.glob","(","os.path.join","(","sys.argv","[","2","]",",","``","consolidated",".","*",".pth","''",")",")",")","assert","len","(","one_files",")","==","len","(","two_files",")",",","``","One","directory","has","{","}","files","while","another","has","{","}","files",".","``",".format","(","len","(","one_files",")",",","len","(","two_files",")",")","deltas","=","[","]","for","i","in","tqdm.trange","(","len","(","one_files",")",",","desc=","''","Comparing","shards","''",")",":","one","=","torch.load","(","one_files","[","i","]",")","two","=","torch.load","(","two_files","[","i","]",")","assert","len","(","one",")","==","len","(","two",")",",","``","shard","should","have","the","same","length",":","{","}","!","=","{","}","''",".format","(","len","(","one",")",",","len","(","two",")",")","for","_",",","(","v",",","w",")","in","enumerate","(","zip","(","one.items","(",")",",","two.items","(",")",")",")",":","assert","v","[","0","]","==","w","[","0","]",",","``","{","}","!","=","{","}","''",".format","(","v","[","0","]",",","w","[","0","]",")","assert","(","v","[","1","]",".shape","==","w","[","1","]",".shape",")",",","``","tensor","{","}","shape","{","}","!","=","{","}","''",".format","(","v","[","0","]",",","v","[","1","]",".shape",",","w","[","1","]",".shape",")","delta","=","(","v","[","1","]","-","w","[","1","]",")",".abs","(",")",".max","(",")",".item","(",")","deltas.append","(","(","i",",","v","[","0","]",",","delta",")",")","del","one","del","two","gc.collect","(",")","deltas","=","sorted","(","deltas",",","key=lambda","x",":","x","[","-1","]",",","reverse=True",")","print","(","``","Top","10","largest","deltas",":","''",")","for","i",",","k",",","v","in","deltas","[",":10","]",":","print","(","f","''","shard","{","i","}","{","k","}",":","{","v","}","''",")"] | 10 | 44 | null | compare_llama_weights.py | llama-recipes/examples/hf_llama_conversion/compare_llama_weights.py | import gc
import glob
import os
import sys
import torch
import tqdm | 7 | null | 6 | null | null | 1 | null | Use image node_id 1 for calling a global function with example usage: main() without return types | 97 | node_id 1 | 166,889 |
|
test_wrong_flows | global | null | false | null | null | null | null | null | def test_wrong_flows():
node_1 = node.Node(node_id=1)
node_2 = node.Node(node_id=2)
node_3 = node.Node(node_id=3)
edge_1 = edge.Edge(node_1, None, node_2)
edge_2 = edge.Edge(node_1, None, node_3)
p = path.Path.new_empty_path()
with pytest.raises(AssertionError):
p.add_edge(edge_1)
p.add_node(node_1)
with pytest.raises(AssertionError):
p.add_node(node_2)
p.add_edge(edge_1)
with pytest.raises(AssertionError):
p.add_edge(edge_2)
| ["def","test_wrong_flows","(",")",":","node_1","=","node.Node","(","node_id=1",")","node_2","=","node.Node","(","node_id=2",")","node_3","=","node.Node","(","node_id=3",")","edge_1","=","edge.Edge","(","node_1",",","None",",","node_2",")","edge_2","=","edge.Edge","(","node_1",",","None",",","node_3",")","p","=","path.Path.new_empty_path","(",")","with","pytest.raises","(","AssertionError",")",":","p.add_edge","(","edge_1",")","p.add_node","(","node_1",")","with","pytest.raises","(","AssertionError",")",":","p.add_node","(","node_2",")","p.add_edge","(","edge_1",")","with","pytest.raises","(","AssertionError",")",":","p.add_edge","(","edge_2",")"] | 24 | 42 | null | test_path.py | redis-py/tests/test_graph_utils/test_path.py | import pytest
from redis.commands.graph import edge, node, path | 7 | null | 2 | null | null | 5 | null | Use image node_id 3 for calling a global function with example usage: test_wrong_flows() without return types | 109 | node_id 3 | 340,399 |
|
setup | GraphSuite | null | true | self | Benchmark for pixel graph routines in scikit-image. | ["Benchmark","for","pixel","graph","routines","in","scikit-image","."] | null | null | null | def setup(self):
retina = color.rgb2gray(data.retina())
t0, _ = filters.threshold_multiotsu(retina, classes=3)
mask = retina > t0
vessels = filters.sato(retina, sigmas=range(1, 10)) * mask
thresholded = filters.apply_hysteresis_threshold(
vessels, 0.01, 0.03
)
labeled = ndi.label(thresholded)[0]
largest_nonzero_label = np.argmax(
np.bincount(labeled[labeled > 0])
)
binary = labeled == largest_nonzero_label
self.skeleton = morphology.skeletonize(binary)
labeled2 = ndi.label(thresholded[::2, ::2])[0]
largest_nonzero_label2 = np.argmax(
np.bincount(labeled2[labeled2 > 0])
)
binary2 = labeled2 == largest_nonzero_label2
small_skeleton = morphology.skeletonize(binary2)
self.g, self.n = graph.pixel_graph(small_skeleton, connectivity=2)
| ["def","setup","(","self",")",":","retina","=","color.rgb2gray","(","data.retina","(",")",")","t0",",","_","=","filters.threshold_multiotsu","(","retina",",","classes=3",")","mask","=","retina",">","t0","vessels","=","filters.sato","(","retina",",","sigmas=range","(","1",",","10",")",")","*","mask","thresholded","=","filters.apply_hysteresis_threshold","(","vessels",",","0.01",",","0.03",")","labeled","=","ndi.label","(","thresholded",")","[","0","]","largest_nonzero_label","=","np.argmax","(","np.bincount","(","labeled","[","labeled",">","0","]",")",")","binary","=","labeled","==","largest_nonzero_label","self.skeleton","=","morphology.skeletonize","(","binary",")","labeled2","=","ndi.label","(","thresholded","[",":",":2",",",":",":2","]",")","[","0","]","largest_nonzero_label2","=","np.argmax","(","np.bincount","(","labeled2","[","labeled2",">","0","]",")",")","binary2","=","labeled2","==","largest_nonzero_label2","small_skeleton","=","morphology.skeletonize","(","binary2",")","self.g",",","self.n","=","graph.pixel_graph","(","small_skeleton",",","connectivity=2",")"] | 12 | 27 | null | benchmark_graph.py | scikit-image/benchmarks/benchmark_graph.py | import numpy
from scipy import ndimage
from skimage import color, data, filters, graph, morphology | 7 | 1 | 3 | 0 | null | 0 | 3 | Use image node_id 1 for calling the GraphSuite obj's underlying member method code with example usage: obj.setup() without return types | 135 | node_id 1 | 344,034 |
test_new_empty_path | global | null | false | null | null | null | null | null | def test_new_empty_path():
new_empty_path = path.Path.new_empty_path()
assert isinstance(new_empty_path, path.Path)
assert new_empty_path._nodes == []
assert new_empty_path._edges == []
| ["def","test_new_empty_path","(",")",":","new_empty_path","=","path.Path.new_empty_path","(",")","assert","isinstance","(","new_empty_path",",","path.Path",")","assert","new_empty_path._nodes","==","[","]","assert","new_empty_path._edges","==","[","]"] | 16 | 20 | null | test_path.py | redis-py/tests/test_graph_utils/test_path.py | import pytest
from redis.commands.graph import edge, node, path | 7 | null | 2 | null | null | 5 | null | Use image node_id 2 for calling a global function with example usage: test_new_empty_path() without return types | 112 | node_id 2 | 340,398 |
|
dense_rocblas | global | null | false | cfg,data,weight,bias,out_dtype | null | null | null | null | matmul | def dense_rocblas(cfg, data, weight, bias=None, out_dtype=None):
"""Dense operator for rocm backend with cblas.
Parameters
----------
data : tvm.te.Tensor
2-D with shape [batch, in_dim]
weight : tvm.te.Tensor
2-D with shape [out_dim, in_dim]
bias : tvm.te.Tensor, optional
1-D with shape [out_dim]
out_dtype : str
The output type. This is used for mixed precision.
Returns
-------
output : tvm.te.Tensor
2-D with shape [batch, out_dim]
"""
if out_dtype is None:
out_dtype = data.dtype
assert out_dtype == data.dtype, "Mixed precision not supported."
matmul = rocblas.matmul(data, weight, False, True)
batch, in_dim = data.shape
out_dim, _ = weight.shape
cfg.add_flop(batch * in_dim * out_dim * 2)
if bias is not None:
matmul = te.compute(
(batch, out_dim),
lambda i, j: matmul[i, j] + bias[j],
tag=tag.BROADCAST,
)
return matmul
| ["def","dense_rocblas","(","cfg",",","data",",","weight",",","bias=None",",","out_dtype=None",")",":","``","''","''","Dense","operator","for","rocm","backend","with","cblas",".","Parameters","--","--","--","--","--","data",":","tvm.te.Tensor","2-D","with","shape","[","batch",",","in_dim","]","weight",":","tvm.te.Tensor","2-D","with","shape","[","out_dim",",","in_dim","]","bias",":","tvm.te.Tensor",",","optional","1-D","with","shape","[","out_dim","]","out_dtype",":","str","The","output","type",".","This","is","used","for","mixed","precision",".","Returns","--","--","--","-","output",":","tvm.te.Tensor","2-D","with","shape","[","batch",",","out_dim","]","``","''","''","if","out_dtype","is","None",":","out_dtype","=","data.dtype","assert","out_dtype","==","data.dtype",",","``","Mixed","precision","not","supported",".","''","matmul","=","rocblas.matmul","(","data",",","weight",",","False",",","True",")","batch",",","in_dim","=","data.shape","out_dim",",","_","=","weight.shape","cfg.add_flop","(","batch","*","in_dim","*","out_dim","*","2",")","if","bias","is","not","None",":","matmul","=","te.compute","(","(","batch",",","out_dim",")",",","lambda","i",",","j",":","matmul","[","i",",","j","]","+","bias","[","j","]",",","tag=tag.BROADCAST",",",")","return","matmul"] | 27 | 60 | null | dense.py | tvm/python/tvm/topi/rocm/dense.py | from tvm import te
from tvm import autotvm
from tvm.contrib import rocblas
from ..None import generic
from ..None import tag | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: dense_rocblas(cfg, data, weight, bias, out_dtype) and returns: matmul | 139 | node_id 1 | 497,565 |
test_type2schema | global | null | false | null | null | null | null | null | def test_type2schema() -> None:
assert type2schema(str) == {"type": "string"}
assert type2schema(int) == {"type": "integer"}
assert type2schema(float) == {"type": "number"}
assert type2schema(bool) == {"type": "boolean"}
assert type2schema(None) == {"type": "null"}
assert type2schema(Optional[int]) == {
"anyOf": [{"type": "integer"}, {"type": "null"}]
}
assert type2schema(List[int]) == {
"items": {"type": "integer"},
"type": "array",
}
assert type2schema(Tuple[int, float, str]) == {
"maxItems": 3,
"minItems": 3,
"prefixItems": [
{"type": "integer"},
{"type": "number"},
{"type": "string"},
],
"type": "array",
}
assert type2schema(Dict[str, int]) == {
"additionalProperties": {"type": "integer"},
"type": "object",
}
assert type2schema(Annotated[str, "some text"]) == {
"type": "string"
}
assert type2schema(Union[int, float]) == {
"anyOf": [{"type": "integer"}, {"type": "number"}]
}
| ["def","test_type2schema","(",")","-",">","None",":","assert","type2schema","(","str",")","==","{","``","type","''",":","``","string","''","}","assert","type2schema","(","int",")","==","{","``","type","''",":","``","integer","''","}","assert","type2schema","(","float",")","==","{","``","type","''",":","``","number","''","}","assert","type2schema","(","bool",")","==","{","``","type","''",":","``","boolean","''","}","assert","type2schema","(","None",")","==","{","``","type","''",":","``","null","''","}","assert","type2schema","(","Optional","[","int","]",")","==","{","``","anyOf","''",":","[","{","``","type","''",":","``","integer","''","}",",","{","``","type","''",":","``","null","''","}","]","}","assert","type2schema","(","List","[","int","]",")","==","{","``","items","''",":","{","``","type","''",":","``","integer","''","}",",","``","type","''",":","``","array","''",",","}","assert","type2schema","(","Tuple","[","int",",","float",",","str","]",")","==","{","``","maxItems","''",":","3",",","``","minItems","''",":","3",",","``","prefixItems","''",":","[","{","``","type","''",":","``","integer","''","}",",","{","``","type","''",":","``","number","''","}",",","{","``","type","''",":","``","string","''","}",",","]",",","``","type","''",":","``","array","''",",","}","assert","type2schema","(","Dict","[","str",",","int","]",")","==","{","``","additionalProperties","''",":","{","``","type","''",":","``","integer","''","}",",","``","type","''",":","``","object","''",",","}","assert","type2schema","(","Annotated","[","str",",","``","some","text","''","]",")","==","{","``","type","''",":","``","string","''","}","assert","type2schema","(","Union","[","int",",","float","]",")","==","{","``","anyOf","''",":","[","{","``","type","''",":","``","integer","''","}",",","{","``","type","''",":","``","number","''","}","]","}"] | 9 | 25 | null | test_pydantic.py | autogen/test/test_pydantic.py | from typing import Dict, List, Optional, Tuple, Union
from pydantic import BaseModel, Field
from typing_extensions import Annotated
from autogen._pydantic import model_dump, model_dump_json, type2schema | 7 | null | 4 | null | null | 3 | null | Use image node_id 1 for calling a global function with example usage: test_type2schema() without return types | 109 | node_id 1 | 33,745 |
|
test_lineage_parent_id | global | null | false | null | null | null | null | null | def test_lineage_parent_id():
task = mock.MagicMock(
dag_id="dag_id",
execution_date="execution_date",
try_number=1,
task=mock.MagicMock(task_id="task_id"),
)
actual = lineage_parent_id(run_id="run_id", task_instance=task)
job_name = str(
uuid.uuid3(
uuid.NAMESPACE_URL,
f"{_DAG_NAMESPACE}.dag_id.task_id.execution_date.1",
)
)
expected = f"{_DAG_NAMESPACE}/{job_name}/run_id"
assert actual == expected
| ["def","test_lineage_parent_id","(",")",":","task","=","mock.MagicMock","(","dag_id=","''","dag_id","''",",","execution_date=","''","execution_date","''",",","try_number=1",",","task=mock.MagicMock","(","task_id=","''","task_id","''",")",",",")","actual","=","lineage_parent_id","(","run_id=","''","run_id","''",",","task_instance=task",")","job_name","=","str","(","uuid.uuid3","(","uuid.NAMESPACE_URL",",","f","''","{","_DAG_NAMESPACE","}",".dag_id.task_id.execution_date.1","''",",",")",")","expected","=","f","''","{","_DAG_NAMESPACE","}","\/","{","job_name","}","\/run_id","''","assert","actual","==","expected"] | 40 | 52 | null | test_macros.py | airflow/tests/providers/openlineage/plugins/test_macros.py | from __future__ import annotations
import uuid
from unittest import mock
from airflow.providers.openlineage.plugins.adapter import _DAG_NAMESPACE
from airflow.providers.openlineage.plugins.macros import lineage_parent_id, lineage_run_id | 7 | null | 5 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: test_lineage_parent_id() without return types | 115 | node_id 2 | 25,834 |
|
test_model_dump | global | null | false | null | null | null | null | null | def test_model_dump() -> None:
class A(BaseModel):
a: str
b: int = 2
assert model_dump(A(a="aaa")) == {"a": "aaa", "b": 2}
| ["def","test_model_dump","(",")","-",">","None",":","class","A","(","BaseModel",")",":","a",":","str","b",":","int","=","2","assert","model_dump","(","A","(","a=","''","aaa","''",")",")","==","{","``","a","''",":","``","aaa","''",",","``","b","''",":","2","}"] | 28 | 33 | null | test_pydantic.py | autogen/test/test_pydantic.py | from typing import Dict, List, Optional, Tuple, Union
from pydantic import BaseModel, Field
from typing_extensions import Annotated
from autogen._pydantic import model_dump, model_dump_json, type2schema | 7 | null | 4 | null | null | 3 | null | Use image node_id 2 for calling a global function with example usage: test_model_dump() without return types | 108 | node_id 2 | 33,746 |
|
test_lineage_run_id | global | null | false | null | null | null | null | null | def test_lineage_run_id():
task = mock.MagicMock(
dag_id="dag_id",
execution_date="execution_date",
try_number=1,
task=mock.MagicMock(task_id="task_id"),
)
actual = lineage_run_id(task)
expected = str(
uuid.uuid3(
uuid.NAMESPACE_URL,
f"{_DAG_NAMESPACE}.dag_id.task_id.execution_date.1",
)
)
assert actual == expected
| ["def","test_lineage_run_id","(",")",":","task","=","mock.MagicMock","(","dag_id=","''","dag_id","''",",","execution_date=","''","execution_date","''",",","try_number=1",",","task=mock.MagicMock","(","task_id=","''","task_id","''",")",",",")","actual","=","lineage_run_id","(","task",")","expected","=","str","(","uuid.uuid3","(","uuid.NAMESPACE_URL",",","f","''","{","_DAG_NAMESPACE","}",".dag_id.task_id.execution_date.1","''",",",")",")","assert","actual","==","expected"] | 26 | 37 | null | test_macros.py | airflow/tests/providers/openlineage/plugins/test_macros.py | from __future__ import annotations
import uuid
from unittest import mock
from airflow.providers.openlineage.plugins.adapter import _DAG_NAMESPACE
from airflow.providers.openlineage.plugins.macros import lineage_parent_id, lineage_run_id | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: test_lineage_run_id() without return types | 112 | node_id 1 | 25,833 |
|
__init__ | OPTActor | Actor | true | self,pretrained,config,checkpoint,lora_rank,lora_train_bias | OPT Actor model.
Args:
pretrained (str): Pretrained model name or path.
config (OPTConfig): Model config.
checkpoint (bool): Enable gradient checkpointing.
lora_rank (int): Rank of the low-rank approximation.
lora_train_bias (str): LoRA bias training mode. | ["OPT","Actor","model",".","Args",":","pretrained","(","str",")",":","Pretrained","model","name","or","path",".","config","(","OPTConfig",")",":","Model","config",".","checkpoint","(","bool",")",":","Enable","gradient","checkpointing",".","lora_rank","(","int",")",":","Rank","of","the","low-rank","approximation",".","lora_train_bias","(","str",")",":","LoRA","bias","training","mode","."] | null | null | OPTActor | def __init__(
self,
pretrained: Optional[str] = None,
config: Optional[OPTConfig] = None,
checkpoint: bool = False,
lora_rank: int = 0,
lora_train_bias: str = "none",
) -> None:
if pretrained is not None:
model = OPTForCausalLM.from_pretrained(pretrained)
elif config is not None:
model = OPTForCausalLM(config)
else:
model = OPTForCausalLM(OPTConfig())
if checkpoint:
model.gradient_checkpointing_enable()
super().__init__(model, lora_rank, lora_train_bias)
| ["def","__init__","(","self",",","pretrained",":","Optional","[","str","]","=","None",",","config",":","Optional","[","OPTConfig","]","=","None",",","checkpoint",":","bool","=","False",",","lora_rank",":","int","=","0",",","lora_train_bias",":","str","=","``","none","''",",",")","-",">","None",":","if","pretrained","is","not","None",":","model","=","OPTForCausalLM.from_pretrained","(","pretrained",")","elif","config","is","not","None",":","model","=","OPTForCausalLM","(","config",")","else",":","model","=","OPTForCausalLM","(","OPTConfig","(",")",")","if","checkpoint",":","model.gradient_checkpointing_enable","(",")","super","(",")",".__init__","(","model",",","lora_rank",",","lora_train_bias",")"] | 21 | 37 | null | opt_actor.py | colossalai/applications/Chat/coati/models/opt/opt_actor.py | from typing import Optional
from transformers.models.opt.configuration_opt import OPTConfig
from transformers.models.opt.modeling_opt import OPTForCausalLM
from ..base import Actor | 7 | 1 | 4 | 1 | 1 | 0 | 1 | Use image node_id 1 to create a new OPTActor object from inherited base classes: Actor with example: obj = OPTActor(pretrained, config, checkpoint, lora_rank, lora_train_bias) | 175 | node_id 1 | 39,285 |
test_init | global | null | false | null | null | null | null | null | def test_init():
with pytest.raises(TypeError):
path.Path(None, None)
path.Path([], None)
path.Path(None, [])
assert isinstance(path.Path([], []), path.Path)
| ["def","test_init","(",")",":","with","pytest.raises","(","TypeError",")",":","path.Path","(","None",",","None",")","path.Path","(","[","]",",","None",")","path.Path","(","None",",","[","]",")","assert","isinstance","(","path.Path","(","[","]",",","[","]",")",",","path.Path",")"] | 6 | 12 | null | test_path.py | redis-py/tests/test_graph_utils/test_path.py | import pytest
from redis.commands.graph import edge, node, path | 7 | null | 2 | null | null | 5 | null | Use image node_id 1 for calling a global function with example usage: test_init() without return types | 102 | node_id 1 | 340,397 |
|
time_build_pixel_graph | GraphSuite | null | true | self | Benchmark for pixel graph routines in scikit-image. | ["Benchmark","for","pixel","graph","routines","in","scikit-image","."] | null | null | null | def time_build_pixel_graph(self):
graph.pixel_graph(self.skeleton, connectivity=2)
| ["def","time_build_pixel_graph","(","self",")",":","graph.pixel_graph","(","self.skeleton",",","connectivity=2",")"] | 29 | 30 | null | benchmark_graph.py | scikit-image/benchmarks/benchmark_graph.py | import numpy
from scipy import ndimage
from skimage import color, data, filters, graph, morphology | 7 | 1 | 3 | 0 | null | 0 | 3 | Use image node_id 2 for calling the GraphSuite obj's underlying member method code with example usage: obj.time_build_pixel_graph() without return types | 152 | node_id 2 | 344,035 |
time_central_pixel | GraphSuite | null | true | self | Benchmark for pixel graph routines in scikit-image. | ["Benchmark","for","pixel","graph","routines","in","scikit-image","."] | null | null | null | def time_central_pixel(self):
graph.central_pixel(self.g, self.n)
| ["def","time_central_pixel","(","self",")",":","graph.central_pixel","(","self.g",",","self.n",")"] | 32 | 33 | null | benchmark_graph.py | scikit-image/benchmarks/benchmark_graph.py | import numpy
from scipy import ndimage
from skimage import color, data, filters, graph, morphology | 7 | 1 | 3 | 0 | null | 0 | 3 | Use image node_id 3 for calling the GraphSuite obj's underlying member method code with example usage: obj.time_central_pixel() without return types | 148 | node_id 3 | 344,036 |
downgrade | global | null | false | null | null | null | null | null | def downgrade():
"""Unapply Increase length of ``Flask-AppBuilder`` ``ab_view_menu.name`` column"""
conn = op.get_bind()
inspector = inspect(conn)
tables = inspector.get_table_names()
if "ab_view_menu" in tables:
if conn.dialect.name == "sqlite":
op.execute("PRAGMA foreign_keys=off")
op.execute(
"""
CREATE TABLE IF NOT EXISTS ab_view_menu_dg_tmp
(
id INTEGER NOT NULL PRIMARY KEY,
name VARCHAR(100) NOT NULL UNIQUE
);
"""
)
op.execute(
"INSERT INTO ab_view_menu_dg_tmp(id, name) select id, name from ab_view_menu;"
)
op.execute("DROP TABLE ab_view_menu")
op.execute(
"ALTER TABLE ab_view_menu_dg_tmp rename to ab_view_menu;"
)
op.execute("PRAGMA foreign_keys=on")
else:
op.alter_column(
table_name="ab_view_menu",
column_name="name",
type_=sa.String(length=100),
nullable=False,
)
| ["def","downgrade","(",")",":","``","''","''","Unapply","Increase","length","of","``","Flask-AppBuilder","``","``","ab_view_menu.name","``","column","''","''","''","conn","=","op.get_bind","(",")","inspector","=","inspect","(","conn",")","tables","=","inspector.get_table_names","(",")","if","``","ab_view_menu","''","in","tables",":","if","conn.dialect.name","==","``","sqlite","''",":","op.execute","(","``","PRAGMA","foreign_keys=off","''",")","op.execute","(","``","''","''","CREATE","TABLE","IF","NOT","EXISTS","ab_view_menu_dg_tmp","(","id","INTEGER","NOT","NULL","PRIMARY","KEY",",","name","VARCHAR","(","100",")","NOT","NULL","UNIQUE",")",";","``","''","''",")","op.execute","(","``","INSERT","INTO","ab_view_menu_dg_tmp","(","id",",","name",")","select","id",",","name","from","ab_view_menu",";","''",")","op.execute","(","``","DROP","TABLE","ab_view_menu","''",")","op.execute","(","``","ALTER","TABLE","ab_view_menu_dg_tmp","rename","to","ab_view_menu",";","''",")","op.execute","(","``","PRAGMA","foreign_keys=on","''",")","else",":","op.alter_column","(","table_name=","''","ab_view_menu","''",",","column_name=","''","name","''",",","type_=sa.String","(","length=100",")",",","nullable=False",",",")"] | 72 | 96 | null | 0058_1_10_13_increase_length_of_fab_ab_view_menu_.py | airflow/airflow/migrations/versions/0058_1_10_13_increase_length_of_fab_ab_view_menu_.py | from __future__ import annotations
import sqlalchemy
from alembic import op
from sqlalchemy import inspect
from airflow.migrations.db_types import StringID | 7 | null | 5 | null | null | 2 | null | Use image node_id 2 for calling a global function with example usage: downgrade() without return types | 102 | node_id 2 | 2,606 |
|
conv2d_nchw_miopen | global | null | false | cfg,data,kernel,strides,padding,dilation,layout,out_dtype | null | null | null | null | miopen | def conv2d_nchw_miopen(
cfg,
data,
kernel,
strides,
padding,
dilation,
layout="NCHW",
out_dtype="float32",
):
"""Conv2D operator for rocm backend.
Parameters
----------
cfg: ConfigEntity
The config for this template
input : tvm.te.Tensor
4-D with shape [batch, in_channel, in_height, in_width]
filter : tvm.te.Tensor
4-D with shape [num_filter, in_channel, filter_height, filter_width]
strides : int or a list/tuple of two ints
stride size, or [stride_height, stride_width]
padding : int or a list/tuple of 2 or 4 ints
padding size, or
[pad_height, pad_width] for 2 ints, or
[pad_top, pad_left, pad_bottom, pad_right] for 4 ints
layout : str
layout of data
Returns
-------
output : tvm.te.Tensor
4-D with shape [batch, out_channel, out_height, out_width]
"""
CO, CI, KH, KW = get_const_tuple(kernel.shape)
N, _, H, W = get_const_tuple(data.shape)
assert layout == "NCHW"
# handle dilation
stride_h, stride_w = (
(strides, strides) if isinstance(strides, int) else strides
)
pt, pl, pb, pr = get_pad_tuple(padding, (KH, KW))
pad_h, pad_w = pt + pb, pl + pr
dilation_h, dilation_w = (
(dilation, dilation)
if isinstance(dilation, int)
else dilation
)
assert (pt == pb) and (pl == pr)
OH = (H + 2 * pad_h - KH) // stride_h + 1
OW = (W + 2 * pad_w - KW) // stride_w + 1
cfg.add_flop(
2
* N
* OH
* OW
* CO
* CI
* ((KH - 1) * dilation_h + 1)
* ((KW - 1) * dilation_w + 1)
)
return miopen.conv2d_forward(
data,
kernel,
stride_h,
stride_w,
pt,
pl,
dilation_h,
dilation_w,
conv_mode=0,
data_type=1,
)
| ["def","conv2d_nchw_miopen","(","cfg",",","data",",","kernel",",","strides",",","padding",",","dilation",",","layout=","''","NCHW","''",",","out_dtype=","''","float32","''",",",")",":","``","''","''","Conv2D","operator","for","rocm","backend",".","Parameters","--","--","--","--","--","cfg",":","ConfigEntity","The","config","for","this","template","input",":","tvm.te.Tensor","4-D","with","shape","[","batch",",","in_channel",",","in_height",",","in_width","]","filter",":","tvm.te.Tensor","4-D","with","shape","[","num_filter",",","in_channel",",","filter_height",",","filter_width","]","strides",":","int","or","a","list\/tuple","of","two","ints","stride","size",",","or","[","stride_height",",","stride_width","]","padding",":","int","or","a","list\/tuple","of","2","or","4","ints","padding","size",",","or","[","pad_height",",","pad_width","]","for","2","ints",",","or","[","pad_top",",","pad_left",",","pad_bottom",",","pad_right","]","for","4","ints","layout",":","str","layout","of","data","Returns","--","--","--","-","output",":","tvm.te.Tensor","4-D","with","shape","[","batch",",","out_channel",",","out_height",",","out_width","]","``","''","''","CO",",","CI",",","KH",",","KW","=","get_const_tuple","(","kernel.shape",")","N",",","_",",","H",",","W","=","get_const_tuple","(","data.shape",")","assert","layout","==","``","NCHW","''","#","handle","dilation","stride_h",",","stride_w","=","(","(","strides",",","strides",")","if","isinstance","(","strides",",","int",")","else","strides",")","pt",",","pl",",","pb",",","pr","=","get_pad_tuple","(","padding",",","(","KH",",","KW",")",")","pad_h",",","pad_w","=","pt","+","pb",",","pl","+","pr","dilation_h",",","dilation_w","=","(","(","dilation",",","dilation",")","if","isinstance","(","dilation",",","int",")","else","dilation",")","assert","(","pt","==","pb",")","and","(","pl","==","pr",")","OH","=","(","H","+","2","*","pad_h","-","KH",")","\/\/","stride_h","+","1","OW","=","(","W","+","2","*","pad_w","-","KW",")","\/\/","stride_w","+","1","cfg.add_flop","(","2","*","N","*","OH","*","OW","*","CO","*","CI","*","(","(","KH","-","1",")","*","dilation_h","+","1",")","*","(","(","KW","-","1",")","*","dilation_w","+","1",")",")","return","miopen.conv2d_forward","(","data",",","kernel",",","stride_h",",","stride_w",",","pt",",","pl",",","dilation_h",",","dilation_w",",","conv_mode=0",",","data_type=1",",",")"] | 28 | 80 | null | conv2d.py | tvm/python/tvm/topi/rocm/conv2d.py | from tvm import autotvm
from tvm.contrib import miopen
from ..None import generic
from ..utils import get_const_tuple
from ..nn.utils import get_pad_tuple | 7 | null | 5 | null | null | 2 | null | Use image node_id 1 for calling a global function with example usage: conv2d_nchw_miopen(cfg, data, kernel, strides, padding, dilation, layout, out_dtype) and returns: miopen | 174 | node_id 1 | 497,563 |
init | LSTMBenchmark | op_bench | true | self,I,H,NL,B,D,dtype | null | null | null | null | null | def init(self, I, H, NL, B, D, dtype):
sequence_len = 128
batch_size = 16
# The quantized.dynamic.LSTM has a bug. That's why we create a regular
# LSTM, and quantize it later. See issue #31192.
scale = 1.0 / 256
zero_point = 0
cell_nn = nn.LSTM(
input_size=I,
hidden_size=H,
num_layers=NL,
bias=B,
batch_first=False,
dropout=0.0,
bidirectional=D,
)
cell_temp = nn.Sequential(cell_nn)
self.cell = torch.ao.quantization.quantize_dynamic(
cell_temp, {nn.LSTM, nn.Linear}, dtype=dtype
)[0]
x = torch.randn(
sequence_len, batch_size, I # sequence length # batch size
) # Number of features in X
h = torch.randn(
NL * (D + 1),
batch_size,
H, # layer_num * dir_num # batch size
) # hidden size
c = torch.randn(
NL * (D + 1),
batch_size,
H, # layer_num * dir_num # batch size
) # hidden size
self.inputs = {"x": x, "h": h, "c": c}
self.set_module_name("QLSTM")
| ["def","init","(","self",",","I",",","H",",","NL",",","B",",","D",",","dtype",")",":","sequence_len","=","128","batch_size","=","16","#","The","quantized.dynamic.LSTM","has","a","bug",".","That","'s","why","we","create","a","regular","#","LSTM",",","and","quantize","it","later",".","See","issue","#","31192.","scale","=","1.0","\/","256","zero_point","=","0","cell_nn","=","nn.LSTM","(","input_size=I",",","hidden_size=H",",","num_layers=NL",",","bias=B",",","batch_first=False",",","dropout=0.0",",","bidirectional=D",",",")","cell_temp","=","nn.Sequential","(","cell_nn",")","self.cell","=","torch.ao.quantization.quantize_dynamic","(","cell_temp",",","{","nn.LSTM",",","nn.Linear","}",",","dtype=dtype",")","[","0","]","x","=","torch.randn","(","sequence_len",",","batch_size",",","I","#","sequence","length","#","batch","size",")","#","Number","of","features","in","X","h","=","torch.randn","(","NL","*","(","D","+","1",")",",","batch_size",",","H",",","#","layer_num","*","dir_num","#","batch","size",")","#","hidden","size","c","=","torch.randn","(","NL","*","(","D","+","1",")",",","batch_size",",","H",",","#","layer_num","*","dir_num","#","batch","size",")","#","hidden","size","self.inputs","=","{","``","x","''",":","x",",","``","h","''",":","h",",","``","c","''",":","c","}","self.set_module_name","(","``","QLSTM","''",")"] | 27 | 60 | null | qrnn_test.py | pytorch/benchmarks/operator_benchmark/pt/qrnn_test.py | import torch
from torch import nn
import operator_benchmark | 7 | 1 | 3 | 1 | 1 | 0 | 2 | Use image node_id 1 for calling the LSTMBenchmark obj's underlying member method code with example usage: obj.init(I, H, NL, B, D, dtype) without return types | 158 | node_id 1 | 271,988 |