Unnamed: 0
int64
0
2.44k
repo
stringlengths
32
81
hash
stringlengths
40
40
diff
stringlengths
113
1.17k
old_path
stringlengths
5
84
rewrite
stringlengths
34
79
initial_state
stringlengths
75
980
final_state
stringlengths
76
980
800
https://:@github.com/ankush13r/BvSalud.git
10216a189e9f63ab1a1d8cf7c647a026b7b941bb
@@ -195,7 +195,7 @@ def get_mesh_major_list(document_dict,decsCodes_list_dict,with_header): #Method else: if len(header_before_slash) != 0: final_header = str(header_code) +'/'+ str(header_after_slash) - elif len(header_after_slash) == 0: + elif len(header_before_slash) == 0: final_header = '/' + str(decsCodes_list_dict) else: print(header,"--",header_before_slash,"--",header_after_slash)
BvSalud/makeSet.py
ReplaceText(target='header_before_slash' @(198,21)->(198,39))
def get_mesh_major_list(document_dict,decsCodes_list_dict,with_header): #Method else: if len(header_before_slash) != 0: final_header = str(header_code) +'/'+ str(header_after_slash) elif len(header_after_slash) == 0: final_header = '/' + str(decsCodes_list_dict) else: print(header,"--",header_before_slash,"--",header_after_slash)
def get_mesh_major_list(document_dict,decsCodes_list_dict,with_header): #Method else: if len(header_before_slash) != 0: final_header = str(header_code) +'/'+ str(header_after_slash) elif len(header_before_slash) == 0: final_header = '/' + str(decsCodes_list_dict) else: print(header,"--",header_before_slash,"--",header_after_slash)
801
https://:@github.com/Hypers-HFA/paraer.git
8a5fd8156012b94e50b1e6a3413244b95f3e323b
@@ -112,7 +112,7 @@ def para_ok_or_400(itemset): ] value = None # 与 '' 区别 para = paramap.get(name) - if required and para not in (None, ''): # 如果是post方法并且传参是json的话,para可能为0 + if required and para in (None, ''): # 如果是post方法并且传参是json的话,para可能为0 result.error(name, 'required') if para is not None: if para:
paraer/para.py
ReplaceText(target=' in ' @(115,36)->(115,44))
def para_ok_or_400(itemset): ] value = None # 与 '' 区别 para = paramap.get(name) if required and para not in (None, ''): # 如果是post方法并且传参是json的话,para可能为0 result.error(name, 'required') if para is not None: if para:
def para_ok_or_400(itemset): ] value = None # 与 '' 区别 para = paramap.get(name) if required and para in (None, ''): # 如果是post方法并且传参是json的话,para可能为0 result.error(name, 'required') if para is not None: if para:
802
https://:@github.com/davidenunes/tensorx.git
34fb24ae4eeca927a7d7b5f376aeeddbb0b7aacd
@@ -2604,7 +2604,7 @@ class SaltPepperNoise(Layer): else: batch_size = noise_shape[0] - noise = salt_pepper_noise(batch_size, noise_shape[1], density, salt_value, pepper_value, seed) + noise = salt_pepper_noise(noise_shape[1], batch_size, density, salt_value, pepper_value, seed) if layer.is_sparse(): tensor = transform.sparse_put(layer.tensor, noise)
tensorx/layers.py
ArgSwap(idxs=0<->1 @(2607,24)->(2607,41))
class SaltPepperNoise(Layer): else: batch_size = noise_shape[0] noise = salt_pepper_noise(batch_size, noise_shape[1], density, salt_value, pepper_value, seed) if layer.is_sparse(): tensor = transform.sparse_put(layer.tensor, noise)
class SaltPepperNoise(Layer): else: batch_size = noise_shape[0] noise = salt_pepper_noise(noise_shape[1], batch_size, density, salt_value, pepper_value, seed) if layer.is_sparse(): tensor = transform.sparse_put(layer.tensor, noise)
803
https://:@github.com/sauliusl/dgw.git
9120263028d2e64983b9b9babe31c46b8f2caedd
@@ -98,7 +98,7 @@ def dtw_std(x, y, metric='sqeuclidean', dist_only=True, constraint=None, k=None, path_rev = (_reverse_path(path[0]), path[1]) if scale_first: - path_rev = _scaled_path(path, scaling_path, flip_paths) + path_rev = _scaled_path(path_rev, scaling_path, flip_paths) # TODO: Reverse cost matrix here cost = None
dgw/dtw/distance.py
ReplaceText(target='path_rev' @(101,40)->(101,44))
def dtw_std(x, y, metric='sqeuclidean', dist_only=True, constraint=None, k=None, path_rev = (_reverse_path(path[0]), path[1]) if scale_first: path_rev = _scaled_path(path, scaling_path, flip_paths) # TODO: Reverse cost matrix here cost = None
def dtw_std(x, y, metric='sqeuclidean', dist_only=True, constraint=None, k=None, path_rev = (_reverse_path(path[0]), path[1]) if scale_first: path_rev = _scaled_path(path_rev, scaling_path, flip_paths) # TODO: Reverse cost matrix here cost = None
804
https://:@bitbucket.org/slicedice/aboutyou-shop-sdk-python.git
c125675219cbb473f47c78c57e9205d1be44d4f2
@@ -439,7 +439,7 @@ class Search(object): else: product = Product(self.easy, p) - self.easy.product_cach[p.id] = product + self.easy.product_cach[product.id] = product self.products.buffer[i+offset] = product
collins/easy.py
ReplaceText(target='product' @(442,35)->(442,36))
class Search(object): else: product = Product(self.easy, p) self.easy.product_cach[p.id] = product self.products.buffer[i+offset] = product
class Search(object): else: product = Product(self.easy, p) self.easy.product_cach[product.id] = product self.products.buffer[i+offset] = product
805
https://:@github.com/neizod/extmath.git
b66496463e3d724a29d652e97867f41e4e1d2aea
@@ -117,7 +117,7 @@ def factorized(n): break while not n % i: factor.append(i) - n /= i + n //= i if n == 1: break return factor
mathapi/__init__.py
ReplaceText(target='//=' @(120,14)->(120,16))
def factorized(n): break while not n % i: factor.append(i) n /= i if n == 1: break return factor
def factorized(n): break while not n % i: factor.append(i) n //= i if n == 1: break return factor
806
https://:@github.com/storborg/axibot.git
4ffd9fc0f715bbe4ae8c1708fbeae45186d35f6e
@@ -64,7 +64,7 @@ async def handle_user_message(app, ws, msg): except Exception as e: notify_error(app, ws, str(e)) else: - app['document'] = msg.document + app['document'] = job.document app['job'] = job app['estimated_time'] = job.duration().total_seconds() notify_new_document(app, exclude_client=ws)
axibot/server/handlers.py
ReplaceText(target='job' @(67,30)->(67,33))
async def handle_user_message(app, ws, msg): except Exception as e: notify_error(app, ws, str(e)) else: app['document'] = msg.document app['job'] = job app['estimated_time'] = job.duration().total_seconds() notify_new_document(app, exclude_client=ws)
async def handle_user_message(app, ws, msg): except Exception as e: notify_error(app, ws, str(e)) else: app['document'] = job.document app['job'] = job app['estimated_time'] = job.duration().total_seconds() notify_new_document(app, exclude_client=ws)
807
https://:@github.com/storborg/axibot.git
4ad752828696a789ef25fa001f1489f87344c3e4
@@ -116,7 +116,7 @@ async def client_handler(request): elif raw_msg.tp == aiohttp.MsgType.closed: break elif raw_msg.tp == aiohttp.MsgType.error: - log.info("User websocket error: %s", msg) + log.info("User websocket error: %s", raw_msg) break else: log.error("Unknown user message type: %s, ignoring.",
axibot/server/handlers.py
ReplaceText(target='raw_msg' @(119,53)->(119,56))
async def client_handler(request): elif raw_msg.tp == aiohttp.MsgType.closed: break elif raw_msg.tp == aiohttp.MsgType.error: log.info("User websocket error: %s", msg) break else: log.error("Unknown user message type: %s, ignoring.",
async def client_handler(request): elif raw_msg.tp == aiohttp.MsgType.closed: break elif raw_msg.tp == aiohttp.MsgType.error: log.info("User websocket error: %s", raw_msg) break else: log.error("Unknown user message type: %s, ignoring.",
808
https://:@github.com/tsuyukimakoto/biisan.git
4465b109a4e71dcb64499c14b579c762e54b4db4
@@ -116,7 +116,7 @@ def process_image(elm, registry, container): img.alt = subitem[1] elif subitem[0] == 'witdh': img.witdh = subitem[1] - elif subitem[1] == 'height': + elif subitem[0] == 'height': img.height = subitem[1] elif subitem[0] == 'uri': img.uri = subitem[1]
biisan/processors/__init__.py
ReplaceText(target='0' @(119,21)->(119,22))
def process_image(elm, registry, container): img.alt = subitem[1] elif subitem[0] == 'witdh': img.witdh = subitem[1] elif subitem[1] == 'height': img.height = subitem[1] elif subitem[0] == 'uri': img.uri = subitem[1]
def process_image(elm, registry, container): img.alt = subitem[1] elif subitem[0] == 'witdh': img.witdh = subitem[1] elif subitem[0] == 'height': img.height = subitem[1] elif subitem[0] == 'uri': img.uri = subitem[1]
809
https://:@github.com/csm10495/csmlog.git
010e00e53545842d2dc1342c375c068c90138ca2
@@ -262,7 +262,7 @@ def test_gsheets_calculate_periodic_loop_sleep_time(gsheets_handler_no_thread): def test_gsheets_add_rows_to_active_sheet_sets_add_rows_time(gsheets_handler_no_thread): gsheets_handler_no_thread._add_rows_time = 99 assert isinstance(gsheets_handler_no_thread._add_rows_to_active_sheet([]), unittest.mock.Mock) - assert gsheets_handler_no_thread._add_rows_time < 99 and gsheets_handler_no_thread._add_rows_time > 0 + assert gsheets_handler_no_thread._add_rows_time < 99 and gsheets_handler_no_thread._add_rows_time >= 0 def test_gsheets_add_rows_to_active_sheet_set_coerce_to_correct_exceptions(gsheets_handler_no_thread): with unittest.mock.patch.object(gsheets_handler_no_thread.sheet, 'append_rows', side_effect=Exception("RESOURCE_EXHAUSTED uh-oh")) as mock:
csmlog/tests/test_google_sheets_handler.py
ReplaceText(target='>=' @(265,102)->(265,103))
def test_gsheets_calculate_periodic_loop_sleep_time(gsheets_handler_no_thread): def test_gsheets_add_rows_to_active_sheet_sets_add_rows_time(gsheets_handler_no_thread): gsheets_handler_no_thread._add_rows_time = 99 assert isinstance(gsheets_handler_no_thread._add_rows_to_active_sheet([]), unittest.mock.Mock) assert gsheets_handler_no_thread._add_rows_time < 99 and gsheets_handler_no_thread._add_rows_time > 0 def test_gsheets_add_rows_to_active_sheet_set_coerce_to_correct_exceptions(gsheets_handler_no_thread): with unittest.mock.patch.object(gsheets_handler_no_thread.sheet, 'append_rows', side_effect=Exception("RESOURCE_EXHAUSTED uh-oh")) as mock:
def test_gsheets_calculate_periodic_loop_sleep_time(gsheets_handler_no_thread): def test_gsheets_add_rows_to_active_sheet_sets_add_rows_time(gsheets_handler_no_thread): gsheets_handler_no_thread._add_rows_time = 99 assert isinstance(gsheets_handler_no_thread._add_rows_to_active_sheet([]), unittest.mock.Mock) assert gsheets_handler_no_thread._add_rows_time < 99 and gsheets_handler_no_thread._add_rows_time >= 0 def test_gsheets_add_rows_to_active_sheet_set_coerce_to_correct_exceptions(gsheets_handler_no_thread): with unittest.mock.patch.object(gsheets_handler_no_thread.sheet, 'append_rows', side_effect=Exception("RESOURCE_EXHAUSTED uh-oh")) as mock:
810
https://:@github.com/JorrandeWit/django-oscar-fees.git
7d18c7bdbeb881f4af5f81c1adf989422e134995
@@ -26,7 +26,7 @@ class FeeApplications(object): 'fee': fee, 'result': result, 'name': fee.name, - 'description': result.description, + 'description': fee.description, 'freq': 0, 'amount': D('0.00')} self.applications[fee.id]['amount'] += result.fee
django_oscar_fees/results.py
ReplaceText(target='fee' @(29,31)->(29,37))
class FeeApplications(object): 'fee': fee, 'result': result, 'name': fee.name, 'description': result.description, 'freq': 0, 'amount': D('0.00')} self.applications[fee.id]['amount'] += result.fee
class FeeApplications(object): 'fee': fee, 'result': result, 'name': fee.name, 'description': fee.description, 'freq': 0, 'amount': D('0.00')} self.applications[fee.id]['amount'] += result.fee
811
https://:@gitlab.com/yhtang/graphdot.git
2ac46e719092b1e78bc2bd860bce1dd23e152638
@@ -27,7 +27,7 @@ class OctileGraph(object): nnz = len(edges) ''' add phantom label if none exists to facilitate C++ interop ''' - assert(len(edges.columns) >= 1) + assert(len(nodes.columns) >= 1) if len(nodes.columns) == 1: nodes['labeled'] = np.zeros(len(nodes), np.bool_)
graphdot/kernel/marginalized/_octilegraph.py
ReplaceText(target='nodes' @(30,19)->(30,24))
class OctileGraph(object): nnz = len(edges) ''' add phantom label if none exists to facilitate C++ interop ''' assert(len(edges.columns) >= 1) if len(nodes.columns) == 1: nodes['labeled'] = np.zeros(len(nodes), np.bool_)
class OctileGraph(object): nnz = len(edges) ''' add phantom label if none exists to facilitate C++ interop ''' assert(len(nodes.columns) >= 1) if len(nodes.columns) == 1: nodes['labeled'] = np.zeros(len(nodes), np.bool_)
812
https://:@gitlab.com/yhtang/graphdot.git
c1d46fe384aaac63a6e32b0a30cf6f3aa2b04183
@@ -315,7 +315,7 @@ class MarginalizedGraphKernel: if traits.eval_gradient is True: output_shape = (output_length, 1 + self.n_dims) else: - output_shape = (output_shape, 1) + output_shape = (output_length, 1) output = backend.empty(int(np.prod(output_shape)), np.float32) timer.toc('creating output buffer')
graphdot/kernel/marginalized/_kernel.py
ReplaceText(target='output_length' @(318,28)->(318,40))
class MarginalizedGraphKernel: if traits.eval_gradient is True: output_shape = (output_length, 1 + self.n_dims) else: output_shape = (output_shape, 1) output = backend.empty(int(np.prod(output_shape)), np.float32) timer.toc('creating output buffer')
class MarginalizedGraphKernel: if traits.eval_gradient is True: output_shape = (output_length, 1 + self.n_dims) else: output_shape = (output_length, 1) output = backend.empty(int(np.prod(output_shape)), np.float32) timer.toc('creating output buffer')
813
https://:@gitlab.com/yhtang/graphdot.git
f1499423ac76e790b77f729e92b2cc3223d64f58
@@ -117,7 +117,7 @@ class SpectralApprox(FactorApprox): def __init__(self, X, rcut=0, acut=0): if isinstance(X, np.ndarray): U, S, _ = np.linalg.svd(X, full_matrices=False) - mask = (S >= S.max() * rcut) | (S >= acut) + mask = (S >= S.max() * rcut) & (S >= acut) self.U = U[:, mask] self.S = S[mask] elif isinstance(X, tuple) and len(X) == 2:
graphdot/linalg/low_rank.py
ReplaceText(target='&' @(120,41)->(120,42))
class SpectralApprox(FactorApprox): def __init__(self, X, rcut=0, acut=0): if isinstance(X, np.ndarray): U, S, _ = np.linalg.svd(X, full_matrices=False) mask = (S >= S.max() * rcut) | (S >= acut) self.U = U[:, mask] self.S = S[mask] elif isinstance(X, tuple) and len(X) == 2:
class SpectralApprox(FactorApprox): def __init__(self, X, rcut=0, acut=0): if isinstance(X, np.ndarray): U, S, _ = np.linalg.svd(X, full_matrices=False) mask = (S >= S.max() * rcut) & (S >= acut) self.U = U[:, mask] self.S = S[mask] elif isinstance(X, tuple) and len(X) == 2:
814
https://:@github.com/straga/micropython.git
1679696612007107dac55d936006b1923eda2475
@@ -7,7 +7,7 @@ desc = { data = bytearray(b"01234567") -S = uctypes.struct(desc, uctypes.addressof(data), uctypes.LITTLE_ENDIAN) +S = uctypes.struct(uctypes.addressof(data), desc, uctypes.LITTLE_ENDIAN) # Arrays of UINT8 are accessed as bytearrays print(S.arr)
tests/extmod/uctypes_bytearray.py
ArgSwap(idxs=0<->1 @(10,4)->(10,18))
desc = { data = bytearray(b"01234567") S = uctypes.struct(desc, uctypes.addressof(data), uctypes.LITTLE_ENDIAN) # Arrays of UINT8 are accessed as bytearrays print(S.arr)
desc = { data = bytearray(b"01234567") S = uctypes.struct(uctypes.addressof(data), desc, uctypes.LITTLE_ENDIAN) # Arrays of UINT8 are accessed as bytearrays print(S.arr)
815
https://:@github.com/straga/micropython.git
1679696612007107dac55d936006b1923eda2475
@@ -22,7 +22,7 @@ desc = { data = bytearray(b"01") -S = uctypes.struct(desc, uctypes.addressof(data), uctypes.LITTLE_ENDIAN) +S = uctypes.struct(uctypes.addressof(data), desc, uctypes.LITTLE_ENDIAN) #print(S) print(hex(S.s0))
tests/extmod/uctypes_le.py
ArgSwap(idxs=0<->1 @(25,4)->(25,18))
desc = { data = bytearray(b"01") S = uctypes.struct(desc, uctypes.addressof(data), uctypes.LITTLE_ENDIAN) #print(S) print(hex(S.s0))
desc = { data = bytearray(b"01") S = uctypes.struct(uctypes.addressof(data), desc, uctypes.LITTLE_ENDIAN) #print(S) print(hex(S.s0))
816
https://:@github.com/straga/micropython.git
1679696612007107dac55d936006b1923eda2475
@@ -31,7 +31,7 @@ desc = { data = bytearray(b"01") -S = uctypes.struct(desc, uctypes.addressof(data), uctypes.NATIVE) +S = uctypes.struct(uctypes.addressof(data), desc, uctypes.NATIVE) #print(S) print(hex(S.s0))
tests/extmod/uctypes_native_le.py
ArgSwap(idxs=0<->1 @(34,4)->(34,18))
desc = { data = bytearray(b"01") S = uctypes.struct(desc, uctypes.addressof(data), uctypes.NATIVE) #print(S) print(hex(S.s0))
desc = { data = bytearray(b"01") S = uctypes.struct(uctypes.addressof(data), desc, uctypes.NATIVE) #print(S) print(hex(S.s0))
817
https://:@github.com/straga/micropython.git
1679696612007107dac55d936006b1923eda2475
@@ -16,7 +16,7 @@ bytes = b"01" addr = uctypes.addressof(bytes) buf = addr.to_bytes(uctypes.sizeof(desc)) -S = uctypes.struct(desc, uctypes.addressof(buf), uctypes.LITTLE_ENDIAN) +S = uctypes.struct(uctypes.addressof(buf), desc, uctypes.LITTLE_ENDIAN) print(S.ptr[0]) assert S.ptr[0] == ord("0")
tests/extmod/uctypes_ptr_le.py
ArgSwap(idxs=0<->1 @(19,4)->(19,18))
bytes = b"01" addr = uctypes.addressof(bytes) buf = addr.to_bytes(uctypes.sizeof(desc)) S = uctypes.struct(desc, uctypes.addressof(buf), uctypes.LITTLE_ENDIAN) print(S.ptr[0]) assert S.ptr[0] == ord("0")
bytes = b"01" addr = uctypes.addressof(bytes) buf = addr.to_bytes(uctypes.sizeof(desc)) S = uctypes.struct(uctypes.addressof(buf), desc, uctypes.LITTLE_ENDIAN) print(S.ptr[0]) assert S.ptr[0] == ord("0")
818
https://:@github.com/IL2HorusTeam/il2fb-ds-airbridge.git
bc69d1a47a4994071d0b46c7dd29aebf6a7d4169
@@ -60,7 +60,7 @@ class TextFileWatchDog: def stop(self) -> None: with self._stop_lock: - self._do_stop = False + self._do_stop = True def run(self) -> None: try:
il2fb/ds/airbridge/watch_dog.py
ReplaceText(target='True' @(63,28)->(63,33))
class TextFileWatchDog: def stop(self) -> None: with self._stop_lock: self._do_stop = False def run(self) -> None: try:
class TextFileWatchDog: def stop(self) -> None: with self._stop_lock: self._do_stop = True def run(self) -> None: try:
819
https://:@github.com/south-coast-science/scs_dfe_eng.git
060bd36c51a20f4586ed199b1e3996c70b0318a7
@@ -49,7 +49,7 @@ class AFE(object): self.__wrk_adc = ADS1115(ADS1115.ADDR_WRK, AFE.__RATE) self.__aux_adc = ADS1115(ADS1115.ADDR_AUX, AFE.__RATE) - self.__pt1000_adc = pt1000_conf.adc(MCP342X.GAIN_4, MCP342X.RATE_15) if pt1000_conf else None + self.__pt1000_adc = pt1000_conf.adc(MCP342X.GAIN_4, MCP342X.RATE_15) if pt1000 else None self.__tconv = self.__wrk_adc.tconv
scs_dfe/gas/afe.py
ReplaceText(target='pt1000' @(52,80)->(52,91))
class AFE(object): self.__wrk_adc = ADS1115(ADS1115.ADDR_WRK, AFE.__RATE) self.__aux_adc = ADS1115(ADS1115.ADDR_AUX, AFE.__RATE) self.__pt1000_adc = pt1000_conf.adc(MCP342X.GAIN_4, MCP342X.RATE_15) if pt1000_conf else None self.__tconv = self.__wrk_adc.tconv
class AFE(object): self.__wrk_adc = ADS1115(ADS1115.ADDR_WRK, AFE.__RATE) self.__aux_adc = ADS1115(ADS1115.ADDR_AUX, AFE.__RATE) self.__pt1000_adc = pt1000_conf.adc(MCP342X.GAIN_4, MCP342X.RATE_15) if pt1000 else None self.__tconv = self.__wrk_adc.tconv
820
https://:@github.com/ionox0/toil.git
3f065a27c433b1cb9c5266ed4c28b643419d4466
@@ -456,7 +456,7 @@ def mainLoop(config, batchSystem): reissueOverLongJobs(updatedJobFiles, jobBatcher, config, batchSystem, childJobFileToParentJob, childCounts) logger.info("Reissued any over long jobs") - hasNoMissingJobs = reissueMissingJobs(updatedJobFiles, jobBatcher, batchSystem, childJobFileToParentJob, config, childCounts) + hasNoMissingJobs = reissueMissingJobs(updatedJobFiles, jobBatcher, batchSystem, childJobFileToParentJob, childCounts, config) if hasNoMissingJobs: timeSinceJobsLastRescued = time.time() else:
src/master.py
ArgSwap(idxs=4<->5 @(459,35)->(459,53))
def mainLoop(config, batchSystem): reissueOverLongJobs(updatedJobFiles, jobBatcher, config, batchSystem, childJobFileToParentJob, childCounts) logger.info("Reissued any over long jobs") hasNoMissingJobs = reissueMissingJobs(updatedJobFiles, jobBatcher, batchSystem, childJobFileToParentJob, config, childCounts) if hasNoMissingJobs: timeSinceJobsLastRescued = time.time() else:
def mainLoop(config, batchSystem): reissueOverLongJobs(updatedJobFiles, jobBatcher, config, batchSystem, childJobFileToParentJob, childCounts) logger.info("Reissued any over long jobs") hasNoMissingJobs = reissueMissingJobs(updatedJobFiles, jobBatcher, batchSystem, childJobFileToParentJob, childCounts, config) if hasNoMissingJobs: timeSinceJobsLastRescued = time.time() else:
821
https://:@github.com/ionox0/toil.git
a0e4eadb87f09791088b6c11dc6aec0bf8002991
@@ -99,7 +99,7 @@ please ensure you re-import targets defined in main" % self.__class__.__name__) """Takes a file (as a path) and uploads it to to the global file store, returns an ID that can be used to retrieve the file. """ - return self.jobStore.writeFile(localFileName, self.job.jobStoreID) + return self.jobStore.writeFile(self.job.jobStoreID, localFileName) def updateGlobalFile(self, fileStoreID, localFileName): """Replaces the existing version of a file in the global file store, keyed by the fileStoreID.
src/target.py
ArgSwap(idxs=0<->1 @(102,15)->(102,38))
please ensure you re-import targets defined in main" % self.__class__.__name__) """Takes a file (as a path) and uploads it to to the global file store, returns an ID that can be used to retrieve the file. """ return self.jobStore.writeFile(localFileName, self.job.jobStoreID) def updateGlobalFile(self, fileStoreID, localFileName): """Replaces the existing version of a file in the global file store, keyed by the fileStoreID.
please ensure you re-import targets defined in main" % self.__class__.__name__) """Takes a file (as a path) and uploads it to to the global file store, returns an ID that can be used to retrieve the file. """ return self.jobStore.writeFile(self.job.jobStoreID, localFileName) def updateGlobalFile(self, fileStoreID, localFileName): """Replaces the existing version of a file in the global file store, keyed by the fileStoreID.
822
https://:@github.com/ionox0/toil.git
06718f425d5981afabd30bcfa39d866edb5c09b4
@@ -46,7 +46,7 @@ class Job( object ): #The IDs of predecessors that have finished. #When len(predecessorsFinished) == predecessorNumber then the #job can be run. - self.predecessorsFinished = predecessorNumber or set() + self.predecessorsFinished = predecessorsFinished or set() #The list of successor jobs to run. Successor jobs are stored #as 4-tuples of the form (jobStoreId, memory, cpu, predecessorNumber).
src/jobTree/job.py
ReplaceText(target='predecessorsFinished' @(49,36)->(49,53))
class Job( object ): #The IDs of predecessors that have finished. #When len(predecessorsFinished) == predecessorNumber then the #job can be run. self.predecessorsFinished = predecessorNumber or set() #The list of successor jobs to run. Successor jobs are stored #as 4-tuples of the form (jobStoreId, memory, cpu, predecessorNumber).
class Job( object ): #The IDs of predecessors that have finished. #When len(predecessorsFinished) == predecessorNumber then the #job can be run. self.predecessorsFinished = predecessorsFinished or set() #The list of successor jobs to run. Successor jobs are stored #as 4-tuples of the form (jobStoreId, memory, cpu, predecessorNumber).
823
https://:@github.com/ionox0/toil.git
47ec5287d747036c524100cc49fa4262a73293cf
@@ -107,7 +107,7 @@ class JobWrapper( object ): """ if self.logJobStoreFileID is not None: self.clearLogFile(jobStore) - self.logJobStoreFileID = jobStore.writeFile( self.jobStoreID, logFile ) + self.logJobStoreFileID = jobStore.writeFile( logFile, self.jobStoreID ) assert self.logJobStoreFileID is not None def getLogFileHandle( self, jobStore ):
src/toil/jobWrapper.py
ArgSwap(idxs=0<->1 @(110,33)->(110,51))
class JobWrapper( object ): """ if self.logJobStoreFileID is not None: self.clearLogFile(jobStore) self.logJobStoreFileID = jobStore.writeFile( self.jobStoreID, logFile ) assert self.logJobStoreFileID is not None def getLogFileHandle( self, jobStore ):
class JobWrapper( object ): """ if self.logJobStoreFileID is not None: self.clearLogFile(jobStore) self.logJobStoreFileID = jobStore.writeFile( logFile, self.jobStoreID ) assert self.logJobStoreFileID is not None def getLogFileHandle( self, jobStore ):
824
https://:@github.com/ionox0/toil.git
cad17f6cbe391264e00675c51a8bfe9221ba1091
@@ -159,7 +159,7 @@ class AbstractJobStore( object ): break #Reset the retry count of the jobWrapper - if jobWrapper.remainingRetryCount < self._defaultTryCount(): + if jobWrapper.remainingRetryCount != self._defaultTryCount(): jobWrapper.remainingRetryCount = self._defaultTryCount() changed = True
src/toil/jobStores/abstractJobStore.py
ReplaceText(target='!=' @(162,46)->(162,47))
class AbstractJobStore( object ): break #Reset the retry count of the jobWrapper if jobWrapper.remainingRetryCount < self._defaultTryCount(): jobWrapper.remainingRetryCount = self._defaultTryCount() changed = True
class AbstractJobStore( object ): break #Reset the retry count of the jobWrapper if jobWrapper.remainingRetryCount != self._defaultTryCount(): jobWrapper.remainingRetryCount = self._defaultTryCount() changed = True
825
https://:@github.com/ionox0/toil.git
dcd5e6f86c0f9d376140f4626e97666348683b73
@@ -437,7 +437,7 @@ class ModuleDescriptor(namedtuple('ModuleDescriptor', ('dirPath', 'name'))): :rtype: toil.resource.Resource """ - if self._runningOnWorker(): + if not self._runningOnWorker(): log.warn('The localize() method should only be invoked on a worker.') resource = Resource.lookup(self._resourcePath) if resource is None:
src/toil/resource.py
ReplaceText(target='not ' @(440,11)->(440,11))
class ModuleDescriptor(namedtuple('ModuleDescriptor', ('dirPath', 'name'))): :rtype: toil.resource.Resource """ if self._runningOnWorker(): log.warn('The localize() method should only be invoked on a worker.') resource = Resource.lookup(self._resourcePath) if resource is None:
class ModuleDescriptor(namedtuple('ModuleDescriptor', ('dirPath', 'name'))): :rtype: toil.resource.Resource """ if not self._runningOnWorker(): log.warn('The localize() method should only be invoked on a worker.') resource = Resource.lookup(self._resourcePath) if resource is None:
826
https://:@github.com/ionox0/toil.git
bee00f3fc4f68af0e7c00b884627f728ca20f792
@@ -649,7 +649,7 @@ def main(args=None, stdout=sys.stdout): if options.logLevel: cwllogger.setLevel(options.logLevel) - useStrict = not args.not_strict + useStrict = not options.not_strict try: t = cwltool.load_tool.load_tool(options.cwltool, cwltool.workflow.defaultMakeTool, resolver=cwltool.resolver.tool_resolver, strict=useStrict)
src/toil/cwl/cwltoil.py
ReplaceText(target='options' @(652,20)->(652,24))
def main(args=None, stdout=sys.stdout): if options.logLevel: cwllogger.setLevel(options.logLevel) useStrict = not args.not_strict try: t = cwltool.load_tool.load_tool(options.cwltool, cwltool.workflow.defaultMakeTool, resolver=cwltool.resolver.tool_resolver, strict=useStrict)
def main(args=None, stdout=sys.stdout): if options.logLevel: cwllogger.setLevel(options.logLevel) useStrict = not options.not_strict try: t = cwltool.load_tool.load_tool(options.cwltool, cwltool.workflow.defaultMakeTool, resolver=cwltool.resolver.tool_resolver, strict=useStrict)
827
https://:@github.com/ionox0/toil.git
231b6665bde956aae5b3469a7147e2358097fca4
@@ -143,7 +143,7 @@ class SingleMachineBatchSystem(BatchSystemSupport): startTime = time.time() # Time job is started popen = None statusCode = None - forkWorker = not (self.debugWorker and "_toil_worker" not in jobCommand) + forkWorker = not (self.debugWorker and "_toil_worker" in jobCommand) if forkWorker: with self.popenLock: popen = subprocess.Popen(jobCommand,
src/toil/batchSystems/singleMachine.py
ReplaceText(target=' in ' @(146,61)->(146,69))
class SingleMachineBatchSystem(BatchSystemSupport): startTime = time.time() # Time job is started popen = None statusCode = None forkWorker = not (self.debugWorker and "_toil_worker" not in jobCommand) if forkWorker: with self.popenLock: popen = subprocess.Popen(jobCommand,
class SingleMachineBatchSystem(BatchSystemSupport): startTime = time.time() # Time job is started popen = None statusCode = None forkWorker = not (self.debugWorker and "_toil_worker" in jobCommand) if forkWorker: with self.popenLock: popen = subprocess.Popen(jobCommand,
828
https://:@github.com/ionox0/toil.git
f5054aca67788f976c72b1e3f42334afe3a1dfaa
@@ -138,7 +138,7 @@ class AbstractProvisioner(with_metaclass(ABCMeta, object)): self._spotBidsMap[nodeType] = bid else: self.nodeTypes.append(nodeTypeStr) - self.nodeShapes.append(self.getNodeShape(nodeType, preemptable=False)) + self.nodeShapes.append(self.getNodeShape(nodeTypeStr, preemptable=False)) @staticmethod def retryPredicate(e):
src/toil/provisioners/abstractProvisioner.py
ReplaceText(target='nodeTypeStr' @(141,57)->(141,65))
class AbstractProvisioner(with_metaclass(ABCMeta, object)): self._spotBidsMap[nodeType] = bid else: self.nodeTypes.append(nodeTypeStr) self.nodeShapes.append(self.getNodeShape(nodeType, preemptable=False)) @staticmethod def retryPredicate(e):
class AbstractProvisioner(with_metaclass(ABCMeta, object)): self._spotBidsMap[nodeType] = bid else: self.nodeTypes.append(nodeTypeStr) self.nodeShapes.append(self.getNodeShape(nodeTypeStr, preemptable=False)) @staticmethod def retryPredicate(e):
829
https://:@github.com/ionox0/toil.git
59f716aed0363adba211e7c1f18ebe3802edb636
@@ -108,7 +108,7 @@ class GridEngineBatchSystem(AbstractGridEngineBatchSystem): if not self.boss.config.manualMemArgs: # for UGE instead of SGE; see #2309 reqline += ['vf=' + memStr, 'h_vmem=' + memStr] - elif self.boss.config.manualMemArgs and sgeArgs: + elif self.boss.config.manualMemArgs and not sgeArgs: raise ValueError("--manualMemArgs set to True, but TOIL_GRIDGENGINE_ARGS is not set." "Please set TOIL_GRIDGENGINE_ARGS to specify memory allocation for " "your system. Default adds the arguments: vf=<mem> h_vmem=<mem> "
src/toil/batchSystems/gridengine.py
ReplaceText(target='not ' @(111,56)->(111,56))
class GridEngineBatchSystem(AbstractGridEngineBatchSystem): if not self.boss.config.manualMemArgs: # for UGE instead of SGE; see #2309 reqline += ['vf=' + memStr, 'h_vmem=' + memStr] elif self.boss.config.manualMemArgs and sgeArgs: raise ValueError("--manualMemArgs set to True, but TOIL_GRIDGENGINE_ARGS is not set." "Please set TOIL_GRIDGENGINE_ARGS to specify memory allocation for " "your system. Default adds the arguments: vf=<mem> h_vmem=<mem> "
class GridEngineBatchSystem(AbstractGridEngineBatchSystem): if not self.boss.config.manualMemArgs: # for UGE instead of SGE; see #2309 reqline += ['vf=' + memStr, 'h_vmem=' + memStr] elif self.boss.config.manualMemArgs and not sgeArgs: raise ValueError("--manualMemArgs set to True, but TOIL_GRIDGENGINE_ARGS is not set." "Please set TOIL_GRIDGENGINE_ARGS to specify memory allocation for " "your system. Default adds the arguments: vf=<mem> h_vmem=<mem> "
830
https://:@github.com/tkf/railgun.git
f49414f39ca7c6bbbd5e000d45db46b77a4e514d
@@ -64,7 +64,7 @@ DATA_TEST_CDEC_PARSE = [ ("int a[i] =2", dict_cdec_parse('int', 'a', tuple('i'), 1, 2)), ("int a[i][j]=3", dict_cdec_parse('int', 'a', tuple('ij'), 2, 3)), ('num_i = 10', dict_cdec_parse('int', 'num_i', default=10)), - (cmem('obj', DmyCDT), dict_cdec_parse(DmyCDT, 'obj', valtype='object')), + (cmem(DmyCDT, 'obj'), dict_cdec_parse(DmyCDT, 'obj', valtype='object')), ]
tests/test_cdata.py
ArgSwap(idxs=0<->1 @(67,5)->(67,9))
DATA_TEST_CDEC_PARSE = [ ("int a[i] =2", dict_cdec_parse('int', 'a', tuple('i'), 1, 2)), ("int a[i][j]=3", dict_cdec_parse('int', 'a', tuple('ij'), 2, 3)), ('num_i = 10', dict_cdec_parse('int', 'num_i', default=10)), (cmem('obj', DmyCDT), dict_cdec_parse(DmyCDT, 'obj', valtype='object')), ]
DATA_TEST_CDEC_PARSE = [ ("int a[i] =2", dict_cdec_parse('int', 'a', tuple('i'), 1, 2)), ("int a[i][j]=3", dict_cdec_parse('int', 'a', tuple('ij'), 2, 3)), ('num_i = 10', dict_cdec_parse('int', 'num_i', default=10)), (cmem(DmyCDT, 'obj'), dict_cdec_parse(DmyCDT, 'obj', valtype='object')), ]
831
https://:@github.com/hubzero/hublib.git
8921395b8feda7af6d59ed2133dd1a4068ae9803
@@ -486,7 +486,7 @@ def copy_files(errnum, etime, toolname, runName): else: # nonparametric run. Results are in current working directory. # Use the timestamp to copy all newer files to the cacheName. - if errnum > 0: + if errnum == 0: files = os.listdir('.') for f in files: if os.path.getmtime(f) > self.start_time:
hublib/ui/submit.py
ReplaceText(target='==' @(489,18)->(489,19))
def copy_files(errnum, etime, toolname, runName): else: # nonparametric run. Results are in current working directory. # Use the timestamp to copy all newer files to the cacheName. if errnum > 0: files = os.listdir('.') for f in files: if os.path.getmtime(f) > self.start_time:
def copy_files(errnum, etime, toolname, runName): else: # nonparametric run. Results are in current working directory. # Use the timestamp to copy all newer files to the cacheName. if errnum == 0: files = os.listdir('.') for f in files: if os.path.getmtime(f) > self.start_time:
832
https://:@github.com/tjstretchalot/ignite-simple.git
944f32fd5421cb01c0c75be5959277e1a1a2c741
@@ -240,7 +240,7 @@ class ParamsTaskQueue(disp.TaskQueue, ps.SweepListener): # remove from params_to_task # add to in_progress # return built disp.Task - if cores > self.sweep_cores and self.sweeps: + if cores >= self.sweep_cores and self.sweeps: swp: ParamsTask = self.sweeps.pop() del self.params_to_sweeps_ind[swp.params] self._len -= 1
ignite_simple/gen_sweep/sweeper.py
ReplaceText(target='>=' @(243,17)->(243,18))
class ParamsTaskQueue(disp.TaskQueue, ps.SweepListener): # remove from params_to_task # add to in_progress # return built disp.Task if cores > self.sweep_cores and self.sweeps: swp: ParamsTask = self.sweeps.pop() del self.params_to_sweeps_ind[swp.params] self._len -= 1
class ParamsTaskQueue(disp.TaskQueue, ps.SweepListener): # remove from params_to_task # add to in_progress # return built disp.Task if cores >= self.sweep_cores and self.sweeps: swp: ParamsTask = self.sweeps.pop() del self.params_to_sweeps_ind[swp.params] self._len -= 1
833
https://:@github.com/jessemyers/lxmlbind.git
35bb1bbc635ea9dca813dffe25165b14576c4a00
@@ -75,7 +75,7 @@ class Base(object): child = etree.SubElement(parent, head) else: return None - return self.search(child, tail, create) if tail else child + return self.search(tail, child, create) if tail else child def __str__(self): """
lxmlbind/base.py
ArgSwap(idxs=0<->1 @(78,15)->(78,26))
class Base(object): child = etree.SubElement(parent, head) else: return None return self.search(child, tail, create) if tail else child def __str__(self): """
class Base(object): child = etree.SubElement(parent, head) else: return None return self.search(tail, child, create) if tail else child def __str__(self): """
834
https://:@github.com/dieterv/etk.docking.git
f7e423f2e5d2a68183238c2d1eb39dcb4cf7ec30
@@ -521,7 +521,7 @@ class DockPaned(gtk.Container): width = max(width, w) height += h # Store the minimum weight for usage in do_size_allocate - item.min_size = w + item.min_size = h # Add handles if self._orientation == gtk.ORIENTATION_HORIZONTAL:
lib/etk/docking/dockpaned.py
ReplaceText(target='h' @(524,32)->(524,33))
class DockPaned(gtk.Container): width = max(width, w) height += h # Store the minimum weight for usage in do_size_allocate item.min_size = w # Add handles if self._orientation == gtk.ORIENTATION_HORIZONTAL:
class DockPaned(gtk.Container): width = max(width, w) height += h # Store the minimum weight for usage in do_size_allocate item.min_size = h # Add handles if self._orientation == gtk.ORIENTATION_HORIZONTAL:
835
https://:@github.com/isnok/asciidrumming.git
66e35371a4b96774d931516f7131e19ccf6e2c43
@@ -42,7 +42,7 @@ def assemble_pieces(phrases, pieces): def clean_phrases(phrases): for phrase in phrases.values(): phrase['pattern'] = ''.join(phrase['pattern']).replace(' ', '') - phrase['length'] = (len(phrase['pattern']) // phrase['beat']) + bool(len(phrase['pattern']) % phrase['beat']) + phrase['length'] = (len(phrase['pattern']) / phrase['beat']) + bool(len(phrase['pattern']) % phrase['beat']) def assemble_phrases(config):
asciidrumming/assemble.py
ReplaceText(target='/' @(45,51)->(45,53))
def assemble_pieces(phrases, pieces): def clean_phrases(phrases): for phrase in phrases.values(): phrase['pattern'] = ''.join(phrase['pattern']).replace(' ', '') phrase['length'] = (len(phrase['pattern']) // phrase['beat']) + bool(len(phrase['pattern']) % phrase['beat']) def assemble_phrases(config):
def assemble_pieces(phrases, pieces): def clean_phrases(phrases): for phrase in phrases.values(): phrase['pattern'] = ''.join(phrase['pattern']).replace(' ', '') phrase['length'] = (len(phrase['pattern']) / phrase['beat']) + bool(len(phrase['pattern']) % phrase['beat']) def assemble_phrases(config):
836
https://:@github.com/FedericoCinus/WoMG.git
75addc91cd520c8e7ebf8975ceaf24b79b2cf49f
@@ -390,7 +390,7 @@ class LDA(TLTTopicModel): saved_model_fname = str(hash(saved_model))+'.model' if os.path.exists(saved_model_fname): #lda_model = pickle.load(os.path.abspath(saved_model)) - lda_model = gensim.models.ldamodel.LdaModel.load(os.path.abspath(saved_model)) + lda_model = gensim.models.ldamodel.LdaModel.load(os.path.abspath(saved_model_fname)) else: lda_model = gensim.models.ldamodel.LdaModel(corpus=corpus, id2word=self.dictionary,
src/test_version/topic/lda.py
ReplaceText(target='saved_model_fname' @(393,77)->(393,88))
class LDA(TLTTopicModel): saved_model_fname = str(hash(saved_model))+'.model' if os.path.exists(saved_model_fname): #lda_model = pickle.load(os.path.abspath(saved_model)) lda_model = gensim.models.ldamodel.LdaModel.load(os.path.abspath(saved_model)) else: lda_model = gensim.models.ldamodel.LdaModel(corpus=corpus, id2word=self.dictionary,
class LDA(TLTTopicModel): saved_model_fname = str(hash(saved_model))+'.model' if os.path.exists(saved_model_fname): #lda_model = pickle.load(os.path.abspath(saved_model)) lda_model = gensim.models.ldamodel.LdaModel.load(os.path.abspath(saved_model_fname)) else: lda_model = gensim.models.ldamodel.LdaModel(corpus=corpus, id2word=self.dictionary,
837
https://:@github.com/waikato-datamining/wai-annotations.git
1c1e29a44b802c45c83af7278b1637aa98855a68
@@ -63,4 +63,4 @@ def get_all_sharded_filenames(filename: str) -> Tuple[str, ...]: if shards == 0: raise ValueError(f"{filename} is not a shard filename") - return tuple(format_sharded_filename(base_name, shards, i) for i in range(index)) + return tuple(format_sharded_filename(base_name, shards, i) for i in range(shards))
src/wai/annotations/tf/utils/_sharded_filenames.py
ReplaceText(target='shards' @(66,78)->(66,83))
def get_all_sharded_filenames(filename: str) -> Tuple[str, ...]: if shards == 0: raise ValueError(f"{filename} is not a shard filename") return tuple(format_sharded_filename(base_name, shards, i) for i in range(index))
def get_all_sharded_filenames(filename: str) -> Tuple[str, ...]: if shards == 0: raise ValueError(f"{filename} is not a shard filename") return tuple(format_sharded_filename(base_name, shards, i) for i in range(shards))
838
https://:@github.com/broundal/Pytolemaic.git
c407b7b2410fe04641d0261ca209a2e57a728ded
@@ -220,7 +220,7 @@ class SensitivityAnalysis(): vulnerability_report = self._vulnerability_report( shuffled_sensitivity=self.shuffled_sensitivity, missing_sensitivity=self.missing_sensitivity, - shuffled_sensitivity_stats=missing_stats_report) + shuffled_sensitivity_stats=shuffle_stats_report) return SensitivityFullReport( shuffle_report=self.shuffled_sensitivity,
pytolemaic/analysis_logic/model_analysis/sensitivity/sensitivity.py
ReplaceText(target='shuffle_stats_report' @(223,39)->(223,59))
class SensitivityAnalysis(): vulnerability_report = self._vulnerability_report( shuffled_sensitivity=self.shuffled_sensitivity, missing_sensitivity=self.missing_sensitivity, shuffled_sensitivity_stats=missing_stats_report) return SensitivityFullReport( shuffle_report=self.shuffled_sensitivity,
class SensitivityAnalysis(): vulnerability_report = self._vulnerability_report( shuffled_sensitivity=self.shuffled_sensitivity, missing_sensitivity=self.missing_sensitivity, shuffled_sensitivity_stats=shuffle_stats_report) return SensitivityFullReport( shuffle_report=self.shuffled_sensitivity,
839
https://:@github.com/damoti/shipmaster.git
f81614acecb36da1343ec9e114679b06f0d8b0ba
@@ -94,7 +94,7 @@ class LayerBase: # Only tag image if container was built successfully. repository, tag = self.image_name, None if ':' in repository: - repository, tag = tag.split(':') + repository, tag = repository.split(':') conf = client.create_container_config(self.image_name, cmd, working_dir=APP_PATH) client.commit(container, repository=repository, tag=tag, conf=conf) client.remove_container(container)
shipmaster/base/builder.py
ReplaceText(target='repository' @(97,34)->(97,37))
class LayerBase: # Only tag image if container was built successfully. repository, tag = self.image_name, None if ':' in repository: repository, tag = tag.split(':') conf = client.create_container_config(self.image_name, cmd, working_dir=APP_PATH) client.commit(container, repository=repository, tag=tag, conf=conf) client.remove_container(container)
class LayerBase: # Only tag image if container was built successfully. repository, tag = self.image_name, None if ':' in repository: repository, tag = repository.split(':') conf = client.create_container_config(self.image_name, cmd, working_dir=APP_PATH) client.commit(container, repository=repository, tag=tag, conf=conf) client.remove_container(container)
840
https://:@github.com/ldo/pybidi.git
ef93e8fdd0220fb446238060c7b4f30422251943
@@ -1352,7 +1352,7 @@ def remove_bidi_marks(string, positions_to_this = None, position_from_this_list if position_from_this_list != None : c_position_from_this_list = seq_to_ct(position_from_this_list, FRIBIDI.StrIndex) else : - position_from_this_list = None + c_position_from_this_list = None #end if if embedding_levels != None : c_embedding_levels = seq_to_ct(embedding_levels, FRIBIDI.StrIndex)
fribidi.py
ReplaceText(target='c_position_from_this_list' @(1355,8)->(1355,31))
def remove_bidi_marks(string, positions_to_this = None, position_from_this_list if position_from_this_list != None : c_position_from_this_list = seq_to_ct(position_from_this_list, FRIBIDI.StrIndex) else : position_from_this_list = None #end if if embedding_levels != None : c_embedding_levels = seq_to_ct(embedding_levels, FRIBIDI.StrIndex)
def remove_bidi_marks(string, positions_to_this = None, position_from_this_list if position_from_this_list != None : c_position_from_this_list = seq_to_ct(position_from_this_list, FRIBIDI.StrIndex) else : c_position_from_this_list = None #end if if embedding_levels != None : c_embedding_levels = seq_to_ct(embedding_levels, FRIBIDI.StrIndex)
841
https://:@github.com/wbolster/cardinality.git
cfc729bc337f8d796d7bfe11a64ef3a3d516575b
@@ -68,7 +68,7 @@ def between(min, max, iterable): """ if min < 0: raise ValueError("'min' must be positive (or zero)") - if min < 0: + if max < 0: raise ValueError("'max' must be positive (or zero)") if min > max: raise ValueError("'max' must be greater or equal than 'min'")
cardinality.py
ReplaceText(target='max' @(71,7)->(71,10))
def between(min, max, iterable): """ if min < 0: raise ValueError("'min' must be positive (or zero)") if min < 0: raise ValueError("'max' must be positive (or zero)") if min > max: raise ValueError("'max' must be greater or equal than 'min'")
def between(min, max, iterable): """ if min < 0: raise ValueError("'min' must be positive (or zero)") if max < 0: raise ValueError("'max' must be positive (or zero)") if min > max: raise ValueError("'max' must be greater or equal than 'min'")
842
https://:@github.com/zongzhenh/pymysql-pool.git
66b07cdf844554245cf209a72de89bd17133269c
@@ -169,7 +169,7 @@ class Pool(object): if self.ping_check: now = int(time()) timeout = now - if isinstance(int, self.ping_check): + if isinstance(self.ping_check, int): timeout = timeout - self.ping_check if not hasattr(c, '__ping_check_timestamp'): c.__ping_check_timestamp = now
pymysqlpool/pool.py
ArgSwap(idxs=0<->1 @(172,15)->(172,25))
class Pool(object): if self.ping_check: now = int(time()) timeout = now if isinstance(int, self.ping_check): timeout = timeout - self.ping_check if not hasattr(c, '__ping_check_timestamp'): c.__ping_check_timestamp = now
class Pool(object): if self.ping_check: now = int(time()) timeout = now if isinstance(self.ping_check, int): timeout = timeout - self.ping_check if not hasattr(c, '__ping_check_timestamp'): c.__ping_check_timestamp = now
843
https://:@github.com/kanzure/python-vba-wrapper.git
98edf03512c8f7960f10fa615867269cdf0821d6
@@ -315,7 +315,7 @@ class VBA(object): # 29 registers buf = (ctypes.c_int32 * self.register_count)() buf[:] = registers - self._vba.set_registers(registers) + self._vba.set_registers(buf) def _get_max_save_size(self): return self._vba.get_max_save_size()
vba_wrapper/core.py
ReplaceText(target='buf' @(318,32)->(318,41))
class VBA(object): # 29 registers buf = (ctypes.c_int32 * self.register_count)() buf[:] = registers self._vba.set_registers(registers) def _get_max_save_size(self): return self._vba.get_max_save_size()
class VBA(object): # 29 registers buf = (ctypes.c_int32 * self.register_count)() buf[:] = registers self._vba.set_registers(buf) def _get_max_save_size(self): return self._vba.get_max_save_size()
844
https://:@github.com/severb/flowy.git
82cbab85d672880b7c3db69b5e1707e90da7686e
@@ -191,7 +191,7 @@ class DecisionPoller(object): elif e_type == 'StartChildWorkflowExecutionFailed': SCWEFEA = 'startChildWorkflowExecutionFailedEventAttributes' id = _subworkflow_id(e[SCWEFEA]['workflowId']) - reason = e[SCWEIEA]['cause'] + reason = e[SCWEFEA]['cause'] errors[id] = reason elif e_type == 'TimerStarted': id = e['timerStartedEventAttributes']['timerId']
flowy/swf/poller.py
ReplaceText(target='SCWEFEA' @(194,27)->(194,34))
class DecisionPoller(object): elif e_type == 'StartChildWorkflowExecutionFailed': SCWEFEA = 'startChildWorkflowExecutionFailedEventAttributes' id = _subworkflow_id(e[SCWEFEA]['workflowId']) reason = e[SCWEIEA]['cause'] errors[id] = reason elif e_type == 'TimerStarted': id = e['timerStartedEventAttributes']['timerId']
class DecisionPoller(object): elif e_type == 'StartChildWorkflowExecutionFailed': SCWEFEA = 'startChildWorkflowExecutionFailedEventAttributes' id = _subworkflow_id(e[SCWEFEA]['workflowId']) reason = e[SCWEFEA]['cause'] errors[id] = reason elif e_type == 'TimerStarted': id = e['timerStartedEventAttributes']['timerId']
845
https://:@github.com/brthor/codetransformer-py2.git
7c327683df810265d01a995d2704c9f8218b0ef7
@@ -141,7 +141,7 @@ class Instruction(InstructionMeta._marker, metaclass=InstructionMeta): 'little', ) - yield cls(arg) + yield instr(arg) @classmethod def from_opcode(cls, opcode):
codetransformer/instructions.py
ReplaceText(target='instr' @(144,18)->(144,21))
class Instruction(InstructionMeta._marker, metaclass=InstructionMeta): 'little', ) yield cls(arg) @classmethod def from_opcode(cls, opcode):
class Instruction(InstructionMeta._marker, metaclass=InstructionMeta): 'little', ) yield instr(arg) @classmethod def from_opcode(cls, opcode):
846
https://:@github.com/bacadra/bacadra.git
74346189dfe3cac99821fd0a32ddeb3290886c4d
@@ -1256,7 +1256,7 @@ class texme(TeXM, object, metaclass=texmemeta): elif type(inherit)==str: self.store(inherit, tex) elif inherit is False: - self.data = [tex + '\n%'] + self.data += [tex + '\n%'] if self.echo: print('[texme.' + name + ']\n' + tex)
bacadra/pinky/texme/texme.py
ReplaceText(target='+=' @(1259,22)->(1259,23))
class texme(TeXM, object, metaclass=texmemeta): elif type(inherit)==str: self.store(inherit, tex) elif inherit is False: self.data = [tex + '\n%'] if self.echo: print('[texme.' + name + ']\n' + tex)
class texme(TeXM, object, metaclass=texmemeta): elif type(inherit)==str: self.store(inherit, tex) elif inherit is False: self.data += [tex + '\n%'] if self.echo: print('[texme.' + name + ']\n' + tex)
847
https://:@github.com/bacadra/bacadra.git
3e4cd62c6a0769ea886aaae2104aeb5dff18dbea
@@ -1131,7 +1131,7 @@ class texme(metaclass=texmemeta): return self.add( submodule = _name1, - code = code, + code = tex, inherit = inherit, echo = echo, )
bacadra/pinky/texme/texme.py
ReplaceText(target='tex' @(1134,24)->(1134,28))
class texme(metaclass=texmemeta): return self.add( submodule = _name1, code = code, inherit = inherit, echo = echo, )
class texme(metaclass=texmemeta): return self.add( submodule = _name1, code = tex, inherit = inherit, echo = echo, )
848
https://:@github.com/ulif/diceware-list.git
744223382d439315c8195d3ddff170e3542d80f0
@@ -92,4 +92,4 @@ def local_android_download(request, monkeypatch, tmpdir): monkeypatch.setattr( "diceware_list.libwordlist.AndroidWordList.base_url", fake_base_url) - return dictfile + return tmpdir
tests/conftest.py
ReplaceText(target='tmpdir' @(95,11)->(95,19))
def local_android_download(request, monkeypatch, tmpdir): monkeypatch.setattr( "diceware_list.libwordlist.AndroidWordList.base_url", fake_base_url) return dictfile
def local_android_download(request, monkeypatch, tmpdir): monkeypatch.setattr( "diceware_list.libwordlist.AndroidWordList.base_url", fake_base_url) return tmpdir
849
https://:@github.com/AeGean-Studio/pysolr_aio.git
056f4e2d4284fc740bcc7862fabab673ea1a439b
@@ -747,7 +747,7 @@ class Solr(object): m = force_unicode(m) end_time = time.time() - self.log.debug("Built add request of %s docs in %0.2f seconds.", len(docs), end_time - start_time) + self.log.debug("Built add request of %s docs in %0.2f seconds.", len(message), end_time - start_time) return self._update(m, commit=commit, waitFlush=waitFlush, waitSearcher=waitSearcher) def delete(self, id=None, q=None, commit=True, waitFlush=None, waitSearcher=None):
pysolr.py
ReplaceText(target='message' @(750,77)->(750,81))
class Solr(object): m = force_unicode(m) end_time = time.time() self.log.debug("Built add request of %s docs in %0.2f seconds.", len(docs), end_time - start_time) return self._update(m, commit=commit, waitFlush=waitFlush, waitSearcher=waitSearcher) def delete(self, id=None, q=None, commit=True, waitFlush=None, waitSearcher=None):
class Solr(object): m = force_unicode(m) end_time = time.time() self.log.debug("Built add request of %s docs in %0.2f seconds.", len(message), end_time - start_time) return self._update(m, commit=commit, waitFlush=waitFlush, waitSearcher=waitSearcher) def delete(self, id=None, q=None, commit=True, waitFlush=None, waitSearcher=None):
850
https://:@bitbucket.org/nidusfr/grapheekdb.git
4f9ca53b0ea54eb7b531290818619a0a2082750f
@@ -49,7 +49,7 @@ class Optimizer(object): def get_kind_ids(self, txn, kind): ENTITY_COUNTER = METADATA_VERTEX_COUNTER if kind == KIND_VERTEX else METADATA_EDGE_COUNTER METADATA_ID_LIST_PREFIX = METADATA_VERTEX_ID_LIST_PREFIX if kind == KIND_VERTEX else METADATA_EDGE_ID_LIST_PREFIX - limit = int(self._graph._get(None, ENTITY_COUNTER)) / CHUNK_SIZE + limit = int(self._graph._get(None, ENTITY_COUNTER)) // CHUNK_SIZE keys = [build_key(METADATA_ID_LIST_PREFIX, i) for i in range(0, limit + 1)] list_entity_ids = self._graph._bulk_get_lst(txn, keys) for entity_ids in list_entity_ids:
grapheekdb/backends/data/optimizer.py
ReplaceText(target='//' @(52,60)->(52,61))
class Optimizer(object): def get_kind_ids(self, txn, kind): ENTITY_COUNTER = METADATA_VERTEX_COUNTER if kind == KIND_VERTEX else METADATA_EDGE_COUNTER METADATA_ID_LIST_PREFIX = METADATA_VERTEX_ID_LIST_PREFIX if kind == KIND_VERTEX else METADATA_EDGE_ID_LIST_PREFIX limit = int(self._graph._get(None, ENTITY_COUNTER)) / CHUNK_SIZE keys = [build_key(METADATA_ID_LIST_PREFIX, i) for i in range(0, limit + 1)] list_entity_ids = self._graph._bulk_get_lst(txn, keys) for entity_ids in list_entity_ids:
class Optimizer(object): def get_kind_ids(self, txn, kind): ENTITY_COUNTER = METADATA_VERTEX_COUNTER if kind == KIND_VERTEX else METADATA_EDGE_COUNTER METADATA_ID_LIST_PREFIX = METADATA_VERTEX_ID_LIST_PREFIX if kind == KIND_VERTEX else METADATA_EDGE_ID_LIST_PREFIX limit = int(self._graph._get(None, ENTITY_COUNTER)) // CHUNK_SIZE keys = [build_key(METADATA_ID_LIST_PREFIX, i) for i in range(0, limit + 1)] list_entity_ids = self._graph._bulk_get_lst(txn, keys) for entity_ids in list_entity_ids:
851
https://:@github.com/AnonSolutions/django-indy-community.git
274dcaf28dd3b433f5205f8bef13063a09047fa7
@@ -559,7 +559,7 @@ def send_claims_for_proof_request(wallet, connection, my_conversation, credentia except: raise - return proof_data + return my_conversation ######################################################################
indy_community_demo/indy_community/agent_utils.py
ReplaceText(target='my_conversation' @(562,11)->(562,21))
def send_claims_for_proof_request(wallet, connection, my_conversation, credentia except: raise return proof_data ######################################################################
def send_claims_for_proof_request(wallet, connection, my_conversation, credentia except: raise return my_conversation ######################################################################
852
https://:@github.com/AnonSolutions/django-indy-community.git
7c1d4634416f608a4886444e51691036e6ba5e70
@@ -336,7 +336,7 @@ def check_connection_status(wallet, connection, initialize_vcx=True): my_connection.status = return_state my_connection.save() - check_connection_callback(connection, prev_status) + check_connection_callback(my_connection, prev_status) except: raise finally:
indy_community_demo/indy_community/agent_utils.py
ReplaceText(target='my_connection' @(339,34)->(339,44))
def check_connection_status(wallet, connection, initialize_vcx=True): my_connection.status = return_state my_connection.save() check_connection_callback(connection, prev_status) except: raise finally:
def check_connection_status(wallet, connection, initialize_vcx=True): my_connection.status = return_state my_connection.save() check_connection_callback(my_connection, prev_status) except: raise finally:
853
https://:@github.com/jisson/django-simple-domain.git
794a491f075a8d43fbc12f18aa6ace33a590bc86
@@ -48,7 +48,7 @@ class DjangoSimpleSiteConfig(AppConfig): if SIMPLE_DOMAIN_ENABLED is True """ std_logger.info("Checking if module should be enabled...") - return simple_site_settings.ENABLED and simple_site_utils.is_item_in_list_a_in_list_b( + return simple_site_settings.ENABLED and not simple_site_utils.is_item_in_list_a_in_list_b( simple_site_settings.DEACTIVATING_COMMANDS, sys.argv )
django_simple_domain/app.py
ReplaceText(target='not ' @(51,48)->(51,48))
class DjangoSimpleSiteConfig(AppConfig): if SIMPLE_DOMAIN_ENABLED is True """ std_logger.info("Checking if module should be enabled...") return simple_site_settings.ENABLED and simple_site_utils.is_item_in_list_a_in_list_b( simple_site_settings.DEACTIVATING_COMMANDS, sys.argv )
class DjangoSimpleSiteConfig(AppConfig): if SIMPLE_DOMAIN_ENABLED is True """ std_logger.info("Checking if module should be enabled...") return simple_site_settings.ENABLED and not simple_site_utils.is_item_in_list_a_in_list_b( simple_site_settings.DEACTIVATING_COMMANDS, sys.argv )
854
https://:@github.com/pkumza/libradar.git
5c3676fd09674f198c20b17d4462e0b6e528bc30
@@ -318,7 +318,7 @@ class Tree(object): # JSON support utg_lib_obj = dict() # untagged library object utg_lib_obj["Package"] = node.pn - utg_lib_obj["Standard Package"] = u + utg_lib_obj["Standard Package"] = a utg_lib_obj["Library"] = "Unknown" utg_lib_obj["Popularity"] = int(c) utg_lib_obj["Weight"] = node.weight
LibRadar/dex_tree.py
ReplaceText(target='a' @(321,42)->(321,43))
class Tree(object): # JSON support utg_lib_obj = dict() # untagged library object utg_lib_obj["Package"] = node.pn utg_lib_obj["Standard Package"] = u utg_lib_obj["Library"] = "Unknown" utg_lib_obj["Popularity"] = int(c) utg_lib_obj["Weight"] = node.weight
class Tree(object): # JSON support utg_lib_obj = dict() # untagged library object utg_lib_obj["Package"] = node.pn utg_lib_obj["Standard Package"] = a utg_lib_obj["Library"] = "Unknown" utg_lib_obj["Popularity"] = int(c) utg_lib_obj["Weight"] = node.weight
855
https://:@github.com/moremoban/gease.git
492b0e368076bf3f48eafce06e4eb960d9188360
@@ -83,5 +83,5 @@ def which_org_has(repo): org_repo = Repo(org_info['repos_url']) for arepo in org_repo.get_all_repos(): if repo == arepo['name']: - return arepo['login'] + return org_info['login'] return None
gease/release.py
ReplaceText(target='org_info' @(86,23)->(86,28))
def which_org_has(repo): org_repo = Repo(org_info['repos_url']) for arepo in org_repo.get_all_repos(): if repo == arepo['name']: return arepo['login'] return None
def which_org_has(repo): org_repo = Repo(org_info['repos_url']) for arepo in org_repo.get_all_repos(): if repo == arepo['name']: return org_info['login'] return None
856
https://:@github.com/eflynch/dspy.git
27d706cdf4e25c5750560d957e34319cfd51e145
@@ -21,7 +21,7 @@ def rechannel(buf, in_channels, out_channels): in_channel = (in_channel + 1) % in_channels elif out_channels > in_channels: out_channel = 0 - for in_channel in range(out_channels): + for in_channel in range(in_channels): output[out_channel::out_channels] += buf[in_channel::in_channels] out_channel = (out_channel + 1) % out_channels
dspy/lib.py
ReplaceText(target='in_channels' @(24,32)->(24,44))
def rechannel(buf, in_channels, out_channels): in_channel = (in_channel + 1) % in_channels elif out_channels > in_channels: out_channel = 0 for in_channel in range(out_channels): output[out_channel::out_channels] += buf[in_channel::in_channels] out_channel = (out_channel + 1) % out_channels
def rechannel(buf, in_channels, out_channels): in_channel = (in_channel + 1) % in_channels elif out_channels > in_channels: out_channel = 0 for in_channel in range(in_channels): output[out_channel::out_channels] += buf[in_channel::in_channels] out_channel = (out_channel + 1) % out_channels
857
https://:@gitlab.com/AmosEgel/smuthi.git
3de13ee6ce0382c2657d94d62dab15c6bc0dca0f
@@ -168,7 +168,7 @@ class PiecewiseFieldExpansion(FieldExpansion): else: pfe_sum.expansion_list.append(fex) if not added: - pfe_sum.expansion_list.append(fex) + pfe_sum.expansion_list.append(other) return pfe_sum
smuthi/field_expansion.py
ReplaceText(target='other' @(171,46)->(171,49))
class PiecewiseFieldExpansion(FieldExpansion): else: pfe_sum.expansion_list.append(fex) if not added: pfe_sum.expansion_list.append(fex) return pfe_sum
class PiecewiseFieldExpansion(FieldExpansion): else: pfe_sum.expansion_list.append(fex) if not added: pfe_sum.expansion_list.append(other) return pfe_sum
858
https://:@github.com/inkeye/cyscore.git
f7b659afcfa4dce73a8915e8fccd916a68248d83
@@ -40,4 +40,4 @@ class Score: "--nodisplays", orcname, sconame]) - return fname + return outname
cyscore/score.py
ReplaceText(target='outname' @(43,15)->(43,20))
class Score: "--nodisplays", orcname, sconame]) return fname
class Score: "--nodisplays", orcname, sconame]) return outname
859
https://:@github.com/DMSC-Instrument-Data/lewis.git
b2797261ac0f3ab9f02dc67f28db1620732d3f0b
@@ -101,7 +101,7 @@ class TestAdapterCollection(unittest.TestCase): collection.device = 'foo' self.assertEqual(mock_adapter1.device, 'foo') - self.assertEqual(mock_adapter1.device, 'foo') + self.assertEqual(mock_adapter2.device, 'foo') mock_adapter3 = MagicMock() mock_adapter3.device = 'other'
test/test_core_adapters.py
ReplaceText(target='mock_adapter2' @(104,25)->(104,38))
class TestAdapterCollection(unittest.TestCase): collection.device = 'foo' self.assertEqual(mock_adapter1.device, 'foo') self.assertEqual(mock_adapter1.device, 'foo') mock_adapter3 = MagicMock() mock_adapter3.device = 'other'
class TestAdapterCollection(unittest.TestCase): collection.device = 'foo' self.assertEqual(mock_adapter1.device, 'foo') self.assertEqual(mock_adapter2.device, 'foo') mock_adapter3 = MagicMock() mock_adapter3.device = 'other'
860
https://:@github.com/DMSC-Instrument-Data/lewis.git
50911fba54fc6f9b4099191c7c81f52f66a6ba47
@@ -82,7 +82,7 @@ class ExposedObject(object): exposed_members = members if members else self._public_members() exclude = list(exclude or []) - if not exclude_inherited: + if exclude_inherited: for base in inspect.getmro(type(obj))[1:]: exclude += dir(base)
lewis/core/control_server.py
ReplaceText(target='' @(85,11)->(85,15))
class ExposedObject(object): exposed_members = members if members else self._public_members() exclude = list(exclude or []) if not exclude_inherited: for base in inspect.getmro(type(obj))[1:]: exclude += dir(base)
class ExposedObject(object): exposed_members = members if members else self._public_members() exclude = list(exclude or []) if exclude_inherited: for base in inspect.getmro(type(obj))[1:]: exclude += dir(base)
861
https://:@github.com/fafhrd91/player.git
8bffaa237459e1d9a79c8ba6cd01a482f72ce2f8
@@ -41,7 +41,7 @@ def add_layer(cfg, layer, name='', path='', description=''): layers.insert(0, intr) cfg.action(discr, introspectables=(intr,)) - log.info("Add layer: %s path:%s"%(layer, directory)) + log.info("Add layer: %s path:%s"%(layer, path)) def add_layers(cfg, name='', path='', description=''):
player/layer.py
ReplaceText(target='path' @(44,45)->(44,54))
def add_layer(cfg, layer, name='', path='', description=''): layers.insert(0, intr) cfg.action(discr, introspectables=(intr,)) log.info("Add layer: %s path:%s"%(layer, directory)) def add_layers(cfg, name='', path='', description=''):
def add_layer(cfg, layer, name='', path='', description=''): layers.insert(0, intr) cfg.action(discr, introspectables=(intr,)) log.info("Add layer: %s path:%s"%(layer, path)) def add_layers(cfg, name='', path='', description=''):
862
https://:@github.com/clumsyme/ziroom_watcher.git
91498a31498833d88401888b7cfc7c32cffb7464
@@ -58,7 +58,7 @@ class Watcher: self.info_url, headers=self.headers) info = json.loads(response.text) status = info['data']['status'] - if status == 'tzpzz': + if status != 'tzpzz': self.sendmail('房源状态已更新', '状态更新了') else: raise NotDoneError(status)
ziroom_watcher.py
ReplaceText(target='!=' @(61,18)->(61,20))
class Watcher: self.info_url, headers=self.headers) info = json.loads(response.text) status = info['data']['status'] if status == 'tzpzz': self.sendmail('房源状态已更新', '状态更新了') else: raise NotDoneError(status)
class Watcher: self.info_url, headers=self.headers) info = json.loads(response.text) status = info['data']['status'] if status != 'tzpzz': self.sendmail('房源状态已更新', '状态更新了') else: raise NotDoneError(status)
863
https://:@github.com/kovpas/itc.cli.git
ed0f4e9946469673f2d75ca713ade2b69ac19c25
@@ -244,7 +244,7 @@ class ITCServerParser(BaseParser): for ratingTr in appRatingTable: inputs = ratingTr.xpath('.//input') - if len(inputs) != 2: + if len(inputs) < 2: continue appRating = {'name': inputs[0].attrib['name'], 'ratings': []} for inpt in inputs:
itc/parsers/serverparser.py
ReplaceText(target='<' @(247,27)->(247,29))
class ITCServerParser(BaseParser): for ratingTr in appRatingTable: inputs = ratingTr.xpath('.//input') if len(inputs) != 2: continue appRating = {'name': inputs[0].attrib['name'], 'ratings': []} for inpt in inputs:
class ITCServerParser(BaseParser): for ratingTr in appRatingTable: inputs = ratingTr.xpath('.//input') if len(inputs) < 2: continue appRating = {'name': inputs[0].attrib['name'], 'ratings': []} for inpt in inputs:
864
https://:@github.com/seebye/media-layer.git
e906f0a4648683f27aa99a5a549020064052edd5
@@ -170,7 +170,7 @@ class ForcedCoverImageScaler(DistortImageScaler, OffsetImageScaler): width: int, height: int): width, height = self.calculate_resolution(image, width, height) image_width, image_height = image.width, image.height - if image_width < image_height: + if image_width > image_height: image_height = int(image_height * width / image_width) image_width = width else:
ueberzug/scaling.py
ReplaceText(target='>' @(173,23)->(173,24))
class ForcedCoverImageScaler(DistortImageScaler, OffsetImageScaler): width: int, height: int): width, height = self.calculate_resolution(image, width, height) image_width, image_height = image.width, image.height if image_width < image_height: image_height = int(image_height * width / image_width) image_width = width else:
class ForcedCoverImageScaler(DistortImageScaler, OffsetImageScaler): width: int, height: int): width, height = self.calculate_resolution(image, width, height) image_width, image_height = image.width, image.height if image_width > image_height: image_height = int(image_height * width / image_width) image_width = width else:
865
https://:@github.com/lvapeab/multimodal_keras_wrapper.git
965b970084e59e945036df9cc7cd836d052cb279
@@ -1286,7 +1286,7 @@ class Dataset(object): elif(type_out == 'binary'): y = np.array(y).astype(np.uint8) elif(type_out == 'text'): - y = self.loadText(y, self.vocabulary[id_out], self.max_text_len[id_out], self.text_offset[id_in]) + y = self.loadText(y, self.vocabulary[id_out], self.max_text_len[id_out], self.text_offset[id_out]) #if max_len == 0: y_aux = np.zeros(list(y.shape)+[self.n_classes_text[id_out]]).astype(np.uint8) for idx in range(y.shape[0]):
keras_wrapper/dataset.py
ReplaceText(target='id_out' @(1289,110)->(1289,115))
class Dataset(object): elif(type_out == 'binary'): y = np.array(y).astype(np.uint8) elif(type_out == 'text'): y = self.loadText(y, self.vocabulary[id_out], self.max_text_len[id_out], self.text_offset[id_in]) #if max_len == 0: y_aux = np.zeros(list(y.shape)+[self.n_classes_text[id_out]]).astype(np.uint8) for idx in range(y.shape[0]):
class Dataset(object): elif(type_out == 'binary'): y = np.array(y).astype(np.uint8) elif(type_out == 'text'): y = self.loadText(y, self.vocabulary[id_out], self.max_text_len[id_out], self.text_offset[id_out]) #if max_len == 0: y_aux = np.zeros(list(y.shape)+[self.n_classes_text[id_out]]).astype(np.uint8) for idx in range(y.shape[0]):
866
https://:@github.com/mic159/komodo.git
9b5fcd652ace81e3ee53bedec1eadc1189b08efc
@@ -94,7 +94,7 @@ class Server(object): scheduler = Scheduler() if not self.without_checks: for name, server in servers.items(): - scheduler.register(name, server) + scheduler.register(server, name) checks_thread = threading.Thread(target=self.start_checks, args=(scheduler, self.thread_stopper, )) checks_thread.daemon = True
python_dashing/server/server.py
ArgSwap(idxs=0<->1 @(97,20)->(97,38))
class Server(object): scheduler = Scheduler() if not self.without_checks: for name, server in servers.items(): scheduler.register(name, server) checks_thread = threading.Thread(target=self.start_checks, args=(scheduler, self.thread_stopper, )) checks_thread.daemon = True
class Server(object): scheduler = Scheduler() if not self.without_checks: for name, server in servers.items(): scheduler.register(server, name) checks_thread = threading.Thread(target=self.start_checks, args=(scheduler, self.thread_stopper, )) checks_thread.daemon = True
867
https://:@github.com/mic159/komodo.git
10d4ba791b59d5873a561d58ee8a096ad104a581
@@ -32,7 +32,7 @@ class JsonDataStore(object): def save(self): with open(self.location, 'wb') as fle: - return json.dump(fle, self.data) + return json.dump(self.data, fle) def set(self, prefix, key, value): self.data[prefix][key] = value
dashmat/datastore.py
ArgSwap(idxs=0<->1 @(35,19)->(35,28))
class JsonDataStore(object): def save(self): with open(self.location, 'wb') as fle: return json.dump(fle, self.data) def set(self, prefix, key, value): self.data[prefix][key] = value
class JsonDataStore(object): def save(self): with open(self.location, 'wb') as fle: return json.dump(self.data, fle) def set(self, prefix, key, value): self.data[prefix][key] = value
868
https://:@github.com/krvss/graph-talk.git
f91b24ea9c154f6659faecc806f562ef1ee448e3
@@ -103,6 +103,6 @@ def get_callable(c): def tuples(*args): res = () for arg in args: - res += tuple(arg) if is_list(args) else (arg, ) + res += tuple(arg) if is_list(arg) else (arg, ) return res
utils.py
ReplaceText(target='arg' @(106,37)->(106,41))
def get_callable(c): def tuples(*args): res = () for arg in args: res += tuple(arg) if is_list(args) else (arg, ) return res
def get_callable(c): def tuples(*args): res = () for arg in args: res += tuple(arg) if is_list(arg) else (arg, ) return res
869
https://:@github.com/reichlab/zoltpy.git
8653aa031af8e1dfd2a33e63eb25b642ab719acd
@@ -70,7 +70,7 @@ def upload_forecast(project_name, model_name, timezero_date, forecast_csv_file): print('* working with', model) # upload a new forecast - upload_file_job = model.upload_forecast(timezero_date, forecast_csv_file) + upload_file_job = model.upload_forecast(forecast_csv_file, timezero_date) busy_poll_upload_file_job(upload_file_job) # get the new forecast from the upload_file_job by parsing the generic 'output_json' field
zoltpy/functions.py
ArgSwap(idxs=0<->1 @(73,22)->(73,43))
def upload_forecast(project_name, model_name, timezero_date, forecast_csv_file): print('* working with', model) # upload a new forecast upload_file_job = model.upload_forecast(timezero_date, forecast_csv_file) busy_poll_upload_file_job(upload_file_job) # get the new forecast from the upload_file_job by parsing the generic 'output_json' field
def upload_forecast(project_name, model_name, timezero_date, forecast_csv_file): print('* working with', model) # upload a new forecast upload_file_job = model.upload_forecast(forecast_csv_file, timezero_date) busy_poll_upload_file_job(upload_file_job) # get the new forecast from the upload_file_job by parsing the generic 'output_json' field
870
https://:@github.com/awbirdsall/popmodel.git
918e2083f541a4c85145fea3f007ed646bc60d19
@@ -34,7 +34,7 @@ def k_solved(hpar, par): def k_sweep(hpar, par): par_sweep = deepcopy(par) par_sweep['sweep']['dosweep'] = True - k_sweep = pm.KineticsRun(hpar, **par) + k_sweep = pm.KineticsRun(hpar, **par_sweep) k_sweep.solveode() return k_sweep
tests/test_main.py
ReplaceText(target='par_sweep' @(37,37)->(37,40))
def k_solved(hpar, par): def k_sweep(hpar, par): par_sweep = deepcopy(par) par_sweep['sweep']['dosweep'] = True k_sweep = pm.KineticsRun(hpar, **par) k_sweep.solveode() return k_sweep
def k_solved(hpar, par): def k_sweep(hpar, par): par_sweep = deepcopy(par) par_sweep['sweep']['dosweep'] = True k_sweep = pm.KineticsRun(hpar, **par_sweep) k_sweep.solveode() return k_sweep
871
https://:@github.com/jezeniel/jellyfish-wheel.git
22e790ad922b0122ad40293d594d32e2819d6387
@@ -422,7 +422,7 @@ def metaphone(s): elif next == 'h' and nextnext and nextnext not in 'aeiou': i += 1 elif c == 'h': - if i == 0 or next in 'aeiou' or s[i-1] in 'aeiou': + if i == 0 or next in 'aeiou' or s[i-1] not in 'aeiou': result.append('h') elif c == 'k': if i == 0 or s[i-1] != 'c':
jellyfish/_jellyfish.py
ReplaceText(target=' not in ' @(425,50)->(425,54))
def metaphone(s): elif next == 'h' and nextnext and nextnext not in 'aeiou': i += 1 elif c == 'h': if i == 0 or next in 'aeiou' or s[i-1] in 'aeiou': result.append('h') elif c == 'k': if i == 0 or s[i-1] != 'c':
def metaphone(s): elif next == 'h' and nextnext and nextnext not in 'aeiou': i += 1 elif c == 'h': if i == 0 or next in 'aeiou' or s[i-1] not in 'aeiou': result.append('h') elif c == 'k': if i == 0 or s[i-1] != 'c':
872
https://:@github.com/tourbillonpy/tourbillon-log.git
cee5b1f59349b7b95b7d9c406b36eab609ef386e
@@ -27,7 +27,7 @@ def get_logfile_metrics(agent): db_config['duration'], db_config['replication'], db_config['name']) - logger.info('database "%s" created successfully', config['name']) + logger.info('database "%s" created successfully', db_config['name']) except: pass
tourbillon/log/log.py
ReplaceText(target='db_config' @(30,58)->(30,64))
def get_logfile_metrics(agent): db_config['duration'], db_config['replication'], db_config['name']) logger.info('database "%s" created successfully', config['name']) except: pass
def get_logfile_metrics(agent): db_config['duration'], db_config['replication'], db_config['name']) logger.info('database "%s" created successfully', db_config['name']) except: pass
873
https://:@github.com/OriHoch/ckanext-upload_via_email.git
9f12f3b51b7a1ac8db665a673f8cca571a840529
@@ -33,7 +33,7 @@ def get_sender_organization_id(from_email, to_email, allowed_senders, config): default_sender_organization_id = config.get('default_sender_organization_id') if default_sender_to_address and default_sender_organization_id: default_sender_to_address = default_sender_to_address.lower().strip() - if is_email_match(from_email, default_sender_to_address): + if is_email_match(to_email, default_sender_to_address): return default_sender_organization_id return None
ckanext/upload_via_email/pipelines/download_messages.py
ReplaceText(target='to_email' @(36,30)->(36,40))
def get_sender_organization_id(from_email, to_email, allowed_senders, config): default_sender_organization_id = config.get('default_sender_organization_id') if default_sender_to_address and default_sender_organization_id: default_sender_to_address = default_sender_to_address.lower().strip() if is_email_match(from_email, default_sender_to_address): return default_sender_organization_id return None
def get_sender_organization_id(from_email, to_email, allowed_senders, config): default_sender_organization_id = config.get('default_sender_organization_id') if default_sender_to_address and default_sender_organization_id: default_sender_to_address = default_sender_to_address.lower().strip() if is_email_match(to_email, default_sender_to_address): return default_sender_organization_id return None
874
https://:@github.com/portfors-lab/sparkle.git
5be995abef4100cdce6e2feb5cee977476140a84
@@ -62,7 +62,7 @@ class ProtocolRunner(ListAcquisitionRunner): self.avg_buffer[irep,:] = response if irep == self.nreps -1: avg_response = self.avg_buffer.mean(axis=0) - self.datafile.append(self.current_dataset_name, response) + self.datafile.append(self.current_dataset_name, avg_response) self.avg_buffer = np.zeros_like(self.avg_buffer) else: self.datafile.append(self.current_dataset_name, response)
sparkle/run/protocol_runner.py
ReplaceText(target='avg_response' @(65,68)->(65,76))
class ProtocolRunner(ListAcquisitionRunner): self.avg_buffer[irep,:] = response if irep == self.nreps -1: avg_response = self.avg_buffer.mean(axis=0) self.datafile.append(self.current_dataset_name, response) self.avg_buffer = np.zeros_like(self.avg_buffer) else: self.datafile.append(self.current_dataset_name, response)
class ProtocolRunner(ListAcquisitionRunner): self.avg_buffer[irep,:] = response if irep == self.nreps -1: avg_response = self.avg_buffer.mean(axis=0) self.datafile.append(self.current_dataset_name, avg_response) self.avg_buffer = np.zeros_like(self.avg_buffer) else: self.datafile.append(self.current_dataset_name, response)
875
https://:@github.com/a1ezzz/wasp-general.git
5795c4cb45c70e120a94c0dd3598762505131b29
@@ -59,7 +59,7 @@ def test_abstract(): pytest.raises(NotImplementedError, WSignalSourceProto.remove_callback, None, 'signal', C()) pytest.raises(TypeError, WSignalCallbackProto) - pytest.raises(NotImplementedError, WSignalCallbackProto.__call__, None, 'signal', S(), 1) + pytest.raises(NotImplementedError, WSignalCallbackProto.__call__, None, S(), 'signal', 1) pytest.raises(TypeError, WSignalProxyProto.ProxiedMessageProto) pytest.raises(NotImplementedError, WSignalProxyProto.ProxiedMessageProto.is_weak, None)
tests/wasp_general_signals_proto_test.py
ArgSwap(idxs=3<->4 @(62,1)->(62,14))
def test_abstract(): pytest.raises(NotImplementedError, WSignalSourceProto.remove_callback, None, 'signal', C()) pytest.raises(TypeError, WSignalCallbackProto) pytest.raises(NotImplementedError, WSignalCallbackProto.__call__, None, 'signal', S(), 1) pytest.raises(TypeError, WSignalProxyProto.ProxiedMessageProto) pytest.raises(NotImplementedError, WSignalProxyProto.ProxiedMessageProto.is_weak, None)
def test_abstract(): pytest.raises(NotImplementedError, WSignalSourceProto.remove_callback, None, 'signal', C()) pytest.raises(TypeError, WSignalCallbackProto) pytest.raises(NotImplementedError, WSignalCallbackProto.__call__, None, S(), 'signal', 1) pytest.raises(TypeError, WSignalProxyProto.ProxiedMessageProto) pytest.raises(NotImplementedError, WSignalProxyProto.ProxiedMessageProto.is_weak, None)
876
https://:@github.com/drasmuss/hessianfree.git
883099364fb68ec67b538fffa804a3783d23ad02
@@ -80,7 +80,7 @@ class CrossEntropy(LossFunction): class ClassificationError(LossFunction): @output_loss def loss(self, output, targets): - return (np.argmax(output, axis=-1) == + return (np.argmax(output, axis=-1) != np.argmax(np.nan_to_num(targets), axis=-1)) # note: not defining d_loss or d2_loss; classification error should only
hessianfree/loss_funcs.py
ReplaceText(target='!=' @(83,43)->(83,45))
class CrossEntropy(LossFunction): class ClassificationError(LossFunction): @output_loss def loss(self, output, targets): return (np.argmax(output, axis=-1) == np.argmax(np.nan_to_num(targets), axis=-1)) # note: not defining d_loss or d2_loss; classification error should only
class CrossEntropy(LossFunction): class ClassificationError(LossFunction): @output_loss def loss(self, output, targets): return (np.argmax(output, axis=-1) != np.argmax(np.nan_to_num(targets), axis=-1)) # note: not defining d_loss or d2_loss; classification error should only
877
https://:@github.com/GalakVince/dermoscopic_symmetry.git
cbad346417689f3a698035f0393fef710ae6dedd
@@ -338,7 +338,7 @@ def classifierTrainer(maxLeafNodes, data=None, data_backup_file='patchesDataSet/ clf: The fitted classifier. acc: The accuracy score of the classifier """ - if data is not None: + if data is None: data = pd.read_csv(f"{package_path()}/data/patchesDataSet/{data_backup_file}.csv", index_col=False) features = list(data) del features[0]
dermoscopic_symmetry/classifier_feeder.py
ReplaceText(target=' is ' @(341,11)->(341,19))
def classifierTrainer(maxLeafNodes, data=None, data_backup_file='patchesDataSet/ clf: The fitted classifier. acc: The accuracy score of the classifier """ if data is not None: data = pd.read_csv(f"{package_path()}/data/patchesDataSet/{data_backup_file}.csv", index_col=False) features = list(data) del features[0]
def classifierTrainer(maxLeafNodes, data=None, data_backup_file='patchesDataSet/ clf: The fitted classifier. acc: The accuracy score of the classifier """ if data is None: data = pd.read_csv(f"{package_path()}/data/patchesDataSet/{data_backup_file}.csv", index_col=False) features = list(data) del features[0]
878
https://:@github.com/interputed/grapi.git
8174ec31bc0572ec562f147a2becc65498774d56
@@ -47,7 +47,7 @@ class Grapi: keys = kwargs.keys() endpoints.check(keys, self._endpoint) - if date_step: + if not date_step: return self._methods(method.lower())(**kwargs) else:
grapi/grapi.py
ReplaceText(target='not ' @(50,11)->(50,11))
class Grapi: keys = kwargs.keys() endpoints.check(keys, self._endpoint) if date_step: return self._methods(method.lower())(**kwargs) else:
class Grapi: keys = kwargs.keys() endpoints.check(keys, self._endpoint) if not date_step: return self._methods(method.lower())(**kwargs) else:
879
https://:@github.com/chen0040/pycompressor.git
273f7e1592566e82b62afa250d29c08b0bbcacdd
@@ -19,7 +19,7 @@ class Node(object): def char_at(text, i): - if len(text) - 1 <= i: + if len(text) - 1 < i: return -1 return ord(text[i])
pycompressor/huffman.py
ReplaceText(target='<' @(22,21)->(22,23))
class Node(object): def char_at(text, i): if len(text) - 1 <= i: return -1 return ord(text[i])
class Node(object): def char_at(text, i): if len(text) - 1 < i: return -1 return ord(text[i])
880
https://:@github.com/scattering-central/saxskit.git
e696feaed288160cf35a34d4f43de44e4af4b90f
@@ -114,7 +114,7 @@ def train_classification_models(data,hyper_parameters_search=False): print('--> average unweighted f1: {}, accuracy: {}'.format(f1_score,acc)) else: print('--> {} untrainable- default value: {}'.format(model_id,model.default_val)) - cls_models['main_classifiers'][struct_nm] = model + cls_models['main_classifiers'][model_id] = model # There are 2**n possible outcomes for n binary classifiers. # For the (2**n)-1 non-null outcomes, a second classifier is used,
xrsdkit/models/train.py
ReplaceText(target='model_id' @(117,39)->(117,48))
def train_classification_models(data,hyper_parameters_search=False): print('--> average unweighted f1: {}, accuracy: {}'.format(f1_score,acc)) else: print('--> {} untrainable- default value: {}'.format(model_id,model.default_val)) cls_models['main_classifiers'][struct_nm] = model # There are 2**n possible outcomes for n binary classifiers. # For the (2**n)-1 non-null outcomes, a second classifier is used,
def train_classification_models(data,hyper_parameters_search=False): print('--> average unweighted f1: {}, accuracy: {}'.format(f1_score,acc)) else: print('--> {} untrainable- default value: {}'.format(model_id,model.default_val)) cls_models['main_classifiers'][model_id] = model # There are 2**n possible outcomes for n binary classifiers. # For the (2**n)-1 non-null outcomes, a second classifier is used,
881
https://:@github.com/scattering-central/saxskit.git
4e384359736c5191648f3a5971ccc7d34fb98b27
@@ -413,7 +413,7 @@ class XRSDFitGUI(object): # TODO: toggles for hyperparam selection? feature selection? dataset_dir = self._vars['io_control']['dataset_dir'].get() output_dir = self._vars['io_control']['output_dir'].get() - model_config_path = os.path.join(dataset_dir,'model_config.yml') + model_config_path = os.path.join(output_dir,'model_config.yml') self._print_to_listbox(display,'LOADING DATASET FROM: {}'.format(dataset_dir)) df, idx_df = read_local_dataset(dataset_dir,downsampling_distance=1., message_callback=partial(self._print_to_listbox,display))
xrsdkit/visualization/gui.py
ReplaceText(target='output_dir' @(416,41)->(416,52))
class XRSDFitGUI(object): # TODO: toggles for hyperparam selection? feature selection? dataset_dir = self._vars['io_control']['dataset_dir'].get() output_dir = self._vars['io_control']['output_dir'].get() model_config_path = os.path.join(dataset_dir,'model_config.yml') self._print_to_listbox(display,'LOADING DATASET FROM: {}'.format(dataset_dir)) df, idx_df = read_local_dataset(dataset_dir,downsampling_distance=1., message_callback=partial(self._print_to_listbox,display))
class XRSDFitGUI(object): # TODO: toggles for hyperparam selection? feature selection? dataset_dir = self._vars['io_control']['dataset_dir'].get() output_dir = self._vars['io_control']['output_dir'].get() model_config_path = os.path.join(output_dir,'model_config.yml') self._print_to_listbox(display,'LOADING DATASET FROM: {}'.format(dataset_dir)) df, idx_df = read_local_dataset(dataset_dir,downsampling_distance=1., message_callback=partial(self._print_to_listbox,display))
882
https://:@github.com/rohinkumar/correlcalc.git
9a13668ff7ef2396ce22175d3f21599af7279766
@@ -18,7 +18,7 @@ def generate_rand_from_pdf(pdf, x_grid, N): cdf = cdf / cdf[-1] values = np.random.rand(N) value_bins = np.searchsorted(cdf, values) - random_from_cdf, nz = x_grid[value_bins], cdf[value_bins] + random_from_cdf, nz = x_grid[value_bins], pdf[value_bins] return random_from_cdf, nz
correlcalc/genrand.py
ReplaceText(target='pdf' @(21,46)->(21,49))
def generate_rand_from_pdf(pdf, x_grid, N): cdf = cdf / cdf[-1] values = np.random.rand(N) value_bins = np.searchsorted(cdf, values) random_from_cdf, nz = x_grid[value_bins], cdf[value_bins] return random_from_cdf, nz
def generate_rand_from_pdf(pdf, x_grid, N): cdf = cdf / cdf[-1] values = np.random.rand(N) value_bins = np.searchsorted(cdf, values) random_from_cdf, nz = x_grid[value_bins], pdf[value_bins] return random_from_cdf, nz
883
https://:@github.com/deepfield/builder.git
4410007757b043e278d5a9a8deec2ae7cd5da2b6
@@ -489,7 +489,7 @@ class Job(object): str_targets = collections.defaultdict(list) for target_type, targets in targets_dict.iteritems(): for target in targets: - str_dependencies[target_type].append(target.unexpanded_id) + str_targets[target_type].append(target.unexpanded_id) this_dict = {"depends": str_dependencies, "targets": str_targets}
jobs.py
ReplaceText(target='str_targets' @(492,16)->(492,32))
class Job(object): str_targets = collections.defaultdict(list) for target_type, targets in targets_dict.iteritems(): for target in targets: str_dependencies[target_type].append(target.unexpanded_id) this_dict = {"depends": str_dependencies, "targets": str_targets}
class Job(object): str_targets = collections.defaultdict(list) for target_type, targets in targets_dict.iteritems(): for target in targets: str_targets[target_type].append(target.unexpanded_id) this_dict = {"depends": str_dependencies, "targets": str_targets}
884
https://:@github.com/allonkleinlab/scrublet.git
afe329d90fa0bddadccf9e707112f835fe83e46e
@@ -66,7 +66,7 @@ def sparse_zscore(E, gene_mean=None, gene_stdev=None): ''' z-score normalize each column of E ''' if gene_mean is None: - gene_stdev = E.mean(0) + gene_mean = E.mean(0) if gene_stdev is None: gene_stdev = np.sqrt(sparse_var(E)) return sparse_multiply((E - gene_mean).T, 1/gene_stdev).T
src/scrublet/helper_functions.py
ReplaceText(target='gene_mean' @(69,8)->(69,18))
def sparse_zscore(E, gene_mean=None, gene_stdev=None): ''' z-score normalize each column of E ''' if gene_mean is None: gene_stdev = E.mean(0) if gene_stdev is None: gene_stdev = np.sqrt(sparse_var(E)) return sparse_multiply((E - gene_mean).T, 1/gene_stdev).T
def sparse_zscore(E, gene_mean=None, gene_stdev=None): ''' z-score normalize each column of E ''' if gene_mean is None: gene_mean = E.mean(0) if gene_stdev is None: gene_stdev = np.sqrt(sparse_var(E)) return sparse_multiply((E - gene_mean).T, 1/gene_stdev).T
885
https://:@github.com/SheldonYS/Zappa.git
513bfe44f24cee37e8dc720f3a55c8f8d0c73163
@@ -1053,7 +1053,7 @@ class Zappa(object): if count: # We can end up in a situation where we have more resources being created # than anticipated. - if (count - current_resources) >= 0: + if (count - current_resources) > 0: progress.update(count - current_resources) current_resources = count progress.close()
zappa/zappa.py
ReplaceText(target='>' @(1056,51)->(1056,53))
class Zappa(object): if count: # We can end up in a situation where we have more resources being created # than anticipated. if (count - current_resources) >= 0: progress.update(count - current_resources) current_resources = count progress.close()
class Zappa(object): if count: # We can end up in a situation where we have more resources being created # than anticipated. if (count - current_resources) > 0: progress.update(count - current_resources) current_resources = count progress.close()
886
https://:@github.com/moser/pylint.git
4d29d8e4559938bef1eb52b1613d2596aeba395b
@@ -174,7 +174,7 @@ class ClassDiagram(Figure, FilterMixIn): value = value._proxied try: ass_obj = self.object_from_node(value) - self.add_relationship(obj, ass_obj, 'association', name) + self.add_relationship(ass_obj, obj, 'association', name) except KeyError: continue
pyreverse/diagrams.py
ArgSwap(idxs=0<->1 @(177,24)->(177,45))
class ClassDiagram(Figure, FilterMixIn): value = value._proxied try: ass_obj = self.object_from_node(value) self.add_relationship(obj, ass_obj, 'association', name) except KeyError: continue
class ClassDiagram(Figure, FilterMixIn): value = value._proxied try: ass_obj = self.object_from_node(value) self.add_relationship(ass_obj, obj, 'association', name) except KeyError: continue
887
https://:@github.com/moser/pylint.git
2ac99b2a24d44a3b56c035a889f6bd78f4f41b7f
@@ -242,7 +242,7 @@ builtins. Remember that you should avoid to define new builtins when possible.' if node.name.startswith('cb_') or \ node.name.endswith('_cb'): continue - self.add_message('W0613', args=name, node=node) + self.add_message('W0613', args=name, node=stmt) else: self.add_message('W0612', args=name, node=stmt)
checkers/variables.py
ReplaceText(target='stmt' @(245,58)->(245,62))
builtins. Remember that you should avoid to define new builtins when possible.' if node.name.startswith('cb_') or \ node.name.endswith('_cb'): continue self.add_message('W0613', args=name, node=node) else: self.add_message('W0612', args=name, node=stmt)
builtins. Remember that you should avoid to define new builtins when possible.' if node.name.startswith('cb_') or \ node.name.endswith('_cb'): continue self.add_message('W0613', args=name, node=stmt) else: self.add_message('W0612', args=name, node=stmt)
888
https://:@github.com/moser/pylint.git
dbb2f3e65b9a452cf981e1bda318ec019b43bd8c
@@ -714,7 +714,7 @@ class BasicErrorChecker(_BasicChecker): if defined_self is not node and not astroid.are_exclusive(node, defined_self): dummy_variables_rgx = lint_utils.get_global_option( self, 'dummy-variables-rgx', default=None) - if dummy_variables_rgx and dummy_variables_rgx.match(defined_self.name): + if dummy_variables_rgx and dummy_variables_rgx.match(node.name): return self.add_message('function-redefined', node=node, args=(redeftype, defined_self.fromlineno))
pylint/checkers/base.py
ReplaceText(target='node' @(717,65)->(717,77))
class BasicErrorChecker(_BasicChecker): if defined_self is not node and not astroid.are_exclusive(node, defined_self): dummy_variables_rgx = lint_utils.get_global_option( self, 'dummy-variables-rgx', default=None) if dummy_variables_rgx and dummy_variables_rgx.match(defined_self.name): return self.add_message('function-redefined', node=node, args=(redeftype, defined_self.fromlineno))
class BasicErrorChecker(_BasicChecker): if defined_self is not node and not astroid.are_exclusive(node, defined_self): dummy_variables_rgx = lint_utils.get_global_option( self, 'dummy-variables-rgx', default=None) if dummy_variables_rgx and dummy_variables_rgx.match(node.name): return self.add_message('function-redefined', node=node, args=(redeftype, defined_self.fromlineno))
889
https://:@github.com/moser/pylint.git
02688062140cc7ffa9c4f1a1f04c13d6f198a400
@@ -697,7 +697,7 @@ def error_of_type(handler: astroid.ExceptHandler, error_type) -> bool: error_type = (error_type,) # type: ignore expected_errors = {stringify_error(error) for error in error_type} # type: ignore if not handler.type: - return True + return False return handler.catch(expected_errors)
pylint/checkers/utils.py
ReplaceText(target='False' @(700,15)->(700,19))
def error_of_type(handler: astroid.ExceptHandler, error_type) -> bool: error_type = (error_type,) # type: ignore expected_errors = {stringify_error(error) for error in error_type} # type: ignore if not handler.type: return True return handler.catch(expected_errors)
def error_of_type(handler: astroid.ExceptHandler, error_type) -> bool: error_type = (error_type,) # type: ignore expected_errors = {stringify_error(error) for error in error_type} # type: ignore if not handler.type: return False return handler.catch(expected_errors)
890
https://:@github.com/ogrisel/pygbm.git
2e019bfc2b3ea35ec5bd7c3c56301f8fc7a64b60
@@ -426,7 +426,7 @@ def _find_best_bin_to_split_helper(context, feature_idx, histogram, n_samples): context.sum_gradients, context.sum_hessians, context.l2_regularization) - if gain > best_split.gain and gain >= context.min_gain_to_split: + if gain > best_split.gain and gain > context.min_gain_to_split: best_split.gain = gain best_split.feature_idx = feature_idx best_split.bin_idx = bin_idx
pygbm/splitting.py
ReplaceText(target='>' @(429,43)->(429,45))
def _find_best_bin_to_split_helper(context, feature_idx, histogram, n_samples): context.sum_gradients, context.sum_hessians, context.l2_regularization) if gain > best_split.gain and gain >= context.min_gain_to_split: best_split.gain = gain best_split.feature_idx = feature_idx best_split.bin_idx = bin_idx
def _find_best_bin_to_split_helper(context, feature_idx, histogram, n_samples): context.sum_gradients, context.sum_hessians, context.l2_regularization) if gain > best_split.gain and gain > context.min_gain_to_split: best_split.gain = gain best_split.feature_idx = feature_idx best_split.bin_idx = bin_idx
891
https://:@github.com/happz/ducky.git
e56b0f3d47fb108865973efe411e6767d51d72a1
@@ -330,7 +330,7 @@ class MathCoprocessor(ISnapshotable, Coprocessor): i = i32(tos.value).value j = i32(lr.value).value D(' i=%i, j=%i (%s, %s)', i, j, type(i), type(j)) - i /= j + i //= j D(' i=%i (%s)', i, type(i)) tos.value = i
ducky/cpu/coprocessor/math_copro.py
ReplaceText(target='//=' @(333,6)->(333,8))
class MathCoprocessor(ISnapshotable, Coprocessor): i = i32(tos.value).value j = i32(lr.value).value D(' i=%i, j=%i (%s, %s)', i, j, type(i), type(j)) i /= j D(' i=%i (%s)', i, type(i)) tos.value = i
class MathCoprocessor(ISnapshotable, Coprocessor): i = i32(tos.value).value j = i32(lr.value).value D(' i=%i, j=%i (%s, %s)', i, j, type(i), type(j)) i //= j D(' i=%i (%s)', i, type(i)) tos.value = i
892
https://:@github.com/happz/ducky.git
fb1a93bdb8055874aef7077850e784dc90e8295c
@@ -680,7 +680,7 @@ class Machine(ISnapshotable, IMachineWorker): self.hdt.create() - pages = self.memory.alloc_pages(segment = 0x00, count = align(PAGE_SIZE, self.hdt.size()) / PAGE_SIZE) + pages = self.memory.alloc_pages(segment = 0x00, count = align(PAGE_SIZE, self.hdt.size()) // PAGE_SIZE) self.memory.update_pages_flags(pages[0].index, len(pages), 'read', True) self.hdt_address = pages[0].base_address
ducky/machine.py
ReplaceText(target='//' @(683,94)->(683,95))
class Machine(ISnapshotable, IMachineWorker): self.hdt.create() pages = self.memory.alloc_pages(segment = 0x00, count = align(PAGE_SIZE, self.hdt.size()) / PAGE_SIZE) self.memory.update_pages_flags(pages[0].index, len(pages), 'read', True) self.hdt_address = pages[0].base_address
class Machine(ISnapshotable, IMachineWorker): self.hdt.create() pages = self.memory.alloc_pages(segment = 0x00, count = align(PAGE_SIZE, self.hdt.size()) // PAGE_SIZE) self.memory.update_pages_flags(pages[0].index, len(pages), 'read', True) self.hdt_address = pages[0].base_address
893
https://:@github.com/happz/ducky.git
f266fb16a7f0e10e19cba8d17eb15d0d69d3c24a
@@ -134,7 +134,7 @@ def show_pages(logger, state): CPR = 32 - for i in range(0, 256 / CPR): + for i in range(0, 256 // CPR): s = [] t = []
ducky/tools/coredump.py
ReplaceText(target='//' @(137,26)->(137,27))
def show_pages(logger, state): CPR = 32 for i in range(0, 256 / CPR): s = [] t = []
def show_pages(logger, state): CPR = 32 for i in range(0, 256 // CPR): s = [] t = []
894
https://:@github.com/tpircher/quine-mccluskey.git
31df780f3ec5a14d70840e755f23da133a8fc282
@@ -372,7 +372,7 @@ class QuineMcCluskey: # Add the unused terms to the list of marked terms for g in list(groups.values()): - marked |= group - used + marked |= g - used if len(used) == 0: done = True
quine_mccluskey/qm.py
ReplaceText(target='g' @(375,26)->(375,31))
class QuineMcCluskey: # Add the unused terms to the list of marked terms for g in list(groups.values()): marked |= group - used if len(used) == 0: done = True
class QuineMcCluskey: # Add the unused terms to the list of marked terms for g in list(groups.values()): marked |= g - used if len(used) == 0: done = True
895
https://:@github.com/SUSE/azurectl.git
09c62dc074497f862c0a4fee6798ffee14f7e61c
@@ -81,7 +81,7 @@ class Image: service_call = service.add_os_image( label, media_link, name, 'Linux' ) - add_os_image_result = service_call.get_operation_status( + add_os_image_result = service.get_operation_status( service_call.request_id ) status = add_os_image_result.status
azurectl/image.py
ReplaceText(target='service' @(84,34)->(84,46))
class Image: service_call = service.add_os_image( label, media_link, name, 'Linux' ) add_os_image_result = service_call.get_operation_status( service_call.request_id ) status = add_os_image_result.status
class Image: service_call = service.add_os_image( label, media_link, name, 'Linux' ) add_os_image_result = service.get_operation_status( service_call.request_id ) status = add_os_image_result.status
896
https://:@github.com/robocomp/learnbot.git
e0af147e7b8ab58add721fefe425c6ac28183043
@@ -1,6 +1,6 @@ def obstacle_free(lbot, threshold= 200, verbose=False): sonarsValue = lbot.getSonars() - if min(sonarsValue) > threshold: + if min(sonarsValue) < threshold: if verbose: print('No obstacles around Learnbot') return True
learnbot_dsl/functions/perceptual/obstacle_free.py
ReplaceText(target='<' @(3,21)->(3,22))
-1,6 +1,6 @@ def obstacle_free(lbot, threshold= 200, verbose=False): sonarsValue = lbot.getSonars() if min(sonarsValue) > threshold: if verbose: print('No obstacles around Learnbot') return True
-1,6 +1,6 @@ def obstacle_free(lbot, threshold= 200, verbose=False): sonarsValue = lbot.getSonars() if min(sonarsValue) < threshold: if verbose: print('No obstacles around Learnbot') return True
897
https://:@github.com/robocomp/learnbot.git
f43f97e705c26d2e6d1a9fded44ee913dbba6e84
@@ -9,7 +9,7 @@ from learnbot_dsl.learnbotCode.Language import getLanguage import tempfile, uuid, sys def str2hex(text): - if sys.version_info[0]>3: + if sys.version_info[0]>=3: return text.encode('utf-8').hex() else: return str(binascii.hexlify(bytes(text)))
learnbot_dsl/learnbotCode/Button.py
ReplaceText(target='>=' @(12,26)->(12,27))
from learnbot_dsl.learnbotCode.Language import getLanguage import tempfile, uuid, sys def str2hex(text): if sys.version_info[0]>3: return text.encode('utf-8').hex() else: return str(binascii.hexlify(bytes(text)))
from learnbot_dsl.learnbotCode.Language import getLanguage import tempfile, uuid, sys def str2hex(text): if sys.version_info[0]>=3: return text.encode('utf-8').hex() else: return str(binascii.hexlify(bytes(text)))
898
https://:@github.com/robocomp/learnbot.git
b5f0d9f8bc119b2c1f05c2b8c0a8e7e8b40ed8e0
@@ -1,4 +1,4 @@ def look_front(lbot): - lbot.setJointAngle("CAMERA",0) + lbot.setJointAngle(0, "CAMERA")
learnbot_dsl/functions/motor/jointmotor/look_front.py
ArgSwap(idxs=0<->1 @(4,4)->(4,22))
-1,4 +1,4 @@ def look_front(lbot): lbot.setJointAngle("CAMERA",0)
-1,4 +1,4 @@ def look_front(lbot): lbot.setJointAngle(0, "CAMERA")
899
https://:@github.com/robocomp/learnbot.git
b5f0d9f8bc119b2c1f05c2b8c0a8e7e8b40ed8e0
@@ -1,4 +1,4 @@ def setAngleCamera(lbot,angle): - lbot.setJointAngle("CAMERA", angle) + lbot.setJointAngle(angle, "CAMERA")
learnbot_dsl/functions/motor/jointmotor/setAngleCamera.py
ArgSwap(idxs=0<->1 @(4,4)->(4,22))
-1,4 +1,4 @@ def setAngleCamera(lbot,angle): lbot.setJointAngle("CAMERA", angle)
-1,4 +1,4 @@ def setAngleCamera(lbot,angle): lbot.setJointAngle(angle, "CAMERA")