Unnamed: 0
int64
0
2.44k
repo
stringlengths
32
81
hash
stringlengths
40
40
diff
stringlengths
113
1.17k
old_path
stringlengths
5
84
rewrite
stringlengths
34
79
initial_state
stringlengths
75
980
final_state
stringlengths
76
980
1,000
https://:@github.com/plonegovbr/brasil.gov.vlibrasnews.git
7d9a28d0b1c48e2599dd6e87fbc9c277c1ea5bd8
@@ -18,6 +18,6 @@ class VLibrasVideoViewlet(ViewletBase): super(VLibrasVideoViewlet, self).update() self.youtube_url = get_video_url(self.context) self.is_ready = self.youtube_url is not None - self.enabled = self.is_ready and not api.user.is_anonymous() + self.enabled = self.is_ready or not api.user.is_anonymous() if self.is_ready: self.klass = 'ready'
src/brasil/gov/vlibrasvideo/browser/vlibrasvideo.py
ReplaceText(target='or' @(21,37)->(21,40))
class VLibrasVideoViewlet(ViewletBase): super(VLibrasVideoViewlet, self).update() self.youtube_url = get_video_url(self.context) self.is_ready = self.youtube_url is not None self.enabled = self.is_ready and not api.user.is_anonymous() if self.is_ready: self.klass = 'ready'
class VLibrasVideoViewlet(ViewletBase): super(VLibrasVideoViewlet, self).update() self.youtube_url = get_video_url(self.context) self.is_ready = self.youtube_url is not None self.enabled = self.is_ready or not api.user.is_anonymous() if self.is_ready: self.klass = 'ready'
1,001
https://:@github.com/ch3pjw/format_cef.git
91a70b8e6d42edf64d1d4a94a447cb87fa77254e
@@ -85,7 +85,7 @@ def str_sanitiser(regex_str='.*', escape_chars='', min_len=0, max_len=None): s = s.encode('utf-8') s = escape(s) if max_len is None: - if len(s) <= min_len: + if len(s) < min_len: raise ValueError( '{}: String longer than {} characters'.format( debug_name, min_len))
format_cef/cef.py
ReplaceText(target='<' @(88,26)->(88,28))
def str_sanitiser(regex_str='.*', escape_chars='', min_len=0, max_len=None): s = s.encode('utf-8') s = escape(s) if max_len is None: if len(s) <= min_len: raise ValueError( '{}: String longer than {} characters'.format( debug_name, min_len))
def str_sanitiser(regex_str='.*', escape_chars='', min_len=0, max_len=None): s = s.encode('utf-8') s = escape(s) if max_len is None: if len(s) < min_len: raise ValueError( '{}: String longer than {} characters'.format( debug_name, min_len))
1,002
https://:@github.com/hyan15/proxyscrape.git
14c34e8e90643885f1839878b3ac55f4b51bcd29
@@ -238,7 +238,7 @@ class Collector: self._extend_filter(combined_filter_opts, filter_opts) self._refresh_resources(False) - return self._store.get_proxy(filter_opts, self._blacklist) + return self._store.get_proxy(combined_filter_opts, self._blacklist) def remove_proxy(self, proxies): if not _is_iterable(proxies):
proxyscrape/proxyscrape.py
ReplaceText(target='combined_filter_opts' @(241,37)->(241,48))
class Collector: self._extend_filter(combined_filter_opts, filter_opts) self._refresh_resources(False) return self._store.get_proxy(filter_opts, self._blacklist) def remove_proxy(self, proxies): if not _is_iterable(proxies):
class Collector: self._extend_filter(combined_filter_opts, filter_opts) self._refresh_resources(False) return self._store.get_proxy(combined_filter_opts, self._blacklist) def remove_proxy(self, proxies): if not _is_iterable(proxies):
1,003
https://:@github.com/Axilent/Djax.git
6ed7214dd303d0358b6adda49eca86511a12e430
@@ -203,7 +203,7 @@ def sync_content(token=None,content_type_to_sync=None): log.info('Syncing %s.' % content_type_to_sync) try: content_type = content_registry[content_type_to_sync] - sync_content_type(content_type) + sync_content_type(content_type_to_sync) except KeyError: log.error('%s is not in the content registry.' % content_type_to_sync) else:
djax/content.py
ReplaceText(target='content_type_to_sync' @(206,30)->(206,42))
def sync_content(token=None,content_type_to_sync=None): log.info('Syncing %s.' % content_type_to_sync) try: content_type = content_registry[content_type_to_sync] sync_content_type(content_type) except KeyError: log.error('%s is not in the content registry.' % content_type_to_sync) else:
def sync_content(token=None,content_type_to_sync=None): log.info('Syncing %s.' % content_type_to_sync) try: content_type = content_registry[content_type_to_sync] sync_content_type(content_type_to_sync) except KeyError: log.error('%s is not in the content registry.' % content_type_to_sync) else:
1,004
https://:@github.com/Axilent/Djax.git
60b81a3f9e104bdebfe9fa0668da6644f8fd058f
@@ -518,7 +518,7 @@ class ContentManager(Manager): channel_results = self._channel(queryset,channel=channel,profile=profile,basekey=basekey,flavor=flavor,limit=limit,include_endorsements=True) remainder_results = queryset.exclude(pk__in=[item.pk for item in channel_results]) final_results = channel_results + [ContentItemWrapper(item,0) for item in remainder_results] - final_results.sort(cmp=lambda x,y: cmp(x.rlevel,y.rlevel)) + final_results.sort(cmp=lambda x,y: cmp(y.rlevel,x.rlevel)) return final_results class ContentItemWrapper(object):
djax/content.py
ArgSwap(idxs=0<->1 @(521,43)->(521,46))
class ContentManager(Manager): channel_results = self._channel(queryset,channel=channel,profile=profile,basekey=basekey,flavor=flavor,limit=limit,include_endorsements=True) remainder_results = queryset.exclude(pk__in=[item.pk for item in channel_results]) final_results = channel_results + [ContentItemWrapper(item,0) for item in remainder_results] final_results.sort(cmp=lambda x,y: cmp(x.rlevel,y.rlevel)) return final_results class ContentItemWrapper(object):
class ContentManager(Manager): channel_results = self._channel(queryset,channel=channel,profile=profile,basekey=basekey,flavor=flavor,limit=limit,include_endorsements=True) remainder_results = queryset.exclude(pk__in=[item.pk for item in channel_results]) final_results = channel_results + [ContentItemWrapper(item,0) for item in remainder_results] final_results.sort(cmp=lambda x,y: cmp(y.rlevel,x.rlevel)) return final_results class ContentItemWrapper(object):
1,005
https://:@github.com/baverman/supplement.git
c1611d29a5ebe5347ddd5107f9f549f256204780
@@ -148,7 +148,7 @@ def assist(project, source, position, filename): if not ctx: names = get_scope_names(scope, lineno) else: - obj = infer(ctx, scope, position) + obj = infer(ctx, scope, lineno) names = [obj.get_names()] elif ctx_type == 'import': names = (project.get_possible_imports(ctx, filename),)
supplement/assistant.py
ReplaceText(target='lineno' @(151,36)->(151,44))
def assist(project, source, position, filename): if not ctx: names = get_scope_names(scope, lineno) else: obj = infer(ctx, scope, position) names = [obj.get_names()] elif ctx_type == 'import': names = (project.get_possible_imports(ctx, filename),)
def assist(project, source, position, filename): if not ctx: names = get_scope_names(scope, lineno) else: obj = infer(ctx, scope, lineno) names = [obj.get_names()] elif ctx_type == 'import': names = (project.get_possible_imports(ctx, filename),)
1,006
https://:@github.com/baverman/supplement.git
82ec55dee23f33c890ab1d8a0e55e694e00dc0b0
@@ -105,7 +105,7 @@ class CallDB(object): if not args: continue try: - func = scope.eval(func, False) + func = s.eval(func, False) except: continue
supplement/calls.py
ReplaceText(target='s' @(108,27)->(108,32))
class CallDB(object): if not args: continue try: func = scope.eval(func, False) except: continue
class CallDB(object): if not args: continue try: func = s.eval(func, False) except: continue
1,007
https://:@github.com/keiserlab/e3fp.git
5f5bed7e36138fddc8c64108f9d352260f061d55
@@ -332,7 +332,7 @@ def mol_to_sdf(mol, out_file, conf_num=None): writer = rdkit.Chem.SDWriter(fobj) conf_ids = [conf.GetId() for conf in mol.GetConformers()] for i in conf_ids: - if conf_num in {-1, None} and i >= conf_num: + if conf_num not in {-1, None} and i >= conf_num: break writer.write(mol, confId=i) writer.close()
e3fp/conformer/util.py
ReplaceText(target=' not in ' @(335,23)->(335,27))
def mol_to_sdf(mol, out_file, conf_num=None): writer = rdkit.Chem.SDWriter(fobj) conf_ids = [conf.GetId() for conf in mol.GetConformers()] for i in conf_ids: if conf_num in {-1, None} and i >= conf_num: break writer.write(mol, confId=i) writer.close()
def mol_to_sdf(mol, out_file, conf_num=None): writer = rdkit.Chem.SDWriter(fobj) conf_ids = [conf.GetId() for conf in mol.GetConformers()] for i in conf_ids: if conf_num not in {-1, None} and i >= conf_num: break writer.write(mol, confId=i) writer.close()
1,008
https://:@github.com/dbrnz/biosignalml-corelib.git
ad805bbdf30919348b05a8e0b6c7a86081dd93dd
@@ -91,7 +91,7 @@ class HDF5Signal(BSMLSignal): if isinstance(self.clock, UniformClock): yield DataSegment(self.clock[startpos], UniformTimeSeries(data, self.clock.rate)) else: - yield DataSegment(0, TimeSeries(self.clock[startpos: startpos+maxpoints], data)) + yield DataSegment(0, TimeSeries(data, self.clock[startpos: startpos+maxpoints])) startpos += len(data) length -= len(data)
formats/hdf5/__init__.py
ArgSwap(idxs=0<->1 @(94,29)->(94,39))
class HDF5Signal(BSMLSignal): if isinstance(self.clock, UniformClock): yield DataSegment(self.clock[startpos], UniformTimeSeries(data, self.clock.rate)) else: yield DataSegment(0, TimeSeries(self.clock[startpos: startpos+maxpoints], data)) startpos += len(data) length -= len(data)
class HDF5Signal(BSMLSignal): if isinstance(self.clock, UniformClock): yield DataSegment(self.clock[startpos], UniformTimeSeries(data, self.clock.rate)) else: yield DataSegment(0, TimeSeries(data, self.clock[startpos: startpos+maxpoints])) startpos += len(data) length -= len(data)
1,009
https://:@github.com/SOBotics/Redunda-lib-Python.git
6187143b32f4f3c4ca9809c347349be1b537cf30
@@ -83,7 +83,7 @@ class Redunda: try: if filename.endswith (".pickle") or ispickle == True: - dict = eval (filename) + dict = eval (filedata) try: pickle.dump (dict, filename) except pickle.PickleError as perr:
Redunda.py
ReplaceText(target='filedata' @(86,29)->(86,37))
class Redunda: try: if filename.endswith (".pickle") or ispickle == True: dict = eval (filename) try: pickle.dump (dict, filename) except pickle.PickleError as perr:
class Redunda: try: if filename.endswith (".pickle") or ispickle == True: dict = eval (filedata) try: pickle.dump (dict, filename) except pickle.PickleError as perr:
1,010
https://:@github.com/averagehuman/kez.git
e923c7dad180ba4f6d8250074b112b2dd7416329
@@ -20,7 +20,7 @@ class BuildController(object): self.src = src self.dst = dst self.options = options or {} - self.settings = options or {} + self.settings = settings or {} self.logfile = pathjoin(self.dst, 'melba.log') self.status = None self.exc_info = None
melba/builders/base.py
ReplaceText(target='settings' @(23,24)->(23,31))
class BuildController(object): self.src = src self.dst = dst self.options = options or {} self.settings = options or {} self.logfile = pathjoin(self.dst, 'melba.log') self.status = None self.exc_info = None
class BuildController(object): self.src = src self.dst = dst self.options = options or {} self.settings = settings or {} self.logfile = pathjoin(self.dst, 'melba.log') self.status = None self.exc_info = None
1,011
https://:@github.com/fsecada01/TextSpitter.git
c7ca706cc19d2d76fa562c045dd8de7a3bc96643
@@ -22,7 +22,7 @@ def PdfFileRead(file): i += 1 else: pdf_file = open(file, 'rb') - pdf_reader = PyPDF2.PdfFileReader(file) + pdf_reader = PyPDF2.PdfFileReader(pdf_file) while i < pdf_reader.numPages: payload = pdf_reader.getPage(i).extractText().replace('\n', '') text += payload.encode('ascii', 'ignore').decode('unicode_escape')
TextSpitter/core.py
ReplaceText(target='pdf_file' @(25,42)->(25,46))
def PdfFileRead(file): i += 1 else: pdf_file = open(file, 'rb') pdf_reader = PyPDF2.PdfFileReader(file) while i < pdf_reader.numPages: payload = pdf_reader.getPage(i).extractText().replace('\n', '') text += payload.encode('ascii', 'ignore').decode('unicode_escape')
def PdfFileRead(file): i += 1 else: pdf_file = open(file, 'rb') pdf_reader = PyPDF2.PdfFileReader(pdf_file) while i < pdf_reader.numPages: payload = pdf_reader.getPage(i).extractText().replace('\n', '') text += payload.encode('ascii', 'ignore').decode('unicode_escape')
1,012
https://:@github.com/WHenderson/HashDb.git
4544c95147266221be975c3da18341641078575f
@@ -11,7 +11,7 @@ def walk(top, topdown=True, onerror=None, followlinks=False): filename = basename(top) try: - scandir_it = scandir(top) + scandir_it = scandir(dirpath) except OSError as error: onerror(error) return
hashdb2/walk.py
ReplaceText(target='dirpath' @(14,33)->(14,36))
def walk(top, topdown=True, onerror=None, followlinks=False): filename = basename(top) try: scandir_it = scandir(top) except OSError as error: onerror(error) return
def walk(top, topdown=True, onerror=None, followlinks=False): filename = basename(top) try: scandir_it = scandir(dirpath) except OSError as error: onerror(error) return
1,013
https://:@github.com/namuyan/bc4py.git
946e2296029195123c663840d49d921b517853a8
@@ -609,7 +609,7 @@ class TransactionBuilder: def put_unconfirmed(self, tx): assert tx.height is None, 'Not unconfirmed tx {}'.format(tx) - if tx.type not in (C.TX_POW_REWARD, C.TX_POS_REWARD): + if tx.type in (C.TX_POW_REWARD, C.TX_POS_REWARD): return # It is Reword tx elif tx.hash in self.unconfirmed: logging.debug('Already unconfirmed tx. {}'.format(tx))
bc4py/database/builder.py
ReplaceText(target=' in ' @(612,18)->(612,26))
class TransactionBuilder: def put_unconfirmed(self, tx): assert tx.height is None, 'Not unconfirmed tx {}'.format(tx) if tx.type not in (C.TX_POW_REWARD, C.TX_POS_REWARD): return # It is Reword tx elif tx.hash in self.unconfirmed: logging.debug('Already unconfirmed tx. {}'.format(tx))
class TransactionBuilder: def put_unconfirmed(self, tx): assert tx.height is None, 'Not unconfirmed tx {}'.format(tx) if tx.type in (C.TX_POW_REWARD, C.TX_POS_REWARD): return # It is Reword tx elif tx.hash in self.unconfirmed: logging.debug('Already unconfirmed tx. {}'.format(tx))
1,014
https://:@github.com/namuyan/bc4py.git
1d2b8ad52307769c0ac32ab65dadb0d0bf606532
@@ -198,7 +198,7 @@ def contract_signature_check(extra_tx: TX, v: Validator, include_block: Block): raise BlockChainError('Unrelated signature include, reject={}'.format(reject_cks)) elif include_block: # check satisfy require? - if len(accept_cks) != v.require: + if len(accept_cks) >= v.require: raise BlockChainError('Not satisfied require signature. [signed={}, accepted={}, require={}]' .format(signed_cks, accept_cks, v.require)) else:
bc4py/chain/checking/tx_contract.py
ReplaceText(target='>=' @(201,27)->(201,29))
def contract_signature_check(extra_tx: TX, v: Validator, include_block: Block): raise BlockChainError('Unrelated signature include, reject={}'.format(reject_cks)) elif include_block: # check satisfy require? if len(accept_cks) != v.require: raise BlockChainError('Not satisfied require signature. [signed={}, accepted={}, require={}]' .format(signed_cks, accept_cks, v.require)) else:
def contract_signature_check(extra_tx: TX, v: Validator, include_block: Block): raise BlockChainError('Unrelated signature include, reject={}'.format(reject_cks)) elif include_block: # check satisfy require? if len(accept_cks) >= v.require: raise BlockChainError('Not satisfied require signature. [signed={}, accepted={}, require={}]' .format(signed_cks, accept_cks, v.require)) else:
1,015
https://:@github.com/namuyan/bc4py.git
87defc8330c37c6c4a4d32a3187c70d36081a8cf
@@ -96,7 +96,7 @@ def validator_fill_iter(v: Validator, best_block=None, best_chain=None): c_address, address, flag, sig_diff = decode(tx.message) if c_address != v.c_address: continue - index = tx.height * 0xffffffff + block.txs.index(tx) + index = block.height * 0xffffffff + block.txs.index(tx) yield index, flag, address, sig_diff, tx.hash # unconfirmed if best_block is None:
bc4py/database/validator.py
ReplaceText(target='block' @(99,20)->(99,22))
def validator_fill_iter(v: Validator, best_block=None, best_chain=None): c_address, address, flag, sig_diff = decode(tx.message) if c_address != v.c_address: continue index = tx.height * 0xffffffff + block.txs.index(tx) yield index, flag, address, sig_diff, tx.hash # unconfirmed if best_block is None:
def validator_fill_iter(v: Validator, best_block=None, best_chain=None): c_address, address, flag, sig_diff = decode(tx.message) if c_address != v.c_address: continue index = block.height * 0xffffffff + block.txs.index(tx) yield index, flag, address, sig_diff, tx.hash # unconfirmed if best_block is None:
1,016
https://:@github.com/namuyan/bc4py.git
2083e976751938af11f53bff7f9b303a1e16a29d
@@ -147,7 +147,7 @@ async def new_address(request): cur = db.cursor() user_name = request.query.get('account', C.account2name[C.ANT_UNKNOWN]) user_id = read_name2user(user_name, cur) - address = create_new_user_keypair(user_name, cur) + address = create_new_user_keypair(user_id, cur) db.commit() if user_id == C.ANT_CONTRACT: address = convert_address(address, V.BLOCK_CONTRACT_PREFIX)
bc4py/user/api/accountinfo.py
ReplaceText(target='user_id' @(150,42)->(150,51))
async def new_address(request): cur = db.cursor() user_name = request.query.get('account', C.account2name[C.ANT_UNKNOWN]) user_id = read_name2user(user_name, cur) address = create_new_user_keypair(user_name, cur) db.commit() if user_id == C.ANT_CONTRACT: address = convert_address(address, V.BLOCK_CONTRACT_PREFIX)
async def new_address(request): cur = db.cursor() user_name = request.query.get('account', C.account2name[C.ANT_UNKNOWN]) user_id = read_name2user(user_name, cur) address = create_new_user_keypair(user_id, cur) db.commit() if user_id == C.ANT_CONTRACT: address = convert_address(address, V.BLOCK_CONTRACT_PREFIX)
1,017
https://:@github.com/namuyan/bc4py.git
21ac4f74d4d4ec9322e90556e2a4506c5ca1a705
@@ -53,7 +53,7 @@ def check_already_started(): new_pid = os.getpid() with open(pid_path, mode='w') as fp: fp.write(str(new_pid)) - log.info("create new process lock file pid={}".format(pid)) + log.info("create new process lock file pid={}".format(new_pid)) class AESCipher:
bc4py/utils.py
ReplaceText(target='new_pid' @(56,58)->(56,61))
def check_already_started(): new_pid = os.getpid() with open(pid_path, mode='w') as fp: fp.write(str(new_pid)) log.info("create new process lock file pid={}".format(pid)) class AESCipher:
def check_already_started(): new_pid = os.getpid() with open(pid_path, mode='w') as fp: fp.write(str(new_pid)) log.info("create new process lock file pid={}".format(new_pid)) class AESCipher:
1,018
https://:@github.com/agtumulak/dabbiew.git
58b7fdb7ede8e0ae216e5b4088ff6c4d1bcd6f0e
@@ -571,7 +571,7 @@ def run(stdscr, df): left, right, top, bottom, moving_right, moving_down = jump( left, right, top, bottom, rows, cols, found_row, found_col, resizing) if keypress in [ord('g')]: - if not keystroke_history and keystroke_history[-1] == 'g': + if keystroke_history and keystroke_history[-1] == 'g': left, right, top, bottom, moving_right, moving_down = jump( left, right, top, bottom, rows, cols, 0, right, resizing) if keypress in [ord('G')]:
src/dabbiew.py
ReplaceText(target='' @(574,15)->(574,19))
def run(stdscr, df): left, right, top, bottom, moving_right, moving_down = jump( left, right, top, bottom, rows, cols, found_row, found_col, resizing) if keypress in [ord('g')]: if not keystroke_history and keystroke_history[-1] == 'g': left, right, top, bottom, moving_right, moving_down = jump( left, right, top, bottom, rows, cols, 0, right, resizing) if keypress in [ord('G')]:
def run(stdscr, df): left, right, top, bottom, moving_right, moving_down = jump( left, right, top, bottom, rows, cols, found_row, found_col, resizing) if keypress in [ord('g')]: if keystroke_history and keystroke_history[-1] == 'g': left, right, top, bottom, moving_right, moving_down = jump( left, right, top, bottom, rows, cols, 0, right, resizing) if keypress in [ord('G')]:
1,019
https://:@github.com/arquolo/glow.git
d5b0d511214ff7588ca148819c5d3aa7a5b635d8
@@ -20,7 +20,7 @@ def timer(name=''): yield finally: duration = time() - start - if name: + if not name: logging.warning('done in %.4g seconds', duration) else: logging.warning('%s - done in %.4g seconds', name, duration)
ort/debug.py
ReplaceText(target='not ' @(23,11)->(23,11))
def timer(name=''): yield finally: duration = time() - start if name: logging.warning('done in %.4g seconds', duration) else: logging.warning('%s - done in %.4g seconds', name, duration)
def timer(name=''): yield finally: duration = time() - start if not name: logging.warning('done in %.4g seconds', duration) else: logging.warning('%s - done in %.4g seconds', name, duration)
1,020
https://:@github.com/joaduo/smoothtest.git
f70861b38425b6511e05469e4fb2bb6f5ee989b6
@@ -155,7 +155,7 @@ class SmokeCommand(SmoothTestBase): for m in s.get_missing(pkg): pth = m.__file__ if pth.endswith('.pyc'): - pth = f[:-1] + pth = pth[:-1] s.log('Missing test in module %s' % m) s.log(formatPathPrint(pth)) #return results
smoothtest/smoke/SmokeTestDiscover.py
ReplaceText(target='pth' @(158,30)->(158,31))
class SmokeCommand(SmoothTestBase): for m in s.get_missing(pkg): pth = m.__file__ if pth.endswith('.pyc'): pth = f[:-1] s.log('Missing test in module %s' % m) s.log(formatPathPrint(pth)) #return results
class SmokeCommand(SmoothTestBase): for m in s.get_missing(pkg): pth = m.__file__ if pth.endswith('.pyc'): pth = pth[:-1] s.log('Missing test in module %s' % m) s.log(formatPathPrint(pth)) #return results
1,021
https://:@github.com/timmykuo/mitopipeline.git
3888bab80cbcff6a6fbd28c15c1941f1c557d611
@@ -36,7 +36,7 @@ class PipelineBuilder(): #write in the steps requested into the pipeline for step in steps: #if Complete Genomics data, i.e., did split gap then pipeline requires different scripts with shorter reads due to splitting into multiple reads at the gap - if 'split_gap' in steps and step == 'remove_numts': + if 'split_gap' not in steps and step == 'remove_numts': job_name = 'remove_numts_no_split_gap.sh' #step only is name of the step, not the name of the script else:
mitopipeline/pipeline_builder.py
ReplaceText(target=' not in ' @(39,30)->(39,34))
class PipelineBuilder(): #write in the steps requested into the pipeline for step in steps: #if Complete Genomics data, i.e., did split gap then pipeline requires different scripts with shorter reads due to splitting into multiple reads at the gap if 'split_gap' in steps and step == 'remove_numts': job_name = 'remove_numts_no_split_gap.sh' #step only is name of the step, not the name of the script else:
class PipelineBuilder(): #write in the steps requested into the pipeline for step in steps: #if Complete Genomics data, i.e., did split gap then pipeline requires different scripts with shorter reads due to splitting into multiple reads at the gap if 'split_gap' not in steps and step == 'remove_numts': job_name = 'remove_numts_no_split_gap.sh' #step only is name of the step, not the name of the script else:
1,022
https://:@github.com/maqifrnswa/scimpy.git
4bec9f14ba49a033bccbaf1fd26fd1102c44126d
@@ -208,7 +208,7 @@ def calc_impedance(plotwidget, ax_power.set_ylabel('SPL (dB 1W1m)', color='b') # ax_phase.set_ylabel('Phase (degrees)') ax_groupdelay.set_ylabel('Group Delay (ms)', color='r') - ax_groupdelay.set_xlabel('Frequency (Hz)') + ax_power.set_xlabel('Frequency (Hz)') ax_power.set_xscale('log') ax_power.set_xlim([20, 20000]) ax_power.xaxis.set_major_formatter(
scimpy/speakermodel.py
ReplaceText(target='ax_power' @(211,4)->(211,17))
def calc_impedance(plotwidget, ax_power.set_ylabel('SPL (dB 1W1m)', color='b') # ax_phase.set_ylabel('Phase (degrees)') ax_groupdelay.set_ylabel('Group Delay (ms)', color='r') ax_groupdelay.set_xlabel('Frequency (Hz)') ax_power.set_xscale('log') ax_power.set_xlim([20, 20000]) ax_power.xaxis.set_major_formatter(
def calc_impedance(plotwidget, ax_power.set_ylabel('SPL (dB 1W1m)', color='b') # ax_phase.set_ylabel('Phase (degrees)') ax_groupdelay.set_ylabel('Group Delay (ms)', color='r') ax_power.set_xlabel('Frequency (Hz)') ax_power.set_xscale('log') ax_power.set_xlim([20, 20000]) ax_power.xaxis.set_major_formatter(
1,023
https://:@github.com/CS207-Project-Group-9/cs207-FinalProject.git
0a9f635bd112ddcee71faa0560cf7c364f1af02d
@@ -428,7 +428,7 @@ def create_r(vals): ''' if np.array(vals).ndim == 0: return rAD(vals) - elif np.array(vals).ndim > 2: + elif np.array(vals).ndim >= 2: raise ValueError('Input is at most 2D.') else: ADs = [rAD(val) for val in vals]
AutoDiff/AutoDiff.py
ReplaceText(target='>=' @(431,29)->(431,30))
def create_r(vals): ''' if np.array(vals).ndim == 0: return rAD(vals) elif np.array(vals).ndim > 2: raise ValueError('Input is at most 2D.') else: ADs = [rAD(val) for val in vals]
def create_r(vals): ''' if np.array(vals).ndim == 0: return rAD(vals) elif np.array(vals).ndim >= 2: raise ValueError('Input is at most 2D.') else: ADs = [rAD(val) for val in vals]
1,024
https://:@github.com/CS207-Project-Group-9/cs207-FinalProject.git
8157501baf379f85b5f84f4516ac0ba83edfac34
@@ -428,7 +428,7 @@ def create_r(vals): ''' if np.array(vals).ndim == 0: return rAD(vals) - elif np.array(vals).ndim >= 2: + elif np.array(vals).ndim > 2: raise ValueError('Input is at most 2D.') else: ADs = [rAD(val) for val in vals]
AutoDiff/AutoDiff.py
ReplaceText(target='>' @(431,29)->(431,31))
def create_r(vals): ''' if np.array(vals).ndim == 0: return rAD(vals) elif np.array(vals).ndim >= 2: raise ValueError('Input is at most 2D.') else: ADs = [rAD(val) for val in vals]
def create_r(vals): ''' if np.array(vals).ndim == 0: return rAD(vals) elif np.array(vals).ndim > 2: raise ValueError('Input is at most 2D.') else: ADs = [rAD(val) for val in vals]
1,025
https://:@github.com/LucumaTeam/DataQualityHDFS.git
d601249bacb744cadc09cfc9d80e4d5fbec61958
@@ -49,6 +49,6 @@ class Table: return self._is_load_information def load_dataset(self): - if(self._source is None): + if(self._source is not None): self._data_set = self._source.retrieve_dataset() self._is_load_information = True
src/main/Core/Table.py
ReplaceText(target=' is not ' @(52,23)->(52,27))
class Table: return self._is_load_information def load_dataset(self): if(self._source is None): self._data_set = self._source.retrieve_dataset() self._is_load_information = True
class Table: return self._is_load_information def load_dataset(self): if(self._source is not None): self._data_set = self._source.retrieve_dataset() self._is_load_information = True
1,026
https://:@github.com/niklasf/python-asyncdgt.git
836508723a0cfd71fc752538ab59608819184ac6
@@ -52,7 +52,7 @@ def usage(): def main(port_globs): loop = asyncio.get_event_loop() - dgt = asyncdgt.auto_connect(port_globs, loop) + dgt = asyncdgt.auto_connect(loop, port_globs) @dgt.on("connected") def on_connected(port):
asyncdgt/__main__.py
ArgSwap(idxs=0<->1 @(55,10)->(55,31))
def usage(): def main(port_globs): loop = asyncio.get_event_loop() dgt = asyncdgt.auto_connect(port_globs, loop) @dgt.on("connected") def on_connected(port):
def usage(): def main(port_globs): loop = asyncio.get_event_loop() dgt = asyncdgt.auto_connect(loop, port_globs) @dgt.on("connected") def on_connected(port):
1,027
https://:@github.com/hackty/elasticsearch-py.git
18f2ac12310a8dc4e69ee0a262663232fe0fb6ea
@@ -205,7 +205,7 @@ def reindex(client, source_index, target_index, target_client=None, chunk_size=5 :arg scroll: Specify how long a consistent view of the index should be maintained for scrolled search """ - target_client = client if target_client is None else target_index + target_client = client if target_client is None else target_client docs = scan(client, index=source_index, scroll=scroll) def _change_doc_index(hits, index):
elasticsearch/helpers.py
ReplaceText(target='target_client' @(208,57)->(208,69))
def reindex(client, source_index, target_index, target_client=None, chunk_size=5 :arg scroll: Specify how long a consistent view of the index should be maintained for scrolled search """ target_client = client if target_client is None else target_index docs = scan(client, index=source_index, scroll=scroll) def _change_doc_index(hits, index):
def reindex(client, source_index, target_index, target_client=None, chunk_size=5 :arg scroll: Specify how long a consistent view of the index should be maintained for scrolled search """ target_client = client if target_client is None else target_client docs = scan(client, index=source_index, scroll=scroll) def _change_doc_index(hits, index):
1,028
https://:@github.com/ShineyDev/github.py.git
84f7746b34d378990a15a0b7a685e1739b4f9350
@@ -815,7 +815,7 @@ class Fragment(): fields = data.get("fields", list()) for (field) in fields: field = Field.from_dict(field) - fragment.add_field(fields) + fragment.add_field(field) return fragment
github/query/builder.py
ReplaceText(target='field' @(818,31)->(818,37))
class Fragment(): fields = data.get("fields", list()) for (field) in fields: field = Field.from_dict(field) fragment.add_field(fields) return fragment
class Fragment(): fields = data.get("fields", list()) for (field) in fields: field = Field.from_dict(field) fragment.add_field(field) return fragment
1,029
https://:@github.com/ShineyDev/github.py.git
2dfbdd085dada2ec7bf898180514644b617e08b5
@@ -47,7 +47,7 @@ class Topic(Node, Type): for (topic) in data: topics.append(cls(topic)) - return topic + return topics @property def name(self) -> str:
github/objects/topic.py
ReplaceText(target='topics' @(50,19)->(50,24))
class Topic(Node, Type): for (topic) in data: topics.append(cls(topic)) return topic @property def name(self) -> str:
class Topic(Node, Type): for (topic) in data: topics.append(cls(topic)) return topics @property def name(self) -> str:
1,030
https://:@github.com/openbadges/badgecheck.git
a1d552fc5142f268640c949dabdeba309c741486
@@ -62,7 +62,7 @@ def detect_input_type(state, task_meta=None, **options): new_actions.append(set_input_type(detected_type)) if detected_type == 'url': new_actions.append(add_task(FETCH_HTTP_NODE, url=id_url)) - new_actions.append(set_validation_subject(input_value)) + new_actions.append(set_validation_subject(id_url)) elif input_is_jws(input_value): detected_type = 'jws' new_actions.append(set_input_type(detected_type))
badgecheck/tasks/input.py
ReplaceText(target='id_url' @(65,54)->(65,65))
def detect_input_type(state, task_meta=None, **options): new_actions.append(set_input_type(detected_type)) if detected_type == 'url': new_actions.append(add_task(FETCH_HTTP_NODE, url=id_url)) new_actions.append(set_validation_subject(input_value)) elif input_is_jws(input_value): detected_type = 'jws' new_actions.append(set_input_type(detected_type))
def detect_input_type(state, task_meta=None, **options): new_actions.append(set_input_type(detected_type)) if detected_type == 'url': new_actions.append(add_task(FETCH_HTTP_NODE, url=id_url)) new_actions.append(set_validation_subject(id_url)) elif input_is_jws(input_value): detected_type = 'jws' new_actions.append(set_input_type(detected_type))
1,031
https://:@github.com/tungminhphan/street_intersection.git
b400ad94e424cae8ef8eec6d898508b2044dde1e
@@ -78,7 +78,7 @@ def collision_free(object1, object2, car_scale_factor, pedestrian_scale_factor): #the if else statements determines whether the object is pedestrian or not so it can unpack its coordinates and angle orientation, and determines if it should get the vertices of a car or pedestrian #it returns True if no collision has happened, False otherwise object1_vertices, x, y, radius = get_bounding_box(object1, car_scale_factor, pedestrian_scale_factor) - object2_vertices, x2, y2, radius2 = get_bounding_box(object1, car_scale_factor, pedestrian_scale_factor) + object2_vertices, x2, y2, radius2 = get_bounding_box(object2, car_scale_factor, pedestrian_scale_factor) #takes the distance of the centers and compares it to the sum of radius, if the distance is greater then collision not possible if no_collision_by_radius_check(x, y, radius, x2, y2, radius2):
traffic_intersection/prepare/collision_check.py
ReplaceText(target='object2' @(81,57)->(81,64))
def collision_free(object1, object2, car_scale_factor, pedestrian_scale_factor): #the if else statements determines whether the object is pedestrian or not so it can unpack its coordinates and angle orientation, and determines if it should get the vertices of a car or pedestrian #it returns True if no collision has happened, False otherwise object1_vertices, x, y, radius = get_bounding_box(object1, car_scale_factor, pedestrian_scale_factor) object2_vertices, x2, y2, radius2 = get_bounding_box(object1, car_scale_factor, pedestrian_scale_factor) #takes the distance of the centers and compares it to the sum of radius, if the distance is greater then collision not possible if no_collision_by_radius_check(x, y, radius, x2, y2, radius2):
def collision_free(object1, object2, car_scale_factor, pedestrian_scale_factor): #the if else statements determines whether the object is pedestrian or not so it can unpack its coordinates and angle orientation, and determines if it should get the vertices of a car or pedestrian #it returns True if no collision has happened, False otherwise object1_vertices, x, y, radius = get_bounding_box(object1, car_scale_factor, pedestrian_scale_factor) object2_vertices, x2, y2, radius2 = get_bounding_box(object2, car_scale_factor, pedestrian_scale_factor) #takes the distance of the centers and compares it to the sum of radius, if the distance is greater then collision not possible if no_collision_by_radius_check(x, y, radius, x2, y2, radius2):
1,032
https://:@github.com/aclark4life/Parse2Plone.git
49817c3b81b7def16843ae55b30e6145ab21a9a5
@@ -676,7 +676,7 @@ class Parse2Plone(object): else: folder = self.create_folder(parent, utils._remove_ext(obj), _replace_types_map) - self.set_title(page, utils._remove_ext(obj)) + self.set_title(folder, utils._remove_ext(obj)) create_spreadsheets(folder, obj, parent_path, import_dir) _COUNT['files'] += 1 commit()
parse2plone.py
ReplaceText(target='folder' @(679,43)->(679,47))
class Parse2Plone(object): else: folder = self.create_folder(parent, utils._remove_ext(obj), _replace_types_map) self.set_title(page, utils._remove_ext(obj)) create_spreadsheets(folder, obj, parent_path, import_dir) _COUNT['files'] += 1 commit()
class Parse2Plone(object): else: folder = self.create_folder(parent, utils._remove_ext(obj), _replace_types_map) self.set_title(folder, utils._remove_ext(obj)) create_spreadsheets(folder, obj, parent_path, import_dir) _COUNT['files'] += 1 commit()
1,033
https://:@github.com/jjyr/mmr.py.git
373d5799f4682f4ee94aada12ad1558f9c38ea2b
@@ -60,7 +60,7 @@ def get_peaks(mmr_size) -> List[int]: poss.append(pos) while height > 0: height, pos = get_right_peak(height, pos, mmr_size) - if height > 0: + if height >= 0: poss.append(pos) return poss
mmr/mmr.py
ReplaceText(target='>=' @(63,18)->(63,19))
def get_peaks(mmr_size) -> List[int]: poss.append(pos) while height > 0: height, pos = get_right_peak(height, pos, mmr_size) if height > 0: poss.append(pos) return poss
def get_peaks(mmr_size) -> List[int]: poss.append(pos) while height > 0: height, pos = get_right_peak(height, pos, mmr_size) if height >= 0: poss.append(pos) return poss
1,034
https://:@github.com/dtkav/datalake-common.git
4093508498f0fabcfa451d333a141d307fdfd615
@@ -25,7 +25,7 @@ def random_hex(length): def random_interval(): now = datetime.now() start = now - timedelta(days=random.randint(0, 365*3)) - end = start - timedelta(days=random.randint(1, 10)) + end = start + timedelta(days=random.randint(1, 10)) return start.isoformat(), end.isoformat() def random_work_id():
datalake_common/tests/conftest.py
ReplaceText(target='+' @(28,16)->(28,17))
def random_hex(length): def random_interval(): now = datetime.now() start = now - timedelta(days=random.randint(0, 365*3)) end = start - timedelta(days=random.randint(1, 10)) return start.isoformat(), end.isoformat() def random_work_id():
def random_hex(length): def random_interval(): now = datetime.now() start = now - timedelta(days=random.randint(0, 365*3)) end = start + timedelta(days=random.randint(1, 10)) return start.isoformat(), end.isoformat() def random_work_id():
1,035
https://:@github.com/cungnv/scrapex.git
9a2b814ea328f79ee330381423ebeaddb9b11852
@@ -208,7 +208,7 @@ class DB(object): for log in self._db.logs.find(query): - logs.append(logs) + logs.append(log) return logs
scrapex/db.py
ReplaceText(target='log' @(211,15)->(211,19))
class DB(object): for log in self._db.logs.find(query): logs.append(logs) return logs
class DB(object): for log in self._db.logs.find(query): logs.append(log) return logs
1,036
https://:@github.com/erdc/quest.git
b0b34875f72a05122cd33f5f1b4148dedecdfe89
@@ -139,7 +139,7 @@ class CoopsPyoos(DataServiceBase): with open(csvFile_path, 'w') as f: f.write(response) - data_files[location][parameter] = filename + data_files[location][parameter] = csvFile_path else: data_files[location][parameter] = None
dsl/services/coops_pyoos.py
ReplaceText(target='csvFile_path' @(142,58)->(142,66))
class CoopsPyoos(DataServiceBase): with open(csvFile_path, 'w') as f: f.write(response) data_files[location][parameter] = filename else: data_files[location][parameter] = None
class CoopsPyoos(DataServiceBase): with open(csvFile_path, 'w') as f: f.write(response) data_files[location][parameter] = csvFile_path else: data_files[location][parameter] = None
1,037
https://:@github.com/erdc/quest.git
016b54452d1170029e57bcbe117e94513a62702c
@@ -96,7 +96,7 @@ class RstMerge(FilterBase): # update feature geometry metadata with rasterio.open(dst) as f: - geometry = util.bbox2poly(f.bounds.left, f.bounds.right, f.bounds.bottom, f.bounds.top, as_shapely=True) + geometry = util.bbox2poly(f.bounds.left, f.bounds.bottom, f.bounds.right, f.bounds.top, as_shapely=True) update_metadata(feature, quest_metadata={'geometry': geometry.to_wkt()}) # update dataset metadata
quest/filters/raster/rst_merge.py
ArgSwap(idxs=1<->2 @(99,23)->(99,37))
class RstMerge(FilterBase): # update feature geometry metadata with rasterio.open(dst) as f: geometry = util.bbox2poly(f.bounds.left, f.bounds.right, f.bounds.bottom, f.bounds.top, as_shapely=True) update_metadata(feature, quest_metadata={'geometry': geometry.to_wkt()}) # update dataset metadata
class RstMerge(FilterBase): # update feature geometry metadata with rasterio.open(dst) as f: geometry = util.bbox2poly(f.bounds.left, f.bounds.bottom, f.bounds.right, f.bounds.top, as_shapely=True) update_metadata(feature, quest_metadata={'geometry': geometry.to_wkt()}) # update dataset metadata
1,038
https://:@github.com/erdc/quest.git
088fce2676801d8ac47d93652d14040296deca36
@@ -140,7 +140,7 @@ class RstWatershedDelineation(FilterBase): if snap.lower() == 'jenson': stream_threshold_pct = options.get('stream_threshold_pct') stream_threshold_abs = options.get('stream_threshold_abs') - outlet_points = snap_points_jenson(flow_accumulation, proj_points, + proj_points = snap_points_jenson(flow_accumulation, proj_points, stream_threshold_pct=stream_threshold_pct, stream_threshold_abs=stream_threshold_abs) if p.is_latlong(): snapped_points = [src.xy(*point) for point in proj_points]
quest/filters/raster/rst_watershed.py
ReplaceText(target='proj_points' @(143,20)->(143,33))
class RstWatershedDelineation(FilterBase): if snap.lower() == 'jenson': stream_threshold_pct = options.get('stream_threshold_pct') stream_threshold_abs = options.get('stream_threshold_abs') outlet_points = snap_points_jenson(flow_accumulation, proj_points, stream_threshold_pct=stream_threshold_pct, stream_threshold_abs=stream_threshold_abs) if p.is_latlong(): snapped_points = [src.xy(*point) for point in proj_points]
class RstWatershedDelineation(FilterBase): if snap.lower() == 'jenson': stream_threshold_pct = options.get('stream_threshold_pct') stream_threshold_abs = options.get('stream_threshold_abs') proj_points = snap_points_jenson(flow_accumulation, proj_points, stream_threshold_pct=stream_threshold_pct, stream_threshold_abs=stream_threshold_abs) if p.is_latlong(): snapped_points = [src.xy(*point) for point in proj_points]
1,039
https://:@github.com/erdc/quest.git
c8e7db0a0b850f8f785e63892e708f2334018d5c
@@ -180,7 +180,7 @@ def copy(uris, destination_collection): if resource == 'datasets': dataset_metadata = get_metadata(uri)[uri] - collection_path = os.path.join(project_path, feature_metadata['collection']) + collection_path = os.path.join(project_path, dataset_metadata['collection']) feature = dataset_metadata['feature']
quest/api/manage.py
ReplaceText(target='dataset_metadata' @(183,57)->(183,73))
def copy(uris, destination_collection): if resource == 'datasets': dataset_metadata = get_metadata(uri)[uri] collection_path = os.path.join(project_path, feature_metadata['collection']) feature = dataset_metadata['feature']
def copy(uris, destination_collection): if resource == 'datasets': dataset_metadata = get_metadata(uri)[uri] collection_path = os.path.join(project_path, dataset_metadata['collection']) feature = dataset_metadata['feature']
1,040
https://:@github.com/next-security-lab/deep-confusables-cli.git
fe2f6fe46f14778cfcb74852d817d1876d171352
@@ -92,7 +92,7 @@ def main(): print('Similar domains to {}'.format(dom)) domains.difference_update(set(dom)) for d in domains: - print_diff(d, args.domain) + print_diff(args.domain, d) if write: f.write(d + "\n") if (args.check):
v2d/main.py
ArgSwap(idxs=0<->1 @(95,16)->(95,26))
def main(): print('Similar domains to {}'.format(dom)) domains.difference_update(set(dom)) for d in domains: print_diff(d, args.domain) if write: f.write(d + "\n") if (args.check):
def main(): print('Similar domains to {}'.format(dom)) domains.difference_update(set(dom)) for d in domains: print_diff(args.domain, d) if write: f.write(d + "\n") if (args.check):
1,041
https://:@github.com/jkittley/RFM69.git
bcf1e92ee49342ee2ac1d39be7b95671c1ddd6d3
@@ -143,7 +143,7 @@ class RFM69(): self.writeReg(REG_PACKETCONFIG2, 0) def readReg(self, addr): - return self.spi.xfer([addr | 0x7F, 0]) + return self.spi.xfer([addr & 0x7F, 0]) def writeReg(self, addr, value): self.spi.xfer([addr | 0x80, value])
RFM69.py
ReplaceText(target='&' @(146,31)->(146,32))
class RFM69(): self.writeReg(REG_PACKETCONFIG2, 0) def readReg(self, addr): return self.spi.xfer([addr | 0x7F, 0]) def writeReg(self, addr, value): self.spi.xfer([addr | 0x80, value])
class RFM69(): self.writeReg(REG_PACKETCONFIG2, 0) def readReg(self, addr): return self.spi.xfer([addr & 0x7F, 0]) def writeReg(self, addr, value): self.spi.xfer([addr | 0x80, value])
1,042
https://:@github.com/schuderer/mllaunchpad.git
93c87d9c160b3270319e6b344f1d1831d4323014
@@ -50,7 +50,7 @@ def train_model(complete_conf): logger.warning("Model's class is not a subclass of ModelInterface: %s", model) model_store = resource.ModelStore(complete_conf) - model_store.dump_trained_model(model_conf, model, metrics) + model_store.dump_trained_model(complete_conf, model, metrics) logger.info("Created and stored trained model %s, version %s, metrics %s", model_conf['name'], model_conf['version'], metrics)
launchpad/train.py
ReplaceText(target='complete_conf' @(53,35)->(53,45))
def train_model(complete_conf): logger.warning("Model's class is not a subclass of ModelInterface: %s", model) model_store = resource.ModelStore(complete_conf) model_store.dump_trained_model(model_conf, model, metrics) logger.info("Created and stored trained model %s, version %s, metrics %s", model_conf['name'], model_conf['version'], metrics)
def train_model(complete_conf): logger.warning("Model's class is not a subclass of ModelInterface: %s", model) model_store = resource.ModelStore(complete_conf) model_store.dump_trained_model(complete_conf, model, metrics) logger.info("Created and stored trained model %s, version %s, metrics %s", model_conf['name'], model_conf['version'], metrics)
1,043
https://:@github.com/daanvdk/is_valid.git
0c78bad8195b8c4b702b368e9d13d1ccaf573382
@@ -158,4 +158,4 @@ def is_set_of(predicate): return (True, explanation) if valid else (False, { elems[i]: value for i, value in explanation.items() }) - return is_if(is_set, predicate, else_valid=False) + return is_if(is_set, is_valid, else_valid=False)
is_valid/structure_predicates.py
ReplaceText(target='is_valid' @(161,25)->(161,34))
def is_set_of(predicate): return (True, explanation) if valid else (False, { elems[i]: value for i, value in explanation.items() }) return is_if(is_set, predicate, else_valid=False)
def is_set_of(predicate): return (True, explanation) if valid else (False, { elems[i]: value for i, value in explanation.items() }) return is_if(is_set, is_valid, else_valid=False)
1,044
https://:@github.com/riptano/cdm.git
cb0de67c9a5418a052d77575a0c81b9159e0089a
@@ -59,7 +59,7 @@ class Importer(object): for _ in pool.imap_unordered(save, self.iter()): i += 1 if i % 10 == 0: - pool.update(i) + p.update(i)
cdm/importer.py
ReplaceText(target='p' @(62,20)->(62,24))
class Importer(object): for _ in pool.imap_unordered(save, self.iter()): i += 1 if i % 10 == 0: pool.update(i)
class Importer(object): for _ in pool.imap_unordered(save, self.iter()): i += 1 if i % 10 == 0: p.update(i)
1,045
https://:@github.com/moggers87/exhibition.git
18f720dd34997c6da7235f2ea1344e09a670f308
@@ -95,7 +95,7 @@ class Config: if self.parent is not None: for k in self.parent.keys(): - if k in _keys_set: + if k not in _keys_set: _keys_set.add(k) yield k
exhibition/main.py
ReplaceText(target=' not in ' @(98,20)->(98,24))
class Config: if self.parent is not None: for k in self.parent.keys(): if k in _keys_set: _keys_set.add(k) yield k
class Config: if self.parent is not None: for k in self.parent.keys(): if k not in _keys_set: _keys_set.add(k) yield k
1,046
https://:@github.com/moggers87/exhibition.git
49a8eb6842d2eedecd239f35c651e4afde69cb87
@@ -486,7 +486,7 @@ def serve(settings): path = pathlib.Path(settings["deploy_path"], path) - if not (path.exists() and path.suffix): + if not (path.exists() or path.suffix): for ext in Node._strip_exts: new_path = path.with_suffix(ext) if new_path.exists():
exhibition/main.py
ReplaceText(target='or' @(489,34)->(489,37))
def serve(settings): path = pathlib.Path(settings["deploy_path"], path) if not (path.exists() and path.suffix): for ext in Node._strip_exts: new_path = path.with_suffix(ext) if new_path.exists():
def serve(settings): path = pathlib.Path(settings["deploy_path"], path) if not (path.exists() or path.suffix): for ext in Node._strip_exts: new_path = path.with_suffix(ext) if new_path.exists():
1,047
https://:@github.com/civrev/rlrisk.git
fb9ba4a28cd8dced235e9cb97b861130baa351a9
@@ -575,7 +575,7 @@ class Risk: t_count=3 #now add to amount recruited - recruit+=t_count + recruit=t_count #calculate for continents for continent in self.continents:
rlrisk/environment/risk.py
ReplaceText(target='=' @(578,15)->(578,17))
class Risk: t_count=3 #now add to amount recruited recruit+=t_count #calculate for continents for continent in self.continents:
class Risk: t_count=3 #now add to amount recruited recruit=t_count #calculate for continents for continent in self.continents:
1,048
https://:@gitlab.com/philn/fb2feed.git
9e6dc0b80947383a54e0fe99c656d419a4ede2c9
@@ -126,7 +126,7 @@ async def page_to_atom(browser, page_id, root_dir, media_dir, media_url_slug): fe = fg.add_entry() fe.author(name=page_id, email="%s.facebook.no-reply@fb2feed.org" % page_id) - fe.id(post_url) + fe.id(post_id) fe.link(href=post_url, rel="alternate") fe.published(timestamp) fe.updated(timestamp)
fb2feed.py
ReplaceText(target='post_id' @(129,18)->(129,26))
async def page_to_atom(browser, page_id, root_dir, media_dir, media_url_slug): fe = fg.add_entry() fe.author(name=page_id, email="%s.facebook.no-reply@fb2feed.org" % page_id) fe.id(post_url) fe.link(href=post_url, rel="alternate") fe.published(timestamp) fe.updated(timestamp)
async def page_to_atom(browser, page_id, root_dir, media_dir, media_url_slug): fe = fg.add_entry() fe.author(name=page_id, email="%s.facebook.no-reply@fb2feed.org" % page_id) fe.id(post_id) fe.link(href=post_url, rel="alternate") fe.published(timestamp) fe.updated(timestamp)
1,049
https://:@github.com/sputt/req-compile.git
f76edf2e9821dcef6d7bc851282731821e569d81
@@ -111,7 +111,7 @@ class DistributionCollection(object): self.nodes[key] = node # If a new extra is being supplied, update the metadata - if reason and node.metadata and reason.extras and set(reason.extras) & node.extras: + if reason and node.metadata and reason.extras and set(reason.extras) - node.extras: metadata_to_apply = node.metadata if source is not None and source.key in self.nodes:
req_compile/dists.py
ReplaceText(target='-' @(114,77)->(114,78))
class DistributionCollection(object): self.nodes[key] = node # If a new extra is being supplied, update the metadata if reason and node.metadata and reason.extras and set(reason.extras) & node.extras: metadata_to_apply = node.metadata if source is not None and source.key in self.nodes:
class DistributionCollection(object): self.nodes[key] = node # If a new extra is being supplied, update the metadata if reason and node.metadata and reason.extras and set(reason.extras) - node.extras: metadata_to_apply = node.metadata if source is not None and source.key in self.nodes:
1,050
https://:@github.com/inovonics/cloud-datastore.git
72b1cb9760973b300b5ebf3bad1d6836b15d908a
@@ -113,7 +113,7 @@ class InoObjectBase: def _validate_oid(self): # Verify the oid is a UUID type variable - if isinstance(getattr(self, 'oid'), uuid.UUID): + if not isinstance(getattr(self, 'oid'), uuid.UUID): return "oid not of type uuid.UUID but type {}, value {}".format(type(getattr(self, 'oid')), getattr(self, 'oid')) return None
inovonics/cloud/datastore/bases.py
ReplaceText(target='not ' @(116,11)->(116,11))
class InoObjectBase: def _validate_oid(self): # Verify the oid is a UUID type variable if isinstance(getattr(self, 'oid'), uuid.UUID): return "oid not of type uuid.UUID but type {}, value {}".format(type(getattr(self, 'oid')), getattr(self, 'oid')) return None
class InoObjectBase: def _validate_oid(self): # Verify the oid is a UUID type variable if not isinstance(getattr(self, 'oid'), uuid.UUID): return "oid not of type uuid.UUID but type {}, value {}".format(type(getattr(self, 'oid')), getattr(self, 'oid')) return None
1,051
https://:@github.com/shawnbrown/squint.git
9912b140bc25d6d8679819ee8c4fd037a0b01b54
@@ -116,7 +116,7 @@ class Result(Iterator): value = __next__(self) finally: self._started_iteration = True - bound_method = __next__.__get__(self.__class__, self) + bound_method = __next__.__get__(self, self.__class__) self.__next__ = bound_method # <- Replace __next__ method! return value
squint/result.py
ArgSwap(idxs=0<->1 @(119,27)->(119,43))
class Result(Iterator): value = __next__(self) finally: self._started_iteration = True bound_method = __next__.__get__(self.__class__, self) self.__next__ = bound_method # <- Replace __next__ method! return value
class Result(Iterator): value = __next__(self) finally: self._started_iteration = True bound_method = __next__.__get__(self, self.__class__) self.__next__ = bound_method # <- Replace __next__ method! return value
1,052
https://:@github.com/kav2k/singularity-pipeline.git
4eacf7dfa1ddcdd7feec8ff928d8cac3ee170c72
@@ -136,7 +136,7 @@ class Pipeline(): self.eprint.bold("# Running pipeline...\n") if not self.dry_run: - if os.path.isfile(self.imagefile): + if not os.path.isfile(self.imagefile): raise RuntimeError("Image {} does not exist".format(self.imagefile)) for spec in self.binds:
singularity_pipeline/pipeline.py
ReplaceText(target='not ' @(139,15)->(139,15))
class Pipeline(): self.eprint.bold("# Running pipeline...\n") if not self.dry_run: if os.path.isfile(self.imagefile): raise RuntimeError("Image {} does not exist".format(self.imagefile)) for spec in self.binds:
class Pipeline(): self.eprint.bold("# Running pipeline...\n") if not self.dry_run: if not os.path.isfile(self.imagefile): raise RuntimeError("Image {} does not exist".format(self.imagefile)) for spec in self.binds:
1,053
https://:@github.com/mvinii94/aws-lambda-log-collector.git
682850f282b70aa18663699c7e5e32bc4f6a8be1
@@ -25,7 +25,7 @@ def cli(function_name, profile, region, output, start_time, end_time, pattern, l epoch_start_time = parse_time(start_time) epoch_end_time = parse_time(end_time) - if epoch_start_time < epoch_end_time: + if epoch_start_time > epoch_end_time: raise Exception(INVALID_DATES) available_profiles = get_profiles()
collector/cli.py
ReplaceText(target='>' @(28,25)->(28,26))
def cli(function_name, profile, region, output, start_time, end_time, pattern, l epoch_start_time = parse_time(start_time) epoch_end_time = parse_time(end_time) if epoch_start_time < epoch_end_time: raise Exception(INVALID_DATES) available_profiles = get_profiles()
def cli(function_name, profile, region, output, start_time, end_time, pattern, l epoch_start_time = parse_time(start_time) epoch_end_time = parse_time(end_time) if epoch_start_time > epoch_end_time: raise Exception(INVALID_DATES) available_profiles = get_profiles()
1,054
https://:@github.com/peteshadbolt/abp.git
9d859145d8a6ec45a7831235bfc05f45932709a7
@@ -58,7 +58,7 @@ class CircuitModel(object): def act_cz(self, control, target): """ Act a CU somewhere """ control = 1 << control - target = 1 << control + target = 1 << target for i in xrange(self.d): if (i & control) and (i & target): self.state[i, 0] *= -1
abp/qi.py
ReplaceText(target='target' @(61,22)->(61,29))
class CircuitModel(object): def act_cz(self, control, target): """ Act a CU somewhere """ control = 1 << control target = 1 << control for i in xrange(self.d): if (i & control) and (i & target): self.state[i, 0] *= -1
class CircuitModel(object): def act_cz(self, control, target): """ Act a CU somewhere """ control = 1 << control target = 1 << target for i in xrange(self.d): if (i & control) and (i & target): self.state[i, 0] *= -1
1,055
https://:@github.com/peteshadbolt/abp.git
1b787c47377d73a0519f60cc77657b2b09577b49
@@ -134,7 +134,7 @@ class GraphState(object): ci = self.get_connection_info(a, b) if ci["non1"] and not clifford.is_diagonal(self.node[a]["vop"]): debug("cphase: left one needs treatment again -> putting it to Id") - self.remove_vop(b, a) + self.remove_vop(a, b) self.cz_with_table(a, b)
abp/graphstate.py
ArgSwap(idxs=0<->1 @(137,12)->(137,27))
class GraphState(object): ci = self.get_connection_info(a, b) if ci["non1"] and not clifford.is_diagonal(self.node[a]["vop"]): debug("cphase: left one needs treatment again -> putting it to Id") self.remove_vop(b, a) self.cz_with_table(a, b)
class GraphState(object): ci = self.get_connection_info(a, b) if ci["non1"] and not clifford.is_diagonal(self.node[a]["vop"]): debug("cphase: left one needs treatment again -> putting it to Id") self.remove_vop(a, b) self.cz_with_table(a, b)
1,056
https://:@github.com/jsbroks/imantics.git
a9919325c1628d65cc0baf150599e3d04d6cf1be
@@ -43,7 +43,7 @@ class Annotation(Semantic): :param polygons: bbox to create annotation from :type polygons: :class:`BBox`, list, tuple """ - return cls(image=image, category=image, bbox=bbox) + return cls(image=image, category=category, bbox=bbox) @classmethod def from_polygons(cls, polygons, image=None, category=None):
imantics/annotation.py
ReplaceText(target='category' @(46,41)->(46,46))
class Annotation(Semantic): :param polygons: bbox to create annotation from :type polygons: :class:`BBox`, list, tuple """ return cls(image=image, category=image, bbox=bbox) @classmethod def from_polygons(cls, polygons, image=None, category=None):
class Annotation(Semantic): :param polygons: bbox to create annotation from :type polygons: :class:`BBox`, list, tuple """ return cls(image=image, category=category, bbox=bbox) @classmethod def from_polygons(cls, polygons, image=None, category=None):
1,057
https://:@github.com/278mt/cotohappy.git
09e71e8a654b93db750a1e5f55b1c063b3070470
@@ -26,7 +26,7 @@ class Reshape(object): Reshape(mode='links', data=link) for link in chunk_info['links'] ] - self.predicate = chunk_info['predicate'] if 'predicate' in data else [] + self.predicate = chunk_info['predicate'] if 'predicate' in chunk_info else [] self.tokens = [ Reshape(mode='tokens', data=token)
cotohappy/reshape.py
ReplaceText(target='chunk_info' @(29,72)->(29,76))
class Reshape(object): Reshape(mode='links', data=link) for link in chunk_info['links'] ] self.predicate = chunk_info['predicate'] if 'predicate' in data else [] self.tokens = [ Reshape(mode='tokens', data=token)
class Reshape(object): Reshape(mode='links', data=link) for link in chunk_info['links'] ] self.predicate = chunk_info['predicate'] if 'predicate' in chunk_info else [] self.tokens = [ Reshape(mode='tokens', data=token)
1,058
https://:@github.com/AntoineToubhans/MongoTs.git
79fa91c7f9f3ecd0e0d22ca01a67a9efc84a9611
@@ -82,4 +82,4 @@ class MongoTSCollection(): raw_data = list(self._collection.aggregate(pipeline)) - return build_dataframe(raw_data, aggregateby, groupby) + return build_dataframe(raw_data, parsed_aggregateby, groupby)
mongots/collection.py
ReplaceText(target='parsed_aggregateby' @(85,41)->(85,52))
class MongoTSCollection(): raw_data = list(self._collection.aggregate(pipeline)) return build_dataframe(raw_data, aggregateby, groupby)
class MongoTSCollection(): raw_data = list(self._collection.aggregate(pipeline)) return build_dataframe(raw_data, parsed_aggregateby, groupby)
1,059
https://:@github.com/thenewguy/django-randomfields.git
983d6a2940e17a42326137b6592e0ed3ea3de8bd
@@ -49,7 +49,7 @@ class RandomStringFieldBase(RandomFieldBase): 'min_length': self.min_length, } defaults.update(kwargs) - return super(RandomStringFieldBase, self).formfield(**kwargs) + return super(RandomStringFieldBase, self).formfield(**defaults) class RandomCharField(RandomStringFieldBase, models.CharField): def check(self, **kwargs):
randomfields/models/fields/string.py
ReplaceText(target='defaults' @(52,62)->(52,68))
class RandomStringFieldBase(RandomFieldBase): 'min_length': self.min_length, } defaults.update(kwargs) return super(RandomStringFieldBase, self).formfield(**kwargs) class RandomCharField(RandomStringFieldBase, models.CharField): def check(self, **kwargs):
class RandomStringFieldBase(RandomFieldBase): 'min_length': self.min_length, } defaults.update(kwargs) return super(RandomStringFieldBase, self).formfield(**defaults) class RandomCharField(RandomStringFieldBase, models.CharField): def check(self, **kwargs):
1,060
https://:@github.com/mathcamp/devbox.git
b014314770fe44557b57c7d9d024959be9fb4f4a
@@ -175,7 +175,7 @@ def git_describe(describe_args): if proc.returncode != 0: print("Error parsing git revision! Make sure that you have tagged a " "commit, and that the tag matches the 'tag_match' argument") - print("Git output: " + output) + print("Git output: " + description) return { 'tag': 'unknown', 'description': 'unknown',
version_helper.py
ReplaceText(target='description' @(178,31)->(178,37))
def git_describe(describe_args): if proc.returncode != 0: print("Error parsing git revision! Make sure that you have tagged a " "commit, and that the tag matches the 'tag_match' argument") print("Git output: " + output) return { 'tag': 'unknown', 'description': 'unknown',
def git_describe(describe_args): if proc.returncode != 0: print("Error parsing git revision! Make sure that you have tagged a " "commit, and that the tag matches the 'tag_match' argument") print("Git output: " + description) return { 'tag': 'unknown', 'description': 'unknown',
1,061
https://:@github.com/mathcamp/devbox.git
54a298fc2018e1b4328a68c08740f5bd49f0864a
@@ -213,7 +213,7 @@ def unbox(repo, dest=None, no_deps=False, *parents): LOG.info("Installing into %s", install_dir) dest_conf = load_conf(install_dir) venv = dest_conf.get('env') - with pushd(dest): + with pushd(install_dir): if venv is not None: venv['path'] = os.path.abspath(venv['path']) run_commands(conf.get('post_setup', []), venv)
devbox/unbox.py
ReplaceText(target='install_dir' @(216,19)->(216,23))
def unbox(repo, dest=None, no_deps=False, *parents): LOG.info("Installing into %s", install_dir) dest_conf = load_conf(install_dir) venv = dest_conf.get('env') with pushd(dest): if venv is not None: venv['path'] = os.path.abspath(venv['path']) run_commands(conf.get('post_setup', []), venv)
def unbox(repo, dest=None, no_deps=False, *parents): LOG.info("Installing into %s", install_dir) dest_conf = load_conf(install_dir) venv = dest_conf.get('env') with pushd(install_dir): if venv is not None: venv['path'] = os.path.abspath(venv['path']) run_commands(conf.get('post_setup', []), venv)
1,062
https://:@github.com/ccxtechnologies/adbus.git
58d6866e77afaac02195d40348e207d9258b4fc2
@@ -177,7 +177,7 @@ class Method: self.name = x.attrib['name'] if x.tag == 'arg': if x.attrib['direction'] == 'out': - self.return_signature = x.attrib['type'] + self.return_signature += x.attrib['type'] elif x.attrib['direction'] == 'in': self.arg_signatures.append(x.attrib['type'])
adbus/client/proxy.py
ReplaceText(target='+=' @(180,42)->(180,43))
class Method: self.name = x.attrib['name'] if x.tag == 'arg': if x.attrib['direction'] == 'out': self.return_signature = x.attrib['type'] elif x.attrib['direction'] == 'in': self.arg_signatures.append(x.attrib['type'])
class Method: self.name = x.attrib['name'] if x.tag == 'arg': if x.attrib['direction'] == 'out': self.return_signature += x.attrib['type'] elif x.attrib['direction'] == 'in': self.arg_signatures.append(x.attrib['type'])
1,063
https://:@github.com/balazstothofficial/models.git
d0b6a34bb0dbd981e3785661f04f04cde9c4222b
@@ -107,7 +107,7 @@ def run_mnist_eager(flags_obj): # Automatically determine device and data_format (device, data_format) = ('/gpu:0', 'channels_first') - if flags_obj.no_gpu or tf.test.is_gpu_available(): + if flags_obj.no_gpu or not tf.test.is_gpu_available(): (device, data_format) = ('/cpu:0', 'channels_last') # If data_format is defined in FLAGS, overwrite automatically set value. if flags_obj.data_format is not None:
official/mnist/mnist_eager.py
ReplaceText(target='not ' @(110,25)->(110,25))
def run_mnist_eager(flags_obj): # Automatically determine device and data_format (device, data_format) = ('/gpu:0', 'channels_first') if flags_obj.no_gpu or tf.test.is_gpu_available(): (device, data_format) = ('/cpu:0', 'channels_last') # If data_format is defined in FLAGS, overwrite automatically set value. if flags_obj.data_format is not None:
def run_mnist_eager(flags_obj): # Automatically determine device and data_format (device, data_format) = ('/gpu:0', 'channels_first') if flags_obj.no_gpu or not tf.test.is_gpu_available(): (device, data_format) = ('/cpu:0', 'channels_last') # If data_format is defined in FLAGS, overwrite automatically set value. if flags_obj.data_format is not None:
1,064
https://:@github.com/balazstothofficial/models.git
e9dbef6be831bac9c5cbacf6b3b13e4557e4777c
@@ -353,7 +353,7 @@ class Worker(threading.Thread): policy_loss = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=memory.actions, logits=logits) policy_loss *= tf.stop_gradient(advantage) - policy_loss = 0.01 * entropy + policy_loss -= 0.01 * entropy total_loss = tf.reduce_mean((0.5 * value_loss + policy_loss)) return total_loss
research/a3c_blogpost/a3c_cartpole.py
ReplaceText(target='-=' @(356,16)->(356,17))
class Worker(threading.Thread): policy_loss = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=memory.actions, logits=logits) policy_loss *= tf.stop_gradient(advantage) policy_loss = 0.01 * entropy total_loss = tf.reduce_mean((0.5 * value_loss + policy_loss)) return total_loss
class Worker(threading.Thread): policy_loss = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=memory.actions, logits=logits) policy_loss *= tf.stop_gradient(advantage) policy_loss -= 0.01 * entropy total_loss = tf.reduce_mean((0.5 * value_loss + policy_loss)) return total_loss
1,065
https://:@github.com/idigbio/idq.git
c3c53aa2eab9d4a958f5ac338c4daac8fae2d192
@@ -31,7 +31,7 @@ def create_harness(w): for f in w.required_fields: f_name = f.replace(":","_") - setattr(SingleForm,f_name,StringField(f_name)) + setattr(SingleForm,f_name,StringField(f)) setattr(SingleForm,"submit",SubmitField(u'Process'))
idq/harness/__init__.py
ReplaceText(target='f' @(34,46)->(34,52))
def create_harness(w): for f in w.required_fields: f_name = f.replace(":","_") setattr(SingleForm,f_name,StringField(f_name)) setattr(SingleForm,"submit",SubmitField(u'Process'))
def create_harness(w): for f in w.required_fields: f_name = f.replace(":","_") setattr(SingleForm,f_name,StringField(f)) setattr(SingleForm,"submit",SubmitField(u'Process'))
1,066
https://:@github.com/iMerica/dj-models.git
659ab9846e81d95bb75dbb3c00147324bf0d6541
@@ -22,7 +22,7 @@ def login(request): else: errors = {} response = HttpResponse() - response.session.set_test_cookie() + request.session.set_test_cookie() t = template_loader.get_template('registration/login') c = Context(request, { 'form': formfields.FormWrapper(manipulator, request.POST, errors),
django/views/auth/login.py
ReplaceText(target='request' @(25,4)->(25,12))
def login(request): else: errors = {} response = HttpResponse() response.session.set_test_cookie() t = template_loader.get_template('registration/login') c = Context(request, { 'form': formfields.FormWrapper(manipulator, request.POST, errors),
def login(request): else: errors = {} response = HttpResponse() request.session.set_test_cookie() t = template_loader.get_template('registration/login') c = Context(request, { 'form': formfields.FormWrapper(manipulator, request.POST, errors),
1,067
https://:@github.com/iMerica/dj-models.git
a97648a7e03fb95b09e888e5d59d82d57fb289b7
@@ -105,7 +105,7 @@ class DecoratorsTest(TestCase): """ def my_view(request): return "response" - my_view_cached = cache_page(123, my_view) + my_view_cached = cache_page(my_view, 123) self.assertEqual(my_view_cached(HttpRequest()), "response") class MethodDecoratorAdapterTests(TestCase):
tests/regressiontests/decorators/tests.py
ArgSwap(idxs=0<->1 @(108,25)->(108,35))
class DecoratorsTest(TestCase): """ def my_view(request): return "response" my_view_cached = cache_page(123, my_view) self.assertEqual(my_view_cached(HttpRequest()), "response") class MethodDecoratorAdapterTests(TestCase):
class DecoratorsTest(TestCase): """ def my_view(request): return "response" my_view_cached = cache_page(my_view, 123) self.assertEqual(my_view_cached(HttpRequest()), "response") class MethodDecoratorAdapterTests(TestCase):
1,068
https://:@github.com/iMerica/dj-models.git
7db68a888b633f2b65f753d50f21463b65a01edf
@@ -399,7 +399,7 @@ def language(parser, token): """ bits = token.split_contents() - if len(bits) < 2: + if len(bits) != 2: raise TemplateSyntaxError("'%s' takes one argument (language)" % bits[0]) language = parser.compile_filter(bits[1]) nodelist = parser.parse(('endlanguage',))
django/templatetags/i18n.py
ReplaceText(target='!=' @(402,17)->(402,18))
def language(parser, token): """ bits = token.split_contents() if len(bits) < 2: raise TemplateSyntaxError("'%s' takes one argument (language)" % bits[0]) language = parser.compile_filter(bits[1]) nodelist = parser.parse(('endlanguage',))
def language(parser, token): """ bits = token.split_contents() if len(bits) != 2: raise TemplateSyntaxError("'%s' takes one argument (language)" % bits[0]) language = parser.compile_filter(bits[1]) nodelist = parser.parse(('endlanguage',))
1,069
https://:@github.com/iMerica/dj-models.git
b2050ff546da4164f90a795e55d7d8c55981783d
@@ -169,7 +169,7 @@ class SQLCompiler(object): if isinstance(col, (list, tuple)): alias, column = col table = self.query.alias_map[alias][TABLE_NAME] - if table in only_load and col not in only_load[table]: + if table in only_load and column not in only_load[table]: continue r = '%s.%s' % (qn(alias), qn(column)) if with_aliases:
django/db/models/sql/compiler.py
ReplaceText(target='column' @(172,46)->(172,49))
class SQLCompiler(object): if isinstance(col, (list, tuple)): alias, column = col table = self.query.alias_map[alias][TABLE_NAME] if table in only_load and col not in only_load[table]: continue r = '%s.%s' % (qn(alias), qn(column)) if with_aliases:
class SQLCompiler(object): if isinstance(col, (list, tuple)): alias, column = col table = self.query.alias_map[alias][TABLE_NAME] if table in only_load and column not in only_load[table]: continue r = '%s.%s' % (qn(alias), qn(column)) if with_aliases:
1,070
https://:@github.com/iMerica/dj-models.git
cfba2460370a6d1808b78e2ba0709ea5c8b7e773
@@ -42,7 +42,7 @@ def check_settings(base_url=None): Checks if the staticfiles settings have sane values. """ - if base_url is not None: + if base_url is None: base_url = settings.STATIC_URL if not base_url: raise ImproperlyConfigured(
django/contrib/staticfiles/utils.py
ReplaceText(target=' is ' @(45,15)->(45,23))
def check_settings(base_url=None): Checks if the staticfiles settings have sane values. """ if base_url is not None: base_url = settings.STATIC_URL if not base_url: raise ImproperlyConfigured(
def check_settings(base_url=None): Checks if the staticfiles settings have sane values. """ if base_url is None: base_url = settings.STATIC_URL if not base_url: raise ImproperlyConfigured(
1,071
https://:@github.com/iMerica/dj-models.git
5f2be4ecbb5df3760f4c6e49170478719d3026d7
@@ -824,7 +824,7 @@ class SQLInsertCompiler(SQLCompiler): for val in values ] if self.return_id and self.connection.features.can_return_id_from_insert: - params = values[0] + params = params[0] col = "%s.%s" % (qn(opts.db_table), qn(opts.pk.column)) result.append("VALUES (%s)" % ", ".join(placeholders[0])) r_fmt, r_params = self.connection.ops.return_insert_id()
django/db/models/sql/compiler.py
ReplaceText(target='params' @(827,21)->(827,27))
class SQLInsertCompiler(SQLCompiler): for val in values ] if self.return_id and self.connection.features.can_return_id_from_insert: params = values[0] col = "%s.%s" % (qn(opts.db_table), qn(opts.pk.column)) result.append("VALUES (%s)" % ", ".join(placeholders[0])) r_fmt, r_params = self.connection.ops.return_insert_id()
class SQLInsertCompiler(SQLCompiler): for val in values ] if self.return_id and self.connection.features.can_return_id_from_insert: params = params[0] col = "%s.%s" % (qn(opts.db_table), qn(opts.pk.column)) result.append("VALUES (%s)" % ", ".join(placeholders[0])) r_fmt, r_params = self.connection.ops.return_insert_id()
1,072
https://:@github.com/iMerica/dj-models.git
d72d5ce8274992ce01e39f866a7a250bc459eefe
@@ -37,7 +37,7 @@ class GeoSQLCompiler(compiler.SQLCompiler): if isinstance(col, (list, tuple)): alias, column = col table = self.query.alias_map[alias][TABLE_NAME] - if table in only_load and col not in only_load[table]: + if table in only_load and column not in only_load[table]: continue r = self.get_field_select(field, alias, column) if with_aliases:
django/contrib/gis/db/models/sql/compiler.py
ReplaceText(target='column' @(40,46)->(40,49))
class GeoSQLCompiler(compiler.SQLCompiler): if isinstance(col, (list, tuple)): alias, column = col table = self.query.alias_map[alias][TABLE_NAME] if table in only_load and col not in only_load[table]: continue r = self.get_field_select(field, alias, column) if with_aliases:
class GeoSQLCompiler(compiler.SQLCompiler): if isinstance(col, (list, tuple)): alias, column = col table = self.query.alias_map[alias][TABLE_NAME] if table in only_load and column not in only_load[table]: continue r = self.get_field_select(field, alias, column) if with_aliases:
1,073
https://:@github.com/iMerica/dj-models.git
6ecbac21a9017a53fe18ac81c9c1d2f28185a292
@@ -111,5 +111,5 @@ class OSMWidget(BaseGeometryWidget): return 900913 def render(self, name, value, attrs=None): - return super(self, OSMWidget).render(name, value, + return super(OSMWidget, self).render(name, value, {'default_lon': self.default_lon, 'default_lat': self.default_lat})
django/contrib/gis/forms/widgets.py
ArgSwap(idxs=0<->1 @(114,15)->(114,20))
class OSMWidget(BaseGeometryWidget): return 900913 def render(self, name, value, attrs=None): return super(self, OSMWidget).render(name, value, {'default_lon': self.default_lon, 'default_lat': self.default_lat})
class OSMWidget(BaseGeometryWidget): return 900913 def render(self, name, value, attrs=None): return super(OSMWidget, self).render(name, value, {'default_lon': self.default_lon, 'default_lat': self.default_lat})
1,074
https://:@github.com/iMerica/dj-models.git
86c248aa646183ef4a1cb407bb3e4cb597272f63
@@ -575,7 +575,7 @@ class SQLCompiler(object): for order, order_params in ordering_group_by: # Even if we have seen the same SQL string, it might have # different params, so, we add same SQL in "has params" case. - if order not in seen or params: + if order not in seen or order_params: result.append(order) params.extend(order_params) seen.add(order)
django/db/models/sql/compiler.py
ReplaceText(target='order_params' @(578,44)->(578,50))
class SQLCompiler(object): for order, order_params in ordering_group_by: # Even if we have seen the same SQL string, it might have # different params, so, we add same SQL in "has params" case. if order not in seen or params: result.append(order) params.extend(order_params) seen.add(order)
class SQLCompiler(object): for order, order_params in ordering_group_by: # Even if we have seen the same SQL string, it might have # different params, so, we add same SQL in "has params" case. if order not in seen or order_params: result.append(order) params.extend(order_params) seen.add(order)
1,075
https://:@github.com/iMerica/dj-models.git
fddb0131d37109c809ec391e1a134ef1d9e442a7
@@ -57,7 +57,7 @@ def check_password(password, encoded, setter=None, preferred='default'): must_update = hasher.algorithm != preferred.algorithm if not must_update: - must_update = hasher.must_update(encoded) + must_update = preferred.must_update(encoded) is_correct = hasher.verify(password, encoded) if setter and is_correct and must_update: setter(password)
django/contrib/auth/hashers.py
ReplaceText(target='preferred' @(60,22)->(60,28))
def check_password(password, encoded, setter=None, preferred='default'): must_update = hasher.algorithm != preferred.algorithm if not must_update: must_update = hasher.must_update(encoded) is_correct = hasher.verify(password, encoded) if setter and is_correct and must_update: setter(password)
def check_password(password, encoded, setter=None, preferred='default'): must_update = hasher.algorithm != preferred.algorithm if not must_update: must_update = preferred.must_update(encoded) is_correct = hasher.verify(password, encoded) if setter and is_correct and must_update: setter(password)
1,076
https://:@github.com/iMerica/dj-models.git
e8223b889aab3b5ac0c2312eb9ee2307ea635c97
@@ -228,7 +228,7 @@ class GenericRelationTests(TestCase): # then wrong results are produced here as the link to b will also match # (b and hs1 have equal pks). self.assertEqual(qs.count(), 1) - self.assertEqual(qs[0].links__sum, l.id) + self.assertEqual(qs[0].links__sum, hs1.id) l.delete() # Now if we don't have proper left join, we will not produce any # results at all here.
tests/generic_relations_regress/tests.py
ReplaceText(target='hs1' @(231,43)->(231,44))
class GenericRelationTests(TestCase): # then wrong results are produced here as the link to b will also match # (b and hs1 have equal pks). self.assertEqual(qs.count(), 1) self.assertEqual(qs[0].links__sum, l.id) l.delete() # Now if we don't have proper left join, we will not produce any # results at all here.
class GenericRelationTests(TestCase): # then wrong results are produced here as the link to b will also match # (b and hs1 have equal pks). self.assertEqual(qs.count(), 1) self.assertEqual(qs[0].links__sum, hs1.id) l.delete() # Now if we don't have proper left join, we will not produce any # results at all here.
1,077
https://:@github.com/iMerica/dj-models.git
3074c5b19e2da5f7a5359c3cf3c5308eb194cdf9
@@ -112,7 +112,7 @@ class ClassDecoratedTestCase(ClassDecoratedTestCaseSuper): @classmethod def setUpClass(cls): - super(cls, ClassDecoratedTestCase).setUpClass() + super(ClassDecoratedTestCase, cls).setUpClass() cls.foo = getattr(settings, 'TEST', 'BUG') def test_override(self):
tests/settings_tests/tests.py
ArgSwap(idxs=0<->1 @(115,8)->(115,13))
class ClassDecoratedTestCase(ClassDecoratedTestCaseSuper): @classmethod def setUpClass(cls): super(cls, ClassDecoratedTestCase).setUpClass() cls.foo = getattr(settings, 'TEST', 'BUG') def test_override(self):
class ClassDecoratedTestCase(ClassDecoratedTestCaseSuper): @classmethod def setUpClass(cls): super(ClassDecoratedTestCase, cls).setUpClass() cls.foo = getattr(settings, 'TEST', 'BUG') def test_override(self):
1,078
https://:@github.com/iMerica/dj-models.git
c2b4967e76fd671e6199e4dd54d2a2c1f096b8eb
@@ -23,7 +23,7 @@ def import_string(dotted_path): return getattr(module, class_name) except AttributeError: msg = 'Module "%s" does not define a "%s" attribute/class' % ( - dotted_path, class_name) + module_path, class_name) six.reraise(ImportError, ImportError(msg), sys.exc_info()[2])
django/utils/module_loading.py
ReplaceText(target='module_path' @(26,12)->(26,23))
def import_string(dotted_path): return getattr(module, class_name) except AttributeError: msg = 'Module "%s" does not define a "%s" attribute/class' % ( dotted_path, class_name) six.reraise(ImportError, ImportError(msg), sys.exc_info()[2])
def import_string(dotted_path): return getattr(module, class_name) except AttributeError: msg = 'Module "%s" does not define a "%s" attribute/class' % ( module_path, class_name) six.reraise(ImportError, ImportError(msg), sys.exc_info()[2])
1,079
https://:@github.com/iMerica/dj-models.git
a3fffdca2472885a99e1ea9159a685753cd45738
@@ -99,7 +99,7 @@ class SessionStore(SessionBase): # Remove expired sessions. expiry_age = self.get_expiry_age(expiry=self._expiry_date(session_data)) - if expiry_age < 0: + if expiry_age <= 0: session_data = {} self.delete() self.create()
django/contrib/sessions/backends/file.py
ReplaceText(target='<=' @(102,30)->(102,31))
class SessionStore(SessionBase): # Remove expired sessions. expiry_age = self.get_expiry_age(expiry=self._expiry_date(session_data)) if expiry_age < 0: session_data = {} self.delete() self.create()
class SessionStore(SessionBase): # Remove expired sessions. expiry_age = self.get_expiry_age(expiry=self._expiry_date(session_data)) if expiry_age <= 0: session_data = {} self.delete() self.create()
1,080
https://:@github.com/iMerica/dj-models.git
abcdb237bb313d116ce2ac8e90f79f61429afc70
@@ -31,7 +31,7 @@ class DatabaseCreation(BaseDatabaseCreation): try: if verbosity >= 1: print("Destroying old test database for alias %s..." % ( - self._get_database_display_str(target_database_name, verbosity), + self._get_database_display_str(verbosity, target_database_name), )) cursor.execute("DROP DATABASE %s" % qn(target_database_name)) cursor.execute("CREATE DATABASE %s" % qn(target_database_name))
django/db/backends/mysql/creation.py
ArgSwap(idxs=0<->1 @(34,28)->(34,58))
class DatabaseCreation(BaseDatabaseCreation): try: if verbosity >= 1: print("Destroying old test database for alias %s..." % ( self._get_database_display_str(target_database_name, verbosity), )) cursor.execute("DROP DATABASE %s" % qn(target_database_name)) cursor.execute("CREATE DATABASE %s" % qn(target_database_name))
class DatabaseCreation(BaseDatabaseCreation): try: if verbosity >= 1: print("Destroying old test database for alias %s..." % ( self._get_database_display_str(verbosity, target_database_name), )) cursor.execute("DROP DATABASE %s" % qn(target_database_name)) cursor.execute("CREATE DATABASE %s" % qn(target_database_name))
1,081
https://:@github.com/iMerica/dj-models.git
542b7f6c50df18f2aa201cf1de81577c1bee643c
@@ -50,7 +50,7 @@ class SeparateDatabaseAndState(Operation): to_state = base_state.clone() for dbop in self.database_operations[:-(pos + 1)]: dbop.state_forwards(app_label, to_state) - from_state = base_state.clone() + from_state = to_state.clone() database_operation.state_forwards(app_label, from_state) database_operation.database_backwards(app_label, schema_editor, from_state, to_state)
django/db/migrations/operations/special.py
ReplaceText(target='to_state' @(53,25)->(53,35))
class SeparateDatabaseAndState(Operation): to_state = base_state.clone() for dbop in self.database_operations[:-(pos + 1)]: dbop.state_forwards(app_label, to_state) from_state = base_state.clone() database_operation.state_forwards(app_label, from_state) database_operation.database_backwards(app_label, schema_editor, from_state, to_state)
class SeparateDatabaseAndState(Operation): to_state = base_state.clone() for dbop in self.database_operations[:-(pos + 1)]: dbop.state_forwards(app_label, to_state) from_state = to_state.clone() database_operation.state_forwards(app_label, from_state) database_operation.database_backwards(app_label, schema_editor, from_state, to_state)
1,082
https://:@github.com/iMerica/dj-models.git
0d9ff873d9f93efbba875efbf582db88bb0e30ce
@@ -147,7 +147,7 @@ class UserAttributeSimilarityValidator(object): continue value_parts = re.split(r'\W+', value) + [value] for value_part in value_parts: - if SequenceMatcher(a=password.lower(), b=value_part.lower()).quick_ratio() > self.max_similarity: + if SequenceMatcher(a=password.lower(), b=value_part.lower()).quick_ratio() >= self.max_similarity: try: verbose_name = force_text(user._meta.get_field(attribute_name).verbose_name) except FieldDoesNotExist:
django/contrib/auth/password_validation.py
ReplaceText(target='>=' @(150,91)->(150,92))
class UserAttributeSimilarityValidator(object): continue value_parts = re.split(r'\W+', value) + [value] for value_part in value_parts: if SequenceMatcher(a=password.lower(), b=value_part.lower()).quick_ratio() > self.max_similarity: try: verbose_name = force_text(user._meta.get_field(attribute_name).verbose_name) except FieldDoesNotExist:
class UserAttributeSimilarityValidator(object): continue value_parts = re.split(r'\W+', value) + [value] for value_part in value_parts: if SequenceMatcher(a=password.lower(), b=value_part.lower()).quick_ratio() >= self.max_similarity: try: verbose_name = force_text(user._meta.get_field(attribute_name).verbose_name) except FieldDoesNotExist:
1,083
https://:@github.com/iMerica/dj-models.git
d5088f838d837fc9e3109c828f18511055f20bea
@@ -383,7 +383,7 @@ class CombinedExpression(Expression): return DurationExpression(self.lhs, self.connector, self.rhs).as_sql(compiler, connection) if (lhs_output and rhs_output and self.connector == self.SUB and lhs_output.get_internal_type() in {'DateField', 'DateTimeField', 'TimeField'} and - lhs_output.get_internal_type() == lhs_output.get_internal_type()): + lhs_output.get_internal_type() == rhs_output.get_internal_type()): return TemporalSubtraction(self.lhs, self.rhs).as_sql(compiler, connection) expressions = [] expression_params = []
django/db/models/expressions.py
ReplaceText(target='rhs_output' @(386,50)->(386,60))
class CombinedExpression(Expression): return DurationExpression(self.lhs, self.connector, self.rhs).as_sql(compiler, connection) if (lhs_output and rhs_output and self.connector == self.SUB and lhs_output.get_internal_type() in {'DateField', 'DateTimeField', 'TimeField'} and lhs_output.get_internal_type() == lhs_output.get_internal_type()): return TemporalSubtraction(self.lhs, self.rhs).as_sql(compiler, connection) expressions = [] expression_params = []
class CombinedExpression(Expression): return DurationExpression(self.lhs, self.connector, self.rhs).as_sql(compiler, connection) if (lhs_output and rhs_output and self.connector == self.SUB and lhs_output.get_internal_type() in {'DateField', 'DateTimeField', 'TimeField'} and lhs_output.get_internal_type() == rhs_output.get_internal_type()): return TemporalSubtraction(self.lhs, self.rhs).as_sql(compiler, connection) expressions = [] expression_params = []
1,084
https://:@github.com/iMerica/dj-models.git
95993a89ce6ca5f5e26b1c22b65c57dcb8c005e9
@@ -42,7 +42,7 @@ class PasswordResetTokenGenerator: return False # Check the timestamp is within limit - if (self._num_days(self._today()) - ts) > settings.PASSWORD_RESET_TIMEOUT_DAYS: + if (self._num_days(self._today()) - ts) >= settings.PASSWORD_RESET_TIMEOUT_DAYS: return False return True
django/contrib/auth/tokens.py
ReplaceText(target='>=' @(45,48)->(45,49))
class PasswordResetTokenGenerator: return False # Check the timestamp is within limit if (self._num_days(self._today()) - ts) > settings.PASSWORD_RESET_TIMEOUT_DAYS: return False return True
class PasswordResetTokenGenerator: return False # Check the timestamp is within limit if (self._num_days(self._today()) - ts) >= settings.PASSWORD_RESET_TIMEOUT_DAYS: return False return True
1,085
https://:@github.com/Cologler/lquery-python.git
2b2d5a249fd80693660433076d8c79ef119c89bb
@@ -55,7 +55,7 @@ class DefaultExprVisitor(ExprVisitor): right = expr.right.accept(self) if left is expr.left and right is expr.right: return expr - return Make.binary_op(left, right, expr.op) + return Make.binary_op(left, expr.op, right) def visit_func_expr(self, expr): body = expr.body.accept(self)
lquery/expr/visitor.py
ArgSwap(idxs=1<->2 @(58,15)->(58,29))
class DefaultExprVisitor(ExprVisitor): right = expr.right.accept(self) if left is expr.left and right is expr.right: return expr return Make.binary_op(left, right, expr.op) def visit_func_expr(self, expr): body = expr.body.accept(self)
class DefaultExprVisitor(ExprVisitor): right = expr.right.accept(self) if left is expr.left and right is expr.right: return expr return Make.binary_op(left, expr.op, right) def visit_func_expr(self, expr): body = expr.body.accept(self)
1,086
https://:@github.com/rkhleics/police-api-client-python.git
2bedbab8eb7d2efb9ff8e39a821fd2796dd4ce3f
@@ -28,7 +28,7 @@ class BaseService(object): def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = {} - if method == 'GET': + if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs
police_api/service.py
ReplaceText(target='verb' @(31,11)->(31,17))
class BaseService(object): def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = {} if method == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs
class BaseService(object): def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = {} if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs
1,087
https://:@github.com/dw/acid.git
7cdd45e62d60ea7e67a2a3d656f8cce8dd9f5571
@@ -94,7 +94,7 @@ class RangeIterator(object): the iterator is still within the bounds of the collection prefix, otherwise False.""" keys_raw, self.data = next(self.it, ('', '')) - keys = keylib.KeyList.from_raw(self.prefix, keys_raw) + keys = keylib.KeyList.from_raw(keys_raw, self.prefix) self.keys = keys return keys is not None
acid/iterators.py
ArgSwap(idxs=0<->1 @(97,15)->(97,38))
class RangeIterator(object): the iterator is still within the bounds of the collection prefix, otherwise False.""" keys_raw, self.data = next(self.it, ('', '')) keys = keylib.KeyList.from_raw(self.prefix, keys_raw) self.keys = keys return keys is not None
class RangeIterator(object): the iterator is still within the bounds of the collection prefix, otherwise False.""" keys_raw, self.data = next(self.it, ('', '')) keys = keylib.KeyList.from_raw(keys_raw, self.prefix) self.keys = keys return keys is not None
1,088
https://:@github.com/dw/acid.git
7cdd45e62d60ea7e67a2a3d656f8cce8dd9f5571
@@ -35,7 +35,7 @@ class IterTest: REVERSE = ITEMS[::-1] def _encode(self, s): - return keylib.packs(self.prefix, s) + return keylib.packs(s, self.prefix) def setUp(self): self.e = acid.engines.ListEngine()
tests/core_test.py
ArgSwap(idxs=0<->1 @(38,15)->(38,27))
class IterTest: REVERSE = ITEMS[::-1] def _encode(self, s): return keylib.packs(self.prefix, s) def setUp(self): self.e = acid.engines.ListEngine()
class IterTest: REVERSE = ITEMS[::-1] def _encode(self, s): return keylib.packs(s, self.prefix) def setUp(self): self.e = acid.engines.ListEngine()
1,089
https://:@github.com/akuendig/RxPython.git
f1a5d48b5c22cf5d39e592299a3760be72ba79f1
@@ -50,7 +50,7 @@ class GroupBy(Producer): else: if not key in self.map: writer = Subject() - self.map[key] = value + self.map[key] = writer fireNewMapEntry = True except Exception as e: self.onError(e)
linq/groupBy.py
ReplaceText(target='writer' @(53,28)->(53,33))
class GroupBy(Producer): else: if not key in self.map: writer = Subject() self.map[key] = value fireNewMapEntry = True except Exception as e: self.onError(e)
class GroupBy(Producer): else: if not key in self.map: writer = Subject() self.map[key] = writer fireNewMapEntry = True except Exception as e: self.onError(e)
1,090
https://:@github.com/ANCIR/grano.git
0c1c013af342409e68adf8e50a80bd72f66cd9a4
@@ -20,7 +20,7 @@ def save(data, project=None): """ Create or update a project with a given slug. """ validator = ProjectValidator() - data = validator.deserialize(validator) + data = validator.deserialize(data) if project is None: project = Project()
grano/logic/projects.py
ReplaceText(target='data' @(23,33)->(23,42))
def save(data, project=None): """ Create or update a project with a given slug. """ validator = ProjectValidator() data = validator.deserialize(validator) if project is None: project = Project()
def save(data, project=None): """ Create or update a project with a given slug. """ validator = ProjectValidator() data = validator.deserialize(data) if project is None: project = Project()
1,091
https://:@github.com/ANCIR/grano.git
051a6d6191ba975e6f741c19b354a9017c825de0
@@ -50,7 +50,7 @@ def update(slug, name): authz.require(authz.project_manage(project)) schema = object_or_404(Schema.by_name(project, name)) data = request_data({'project': project}) - project = schemata.save(data, schema=schema) + schema = schemata.save(data, schema=schema) db.session.commit() return jsonify(schemata.to_rest(schema))
grano/views/schemata_api.py
ReplaceText(target='schema' @(53,4)->(53,11))
def update(slug, name): authz.require(authz.project_manage(project)) schema = object_or_404(Schema.by_name(project, name)) data = request_data({'project': project}) project = schemata.save(data, schema=schema) db.session.commit() return jsonify(schemata.to_rest(schema))
def update(slug, name): authz.require(authz.project_manage(project)) schema = object_or_404(Schema.by_name(project, name)) data = request_data({'project': project}) schema = schemata.save(data, schema=schema) db.session.commit() return jsonify(schemata.to_rest(schema))
1,092
https://:@github.com/eljost/pysisyphus.git
b5fd2ffdadc4b9d84a4b553b2e24a739509659be
@@ -100,7 +100,7 @@ class NEB(ChainOfStates): if self._forces is None: # Parallel calculation - if self.parallel != 0: + if self.parallel > 0: with Pool(processes=self.parallel) as pool: image_number = len(self.images) par_images = pool.map(self.par_calc, range(image_number))
pysisyphus/cos/NEB.py
ReplaceText(target='>' @(103,29)->(103,31))
class NEB(ChainOfStates): if self._forces is None: # Parallel calculation if self.parallel != 0: with Pool(processes=self.parallel) as pool: image_number = len(self.images) par_images = pool.map(self.par_calc, range(image_number))
class NEB(ChainOfStates): if self._forces is None: # Parallel calculation if self.parallel > 0: with Pool(processes=self.parallel) as pool: image_number = len(self.images) par_images = pool.map(self.par_calc, range(image_number))
1,093
https://:@github.com/eljost/pysisyphus.git
f2aa6358798813b4e77862d57a38ec3286f3a142
@@ -425,7 +425,7 @@ class ORCA(Calculator): self.store_wfo_data(atoms, coords) # In the first iteration we have nothing to compare to old_root = self.root - if self.calc_counter >= 1: + if self.calc_counter > 1: last_two_coords = self.wfow.last_two_coords self.root = self.wfow.track(old_root=self.root) if self.root != old_root:
pysisyphus/calculators/ORCA.py
ReplaceText(target='>' @(428,29)->(428,31))
class ORCA(Calculator): self.store_wfo_data(atoms, coords) # In the first iteration we have nothing to compare to old_root = self.root if self.calc_counter >= 1: last_two_coords = self.wfow.last_two_coords self.root = self.wfow.track(old_root=self.root) if self.root != old_root:
class ORCA(Calculator): self.store_wfo_data(atoms, coords) # In the first iteration we have nothing to compare to old_root = self.root if self.calc_counter > 1: last_two_coords = self.wfow.last_two_coords self.root = self.wfow.track(old_root=self.root) if self.root != old_root:
1,094
https://:@github.com/eljost/pysisyphus.git
f00266de85fe1696b618f979711c52bafb2312be
@@ -58,7 +58,7 @@ def get_geoms(xyz_fns, idpp=False, between=0, dump=False, multiple_geoms=False): # How is this different from above? elif isinstance(xyz_fns, str) and xyz_fns.endswith(".trj"): geoms = geoms_from_trj(xyz_fns) - elif multiple_geoms: + elif not multiple_geoms: geoms = geoms_from_trj(xyz_fns[0]) # Handle multiple .xyz files else:
pysisyphus/trj.py
ReplaceText(target='not ' @(61,9)->(61,9))
def get_geoms(xyz_fns, idpp=False, between=0, dump=False, multiple_geoms=False): # How is this different from above? elif isinstance(xyz_fns, str) and xyz_fns.endswith(".trj"): geoms = geoms_from_trj(xyz_fns) elif multiple_geoms: geoms = geoms_from_trj(xyz_fns[0]) # Handle multiple .xyz files else:
def get_geoms(xyz_fns, idpp=False, between=0, dump=False, multiple_geoms=False): # How is this different from above? elif isinstance(xyz_fns, str) and xyz_fns.endswith(".trj"): geoms = geoms_from_trj(xyz_fns) elif not multiple_geoms: geoms = geoms_from_trj(xyz_fns[0]) # Handle multiple .xyz files else:
1,095
https://:@github.com/eljost/pysisyphus.git
98f7e8f2e9a295c727cf57b2dd6bca90fd829b16
@@ -837,7 +837,7 @@ def plot_opt(h5_fn="optimization.h5", group_name="opt"): ax1.set_title("max(forces)") ax1.set_ylabel("$E_h$ Bohr⁻¹ (rad)⁻¹") - ax2.plot(max_forces, **ax_kwargs) + ax2.plot(rms_forces, **ax_kwargs) ax2.set_title("rms(forces)") ax2.set_xlabel("Step") ax2.set_ylabel("$E_h$ Bohr⁻¹ (rad)⁻¹")
pysisyphus/plot.py
ReplaceText(target='rms_forces' @(840,13)->(840,23))
def plot_opt(h5_fn="optimization.h5", group_name="opt"): ax1.set_title("max(forces)") ax1.set_ylabel("$E_h$ Bohr⁻¹ (rad)⁻¹") ax2.plot(max_forces, **ax_kwargs) ax2.set_title("rms(forces)") ax2.set_xlabel("Step") ax2.set_ylabel("$E_h$ Bohr⁻¹ (rad)⁻¹")
def plot_opt(h5_fn="optimization.h5", group_name="opt"): ax1.set_title("max(forces)") ax1.set_ylabel("$E_h$ Bohr⁻¹ (rad)⁻¹") ax2.plot(rms_forces, **ax_kwargs) ax2.set_title("rms(forces)") ax2.set_xlabel("Step") ax2.set_ylabel("$E_h$ Bohr⁻¹ (rad)⁻¹")
1,096
https://:@github.com/eljost/pysisyphus.git
2aaee0dec2e2c07da386ede88d22bd3c2b2baf5e
@@ -131,7 +131,7 @@ def run(): inp = make_input(**inp_kwargs) inp_fn = "packmol.inp" with open(inp_fn, "w") as handle: - handle.write(inp_fn) + handle.write(inp) print(f"Wrote packmol input to '{inp_fn}'") proc = call_packmol(inp)
pysisyphus/pack.py
ReplaceText(target='inp' @(134,21)->(134,27))
def run(): inp = make_input(**inp_kwargs) inp_fn = "packmol.inp" with open(inp_fn, "w") as handle: handle.write(inp_fn) print(f"Wrote packmol input to '{inp_fn}'") proc = call_packmol(inp)
def run(): inp = make_input(**inp_kwargs) inp_fn = "packmol.inp" with open(inp_fn, "w") as handle: handle.write(inp) print(f"Wrote packmol input to '{inp_fn}'") proc = call_packmol(inp)
1,097
https://:@github.com/pashango2/Image4Layer.git
24060ba346c534032314d6f28011d2edfceec1a2
@@ -191,7 +191,7 @@ def separate_blend(cb, cs, func, eval_str="func(float(a), float(b))"): # cs has alpha if cs_alpha: - base_img = img.copy() + base_img = cb.copy() base_img.paste(img, mask=cs_alpha) img = base_img
image4layer/image4layer.py
ReplaceText(target='cb' @(194,23)->(194,26))
def separate_blend(cb, cs, func, eval_str="func(float(a), float(b))"): # cs has alpha if cs_alpha: base_img = img.copy() base_img.paste(img, mask=cs_alpha) img = base_img
def separate_blend(cb, cs, func, eval_str="func(float(a), float(b))"): # cs has alpha if cs_alpha: base_img = cb.copy() base_img.paste(img, mask=cs_alpha) img = base_img
1,098
https://:@github.com/BadrYoubiIdrissi/hydra-plugins.git
d690e35407ad42bbf99d99ced25e7f3f77dae25b
@@ -46,7 +46,7 @@ class RangeSweeper(Sweeper): src_lists = [] for s in arguments: key, value = s.split("=") - gl = re.match(r'glob\((.+)\)', s) + gl = re.match(r'glob\((.+)\)', value) if ',' in value: possible_values=value.split(',') elif ':' in value:
badr_range_sweeper/hydra_plugins/range_sweeper_badr/range_sweeper_badr.py
ReplaceText(target='value' @(49,43)->(49,44))
class RangeSweeper(Sweeper): src_lists = [] for s in arguments: key, value = s.split("=") gl = re.match(r'glob\((.+)\)', s) if ',' in value: possible_values=value.split(',') elif ':' in value:
class RangeSweeper(Sweeper): src_lists = [] for s in arguments: key, value = s.split("=") gl = re.match(r'glob\((.+)\)', value) if ',' in value: possible_values=value.split(',') elif ':' in value:
1,099
https://:@github.com/annoys-parrot/multi_view_network.git
df17d237d8e022cfa14fe76cce812b3914360e20
@@ -340,7 +340,7 @@ def BuildMultiViewNetwork( [v1, v2, v3, v4], name='concatenation') fully_connected = keras.layers.Dense( units=hidden_units, name='fully_connected')(concatenation) - dropout = keras.layers.Dropout(rate=dropout_rate)(concatenation) + dropout = keras.layers.Dropout(rate=dropout_rate)(fully_connected) softmax = keras.layers.Dense( units=output_units, activation='softmax', name='softmax')(dropout)
multi_view_network/models.py
ReplaceText(target='fully_connected' @(343,54)->(343,67))
def BuildMultiViewNetwork( [v1, v2, v3, v4], name='concatenation') fully_connected = keras.layers.Dense( units=hidden_units, name='fully_connected')(concatenation) dropout = keras.layers.Dropout(rate=dropout_rate)(concatenation) softmax = keras.layers.Dense( units=output_units, activation='softmax', name='softmax')(dropout)
def BuildMultiViewNetwork( [v1, v2, v3, v4], name='concatenation') fully_connected = keras.layers.Dense( units=hidden_units, name='fully_connected')(concatenation) dropout = keras.layers.Dropout(rate=dropout_rate)(fully_connected) softmax = keras.layers.Dense( units=output_units, activation='softmax', name='softmax')(dropout)