input
stringlengths
11
7.65k
target
stringlengths
22
8.26k
def prefix(self): return True
def test_validate_question_reviewer_count_fails_for_negative_count(self): community_contribution_stats = ( suggestion_services.get_community_contribution_stats() ) community_contribution_stats.question_reviewer_count = ( self.negative_count ) with self.assertRaisesRegex( utils.ValidationError, 'Expected the question reviewer count to be non-negative, ' 'received: %s.' % ( community_contribution_stats.question_reviewer_count) ): community_contribution_stats.validate()
def prefix(self): return True
def test_validate_question_suggestion_count_fails_for_negative_count(self): community_contribution_stats = ( suggestion_services.get_community_contribution_stats() ) community_contribution_stats.question_suggestion_count = ( self.negative_count ) with self.assertRaisesRegex( utils.ValidationError, 'Expected the question suggestion count to be non-negative, ' 'received: %s.' % ( community_contribution_stats.question_suggestion_count) ): community_contribution_stats.validate()
def prefix(self): return True
def test_validate_question_reviewer_count_fails_for_non_integer_count( self): community_contribution_stats = ( suggestion_services.get_community_contribution_stats() ) community_contribution_stats.question_reviewer_count = ( self.non_integer_count ) with self.assertRaisesRegex( utils.ValidationError, 'Expected the question reviewer count to be an integer, ' 'received: %s.' % ( community_contribution_stats.question_reviewer_count) ): community_contribution_stats.validate()
def prefix(self): return True
def test_validate_question_suggestion_count_fails_for_non_integer_count( self): community_contribution_stats = ( suggestion_services.get_community_contribution_stats() ) community_contribution_stats.question_suggestion_count = ( self.non_integer_count ) with self.assertRaisesRegex( utils.ValidationError, 'Expected the question suggestion count to be an integer, ' 'received: %s.' % ( community_contribution_stats.question_suggestion_count) ): community_contribution_stats.validate()
def prefix(self): return True
def main(unused_argv): tpu_cluster_resolver = contrib_cluster_resolver.TPUClusterResolver( FLAGS.tpu, zone=FLAGS.tpu_zone, project=FLAGS.gcp_project) config = contrib_tpu.RunConfig( cluster=tpu_cluster_resolver, model_dir=FLAGS.model_dir, save_checkpoints_steps=FLAGS.iterations_per_loop, keep_checkpoint_max=None, tpu_config=contrib_tpu.TPUConfig( iterations_per_loop=FLAGS.iterations_per_loop, num_shards=FLAGS.num_cores, per_host_input_for_training=contrib_tpu.InputPipelineConfig.PER_HOST_V2)) # pylint: disable=line-too-long # Input pipelines are slightly different (with regards to shuffling and # preprocessing) between training and evaluation. imagenet_train = imagenet_input.ImageNetInput( is_training=True, data_dir=FLAGS.data_dir, use_bfloat16=True, transpose_input=FLAGS.transpose_input) imagenet_eval = imagenet_input.ImageNetInput( is_training=False, data_dir=FLAGS.data_dir, use_bfloat16=True, transpose_input=FLAGS.transpose_input) if FLAGS.use_fast_lr: resnet_main.LR_SCHEDULE = [ # (multiplier, epoch to start) tuples (1.0, 4), (0.1, 21), (0.01, 35), (0.001, 43) ] imagenet_train_small = imagenet_input.ImageNetInput( is_training=True, image_size=128, data_dir=FLAGS.data_dir_small, num_parallel_calls=FLAGS.num_parallel_calls, use_bfloat16=True, transpose_input=FLAGS.transpose_input, cache=True) imagenet_eval_small = imagenet_input.ImageNetInput( is_training=False, image_size=128, data_dir=FLAGS.data_dir_small, num_parallel_calls=FLAGS.num_parallel_calls, use_bfloat16=True, transpose_input=FLAGS.transpose_input, cache=True) imagenet_train_large = imagenet_input.ImageNetInput( is_training=True, image_size=288, data_dir=FLAGS.data_dir, num_parallel_calls=FLAGS.num_parallel_calls, use_bfloat16=True, transpose_input=FLAGS.transpose_input) imagenet_eval_large = imagenet_input.ImageNetInput( is_training=False, image_size=288, data_dir=FLAGS.data_dir, num_parallel_calls=FLAGS.num_parallel_calls, use_bfloat16=True, transpose_input=FLAGS.transpose_input) resnet_classifier = contrib_tpu.TPUEstimator( use_tpu=FLAGS.use_tpu, model_fn=resnet_main.resnet_model_fn, config=config, train_batch_size=FLAGS.train_batch_size, eval_batch_size=FLAGS.eval_batch_size) if FLAGS.mode == 'train': current_step = estimator._load_global_step_from_checkpoint_dir(FLAGS.model_dir) # pylint: disable=protected-access,line-too-long batches_per_epoch = NUM_TRAIN_IMAGES / FLAGS.train_batch_size tf.logging.info('Training for %d steps (%.2f epochs in total). Current' ' step %d.' % (FLAGS.train_steps, FLAGS.train_steps / batches_per_epoch, current_step)) start_timestamp = time.time() # This time will include compilation time # Write a dummy file at the start of training so that we can measure the # runtime at each checkpoint from the file write time. tf.gfile.MkDir(FLAGS.model_dir) if not tf.gfile.Exists(os.path.join(FLAGS.model_dir, 'START')): with tf.gfile.GFile(os.path.join(FLAGS.model_dir, 'START'), 'w') as f: f.write(str(start_timestamp)) if FLAGS.use_fast_lr: small_steps = int(18 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size) normal_steps = int(41 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size) large_steps = int(min(50 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size, FLAGS.train_steps)) resnet_classifier.train( input_fn=imagenet_train_small.input_fn, max_steps=small_steps) resnet_classifier.train( input_fn=imagenet_train.input_fn, max_steps=normal_steps) resnet_classifier.train( input_fn=imagenet_train_large.input_fn, max_steps=large_steps) else: resnet_classifier.train( input_fn=imagenet_train.input_fn, max_steps=FLAGS.train_steps) else: assert FLAGS.mode == 'eval' start_timestamp = tf.gfile.Stat( os.path.join(FLAGS.model_dir, 'START')).mtime_nsec results = [] eval_steps = NUM_EVAL_IMAGES // FLAGS.eval_batch_size ckpt_steps = set() all_files = tf.gfile.ListDirectory(FLAGS.model_dir) for f in all_files: mat = re.match(CKPT_PATTERN, f) if mat is not None: ckpt_steps.add(int(mat.group('gs'))) ckpt_steps = sorted(list(ckpt_steps)) tf.logging.info('Steps to be evaluated: %s' % str(ckpt_steps)) for step in ckpt_steps: ckpt = os.path.join(FLAGS.model_dir, 'model.ckpt-%d' % step) batches_per_epoch = NUM_TRAIN_IMAGES // FLAGS.train_batch_size current_epoch = step // batches_per_epoch if FLAGS.use_fast_lr: if current_epoch < 18: eval_input_fn = imagenet_eval_small.input_fn if current_epoch >= 18 and current_epoch < 41: eval_input_fn = imagenet_eval.input_fn if current_epoch >= 41: # 41: eval_input_fn = imagenet_eval_large.input_fn else: eval_input_fn = imagenet_eval.input_fn end_timestamp = tf.gfile.Stat(ckpt + '.index').mtime_nsec elapsed_hours = (end_timestamp - start_timestamp) / (1e9 * 3600.0) tf.logging.info('Starting to evaluate.') eval_start = time.time() # This time will include compilation time eval_results = resnet_classifier.evaluate( input_fn=eval_input_fn, steps=eval_steps, checkpoint_path=ckpt) eval_time = int(time.time() - eval_start) tf.logging.info('Eval results: %s. Elapsed seconds: %d' % (eval_results, eval_time)) results.append([ current_epoch, elapsed_hours, '%.2f' % (eval_results['top_1_accuracy'] * 100), '%.2f' % (eval_results['top_5_accuracy'] * 100), ]) time.sleep(60) with tf.gfile.GFile(os.path.join(FLAGS.model_dir, 'results.tsv'), 'wb') as tsv_file: # pylint: disable=line-too-long writer = csv.writer(tsv_file, delimiter='\t') writer.writerow(['epoch', 'hours', 'top1Accuracy', 'top5Accuracy']) writer.writerows(results)
def prefix(self): return True
def test_telemetry_finish(runner, live_mock_server, parse_ctx): with runner.isolated_filesystem(): run = wandb.init() run.finish() ctx_util = parse_ctx(live_mock_server.get_ctx()) telemetry = ctx_util.telemetry assert telemetry and 2 in telemetry.get("3", [])
def prefix(self): return True
def allowed_file(filename): return '.' in filename and \ filename.rsplit('.', 1)[1] in app.config.get('ALLOWED_EXTENSIONS')
def prefix(self): return True
def test_telemetry_imports_hf(runner, live_mock_server, parse_ctx): with runner.isolated_filesystem(): run = wandb.init() with mock.patch.dict("sys.modules", {"transformers": mock.Mock()}): import transformers run.finish() ctx_util = parse_ctx(live_mock_server.get_ctx()) telemetry = ctx_util.telemetry # hf in finish modules but not in init modules assert telemetry and 11 not in telemetry.get("1", []) assert telemetry and 11 in telemetry.get("2", [])
def prefix(self): return True
def unique_name(): now_time = datetime.datetime.now().strftime("%Y%m%d%H%M%S") random_num = random.randint(0, 100) if random_num <= 10: random_num = str(0) + str(random_num) unique_num = str(now_time) + str(random_num) return unique_num
def prefix(self): return True
def test_telemetry_imports_catboost(runner, live_mock_server, parse_ctx): with runner.isolated_filesystem(): with mock.patch.dict("sys.modules", {"catboost": mock.Mock()}): import catboost run = wandb.init() run.finish() ctx_util = parse_ctx(live_mock_server.get_ctx()) telemetry = ctx_util.telemetry # catboost in both init and finish modules assert telemetry and 7 in telemetry.get("1", []) assert telemetry and 7 in telemetry.get("2", [])
def prefix(self): return True
def image_thumbnail(filename): filepath = os.path.join(app.config.get('UPLOAD_FOLDER'), filename) im = Image.open(filepath) w, h = im.size if w > h: im.thumbnail((106, 106*h/w)) else: im.thumbnail((106*w/h, 106)) im.save(os.path.join(app.config.get('UPLOAD_FOLDER'), os.path.splitext(filename)[0] + '_thumbnail' + os.path.splitext(filename)[1]))
def prefix(self): return True
def test_telemetry_imports_jax(runner, live_mock_server, parse_ctx): with runner.isolated_filesystem(): import jax wandb.init() wandb.finish() ctx_util = parse_ctx(live_mock_server.get_ctx()) telemetry = ctx_util.telemetry # jax in finish modules but not in init modules assert telemetry and 12 in telemetry.get("1", []) assert telemetry and 12 in telemetry.get("2", [])
def prefix(self): return True
def image_delete(filename): thumbnail_filepath = os.path.join(app.config.get('UPLOAD_FOLDER'), filename) filepath = thumbnail_filepath.replace('_thumbnail', '') os.remove(filepath) os.remove(thumbnail_filepath)
def prefix(self): return True
def test_telemetry_run_organizing_init(runner, live_mock_server, parse_ctx): with runner.isolated_filesystem(): wandb.init(name="test_name", tags=["my-tag"], config={"abc": 123}, id="mynewid") wandb.finish() ctx_util = parse_ctx(live_mock_server.get_ctx()) telemetry = ctx_util.telemetry assert telemetry and 13 in telemetry.get("3", []) # name assert telemetry and 14 in telemetry.get("3", []) # id assert telemetry and 15 in telemetry.get("3", []) # tags assert telemetry and 16 in telemetry.get("3", []) # config
def prefix(self): return True
def _rematcher(regex): """compile the regexp with the best available regexp engine and return a matcher function""" m = util.re.compile(regex) try: # slightly faster, provided by facebook's re2 bindings return m.test_match except AttributeError: return m.match
def prefix(self): return True
def refresh(): win.Refresh() print type(win)
def prefix(self): return True
def _expandsets(kindpats, ctx): """Returns the kindpats list with the 'set' patterns expanded.""" fset = set() other = [] for kind, pat, source in kindpats: if kind == "set": if not ctx: raise error.ProgrammingError("fileset expression with no " "context") s = ctx.getfileset(pat) fset.update(s) continue other.append((kind, pat, source)) return fset, other
def prefix(self): return True
def create_texture(): global rgbtex rgbtex = glGenTextures(1) glBindTexture(TEXTURE_TARGET, rgbtex) glTexImage2D(TEXTURE_TARGET,0,GL_RGB,640,480,0,GL_RGB,GL_UNSIGNED_BYTE,None)
def prefix(self): return True
def _expandsubinclude(kindpats, root): """Returns the list of subinclude matcher args and the kindpats without the subincludes in it.""" relmatchers = [] other = [] for kind, pat, source in kindpats: if kind == "subinclude": sourceroot = pathutil.dirname(util.normpath(source)) pat = util.pconvert(pat) path = pathutil.join(sourceroot, pat) newroot = pathutil.dirname(path) matcherargs = (newroot, "", [], ["include:%s" % path]) prefix = pathutil.canonpath(root, root, newroot) if prefix: prefix += "/" relmatchers.append((prefix, matcherargs)) else: other.append((kind, pat, source)) return relmatchers, other
def prefix(self): return True
def EVT_LEFT_DOWN(event): global _mpos _mpos = event.Position
def prefix(self): return True
def _kindpatsalwaysmatch(kindpats): """ "Checks whether the kindspats match everything, as e.g. 'relpath:.' does. """ for kind, pat, source in kindpats: # TODO: update me? if pat != "" or kind not in ["relpath", "glob"]: return False return True
def prefix(self): return True
def EVT_LEFT_UP(event): global _mpos _mpos = None
def prefix(self): return True
def normalize(patterns, default, root, cwd, auditor, warn): kp = _donormalize(patterns, default, root, cwd, auditor, warn) kindpats = [] for kind, pats, source in kp: if kind not in ("re", "relre"): # regex can't be normalized p = pats pats = dsnormalize(pats) # Preserve the original to handle a case only rename. if p != pats and p in dirstate: kindpats.append((kind, p, source)) kindpats.append((kind, pats, source)) return kindpats
def prefix(self): return True
def EVT_MOTION(event): global _mpos if event.LeftIsDown(): if _mpos: (x,y),(mx,my) = event.Position,_mpos rotangles[0] += y-my rotangles[1] += x-mx refresh() _mpos = event.Position
def prefix(self): return True
def exact(root, cwd, files, badfn=None): return exactmatcher(root, cwd, files, badfn=badfn)
def prefix(self): return True
def EVT_MOUSEWHEEL(event): global zoomdist dy = event.WheelRotation zoomdist *= np.power(0.95, -dy) refresh()
def prefix(self): return True
def always(root, cwd): return alwaysmatcher(root, cwd)
def prefix(self): return True
def mouse_rotate(xAngle, yAngle, zAngle): glRotatef(xAngle, 1.0, 0.0, 0.0); glRotatef(yAngle, 0.0, 1.0, 0.0); glRotatef(zAngle, 0.0, 0.0, 1.0);
def prefix(self): return True
def never(root, cwd): return nevermatcher(root, cwd)
def prefix(self): return True
def playcolors(): while 1: global clearcolor clearcolor = [np.random.random(),0,0,0] time.sleep(0.1) refresh()
def prefix(self): return True
def union(matches, root, cwd): """Union a list of matchers. If the list is empty, return nevermatcher. If the list only contains one non-None value, return that matcher. Otherwise return a union matcher. """ matches = list(filter(None, matches)) if len(matches) == 0: return nevermatcher(root, cwd) elif len(matches) == 1: return matches[0] else: return unionmatcher(matches)
def prefix(self): return True
def update(dt=0): global projpts, rgb, depth depth,_ = freenect.sync_get_depth() rgb,_ = freenect.sync_get_video() q = depth X,Y = np.meshgrid(range(640),range(480)) # YOU CAN CHANGE THIS AND RERUN THE PROGRAM! # Point cloud downsampling d = 4 projpts = calibkinect.depth2xyzuv(q[::d,::d],X[::d,::d],Y[::d,::d]) refresh()
def prefix(self): return True
def badmatch(match, badfn): """Make a copy of the given matcher, replacing its bad method with the given one. """ m = copy.copy(match) m.bad = badfn return m
def prefix(self): return True
def update_join(): update_on() try: _thread.join() except: update_off()
def prefix(self): return True
def _donormalize(patterns, default, root, cwd, auditor, warn): """Convert 'kind:pat' from the patterns list to tuples with kind and normalized and rooted patterns and with listfiles expanded.""" kindpats = [] for kind, pat in [_patsplit(p, default) for p in patterns]: if kind in cwdrelativepatternkinds: pat = pathutil.canonpath(root, cwd, pat, auditor) elif kind in ("relglob", "path", "rootfilesin"): pat = util.normpath(pat) elif kind in ("listfile", "listfile0"): try: files = decodeutf8(util.readfile(pat)) if kind == "listfile0": files = files.split("\0") else: files = files.splitlines() files = [f for f in files if f] except EnvironmentError: raise error.Abort(_("unable to read file list (%s)") % pat) for k, p, source in _donormalize(files, default, root, cwd, auditor, warn): kindpats.append((k, p, pat)) continue elif kind == "include": try: fullpath = os.path.join(root, util.localpath(pat)) includepats = readpatternfile(fullpath, warn) for k, p, source in _donormalize( includepats, default, root, cwd, auditor, warn ): kindpats.append((k, p, source or pat)) except error.Abort as inst: raise error.Abort("%s: %s" % (pat, inst[0])) except IOError as inst: if warn: warn( _("skipping unreadable pattern file '%s': %s\n") % (pat, inst.strerror) ) continue # else: re or relre - which cannot be normalized kindpats.append((kind, pat, "")) return kindpats
def prefix(self): return True
def update_on(): global _updating if not '_updating' in globals(): _updating = False if _updating: return
def prefix(self): return True
def _testrefastpath(repat): """Test if a re pattern can use fast path. That is, for every "$A/$B" path the pattern matches, "$A" must also be matched, Return True if we're sure it is. Return False otherwise. """ # XXX: It's very hard to implement this. These are what need to be # supported in production and tests. Very hacky. But we plan to get rid # of re matchers eventually. # Rules like "(?!experimental/)" if repat.startswith("(?!") and repat.endswith(")") and repat.count(")") == 1: return True # Rules used in doctest if repat == "(i|j)$": return True return False
def prefix(self): return True
def _run(): while _updating: update()
def prefix(self): return True
def _globpatsplit(pat): """Split a glob pattern. Return a list. A naive version is "path.split("/")". This function handles more cases, like "{*,{a,b}*/*}". >>> _globpatsplit("*/**/x/{a,b/c}") ['*', '**', 'x', '{a,b/c}'] """ result = [] buf = "" parentheses = 0 for ch in pat: if ch == "{": parentheses += 1 elif ch == "}": parentheses -= 1 if parentheses == 0 and ch == "/": if buf: result.append(buf) buf = "" else: buf += ch if buf: result.append(buf) return result
def prefix(self): return True
def update_off(): global _updating _updating = False
def prefix(self): return True
def __init__(self, *args, **kwargs): # If True, avoid entering subdirectories, and match everything recursively, # unconditionally. self.matchrecursive = False # If True, avoid entering subdirectories, and return "unsure" for # everything. This is set to True when complex re patterns (potentially # including "/") are used. self.unsurerecursive = False # Patterns for matching paths in this directory. self._kindpats = [] # Glob patterns used to match parent directories of another glob # pattern. self._globdirpats = [] super(_tree, self).__init__(*args, **kwargs)
def prefix(self): return True
def loopcv(): import cv while 1: cv.ShowImage('hi',get_depth().astype(np.uint8)) cv.WaitKey(10)
def prefix(self): return True
def insert(self, path, matchrecursive=True, globpats=None, repats=None): """Insert a directory path to this tree. If matchrecursive is True, mark the directory as unconditionally include files and subdirs recursively. If globpats or repats are specified, append them to the patterns being applied at this directory. The tricky part is those patterns can match "x/y/z" and visit("x"), visit("x/y") need to return True, while we still want visit("x/a") to return False. """ if path == "": self.matchrecursive |= matchrecursive if globpats: # Need to match parent directories too. for pat in globpats: components = _globpatsplit(pat) parentpat = "" for comp in components: if parentpat: parentpat += "/" parentpat += comp if "/" in comp: # Giving up - fallback to slow paths. self.unsurerecursive = True self._globdirpats.append(parentpat) if any("**" in p for p in globpats): # Giving up - "**" matches paths including "/" self.unsurerecursive = True self._kindpats += [("glob", pat, "") for pat in globpats] if repats: if not all(map(_testrefastpath, repats)): # Giving up - fallback to slow paths. self.unsurerecursive = True self._kindpats += [("re", pat, "") for pat in repats] return subdir, rest = self._split(path) self.setdefault(subdir, _tree()).insert(rest, matchrecursive, globpats, repats)
def prefix(self): return True
def visitdir(self, path): """Similar to matcher.visitdir""" path = normalizerootdir(path, "visitdir") if self.matchrecursive: return "all" elif self.unsurerecursive: return True elif path == "": return True if self._kindpats and self._compiledpats(path): # XXX: This is incorrect. But re patterns are already used in # production. We should kill them! # Need to test "if every string starting with 'path' matches". # Obviously it's impossible to test *every* string with the # standard regex API, therefore pick a random strange path to test # it approximately. if self._compiledpats("%s/*/_/-/0/*" % path): return "all" else: return True if self._globdirpats and self._compileddirpats(path): return True subdir, rest = self._split(path) subtree = self.get(subdir) if subtree is None: return False else: return subtree.visitdir(rest)
def prefix(self): return True
def _compiledpats(self): pat, matchfunc = _buildregexmatch(self._kindpats, "") return matchfunc
def prefix(self): return True
def _compileddirpats(self): pat, matchfunc = _buildregexmatch( [("glob", p, "") for p in self._globdirpats], "$" ) return matchfunc
def prefix(self): return True
def _split(self, path): if "/" in path: subdir, rest = path.split("/", 1) else: subdir, rest = path, "" if not subdir: raise error.ProgrammingError("path cannot be absolute") return subdir, rest
def prefix(self): return True
def _remainingpats(pat, prefix): """list of patterns with prefix stripped >>> _remainingpats("a/b/c", "") ['a/b/c'] >>> _remainingpats("a/b/c", "a") ['b/c'] >>> _remainingpats("a/b/c", "a/b") ['c'] >>> _remainingpats("a/b/c", "a/b/c") [] >>> _remainingpats("", "") [] """ if prefix: if prefix == pat: return [] else: assert pat[len(prefix)] == "/" return [pat[len(prefix) + 1 :]] else: if pat: return [pat] else: return []
def prefix(self): return True
def _buildvisitdir(kindpats): """Try to build an efficient visitdir function Return a visitdir function if it's built. Otherwise return None if there are unsupported patterns. >>> _buildvisitdir([('include', 'foo', '')]) >>> _buildvisitdir([('relglob', 'foo', '')]) >>> t = _buildvisitdir([ ... ('glob', 'a/b', ''), ... ('glob', 'c/*.d', ''), ... ('glob', 'e/**/*.c', ''), ... ('re', '^f/(?!g)', ''), # no "$", only match prefix ... ('re', '^h/(i|j)$', ''), ... ('glob', 'i/a*/b*/c*', ''), ... ('glob', 'i/a5/b7/d', ''), ... ('glob', 'j/**.c', ''), ... ]) >>> t('a') True >>> t('a/b') 'all' >>> t('a/b/c') 'all' >>> t('c') True >>> t('c/d') False >>> t('c/rc.d') 'all' >>> t('c/rc.d/foo') 'all' >>> t('e') True >>> t('e/a') True >>> t('e/a/b.c') True >>> t('e/a/b.d') True >>> t('f') True >>> t('f/g') False >>> t('f/g2') False >>> t('f/g/a') False >>> t('f/h') 'all' >>> t('f/h/i') 'all' >>> t('h/i') True >>> t('h/i/k') False >>> t('h/k') False >>> t('i') True >>> t('i/a1') True >>> t('i/b2') False >>> t('i/a/b2/c3') 'all' >>> t('i/a/b2/d4') False >>> t('i/a5/b7/d') 'all' >>> t('j/x/y') True >>> t('z') False """ tree = _tree() for kind, pat, _source in kindpats: if kind == "glob": components = [] for p in pat.split("/"): if "[" in p or "{" in p or "*" in p or "?" in p: break components.append(p) prefix = "/".join(components) matchrecursive = prefix == pat tree.insert( prefix, matchrecursive=matchrecursive, globpats=_remainingpats(pat, prefix), ) elif kind == "re": # Still try to get a plain prefix from the regular expression so we # can still have fast paths. if pat.startswith("^"): # "re" already matches from the beginning, unlike "relre" pat = pat[1:] components = [] for p in pat.split("/"): if re.escape(p) != p: # contains special characters break components.append(p) prefix = "/".join(components) tree.insert( prefix, matchrecursive=False, repats=_remainingpats(pat, prefix) ) else: # Unsupported kind return None return tree.visitdir
def prefix(self): return True
def __init__(self, root, cwd, badfn=None, relativeuipath=True): self._root = root self._cwd = cwd if badfn is not None: self.bad = badfn self._relativeuipath = relativeuipath
def prefix(self): return True
def __repr__(self): return "<%s>" % self.__class__.__name__
def prefix(self): return True
def __call__(self, fn): return self.matchfn(fn)
def prefix(self): return True
def __iter__(self): for f in self._files: yield f
def prefix(self): return True
def bad(self, f, msg): """Callback from dirstate.walk for each explicit file that can't be found/accessed, with an error message."""
def prefix(self): return True
def abs(self, f): """Convert a repo path back to path that is relative to the root of the matcher.""" return f
def prefix(self): return True
def rel(self, f): """Convert repo path back to path that is relative to cwd of matcher.""" return util.pathto(self._root, self._cwd, f)
def prefix(self): return True
def uipath(self, f): """Convert repo path to a display path. If patterns or -I/-X were used to create this matcher, the display path will be relative to cwd. Otherwise it is relative to the root of the repo.""" return (self._relativeuipath and self.rel(f)) or self.abs(f)
def prefix(self): return True
def files(self): """Explicitly listed files or patterns or roots: if no patterns or .always(): empty list, if exact: list exact files, if not .anypats(): list all files and dirs, else: optimal roots""" return self._files
def prefix(self): return True
def _fileset(self): return set(self._files)
def prefix(self): return True
def exact(self, f): """Returns True if f is in .files().""" return f in self._fileset
def prefix(self): return True
def matchfn(self, f): return False
def prefix(self): return True
def visitdir(self, dir): """Decides whether a directory should be visited based on whether it has potential matches in it or one of its subdirectories. This is based on the match's primary, included, and excluded patterns. Returns the string 'all' if the given directory and all subdirectories should be visited. Otherwise returns True or False indicating whether the given directory should be visited. """ return True
def prefix(self): return True
def always(self): """Matcher will match everything and .files() will be empty -- optimization might be possible.""" return False
def prefix(self): return True
def isexact(self): """Matcher will match exactly the list of files in .files() -- optimization might be possible.""" return False
def prefix(self): return True
def prefix(self): """Matcher will match the paths in .files() recursively -- optimization might be possible.""" return False
def prefix(self): return True
def anypats(self): """None of .always(), .isexact(), and .prefix() is true -- optimizations will be difficult.""" return not self.always() and not self.isexact() and not self.prefix()
def prefix(self): return True
def __init__(self, root, cwd, badfn=None, relativeuipath=False): super(alwaysmatcher, self).__init__( root, cwd, badfn, relativeuipath=relativeuipath )
def prefix(self): return True
def always(self): return True
def prefix(self): return True
def matchfn(self, f): return True
def prefix(self): return True
def visitdir(self, dir): return "all"
def prefix(self): return True
def __repr__(self): return "<alwaysmatcher>"
def prefix(self): return True
def __init__(self, root, cwd, badfn=None): super(nevermatcher, self).__init__(root, cwd, badfn)
def prefix(self): return True
def isexact(self): return True
def prefix(self): return True
def visitdir(self, dir): return False
def prefix(self): return True
def __repr__(self): return "<nevermatcher>"
def prefix(self): return True
def __init__(self, root, cwd, badfn=None, gitignorepaths=None): super(gitignorematcher, self).__init__(root, cwd, badfn) gitignorepaths = gitignorepaths or [] self._matcher = pathmatcher.gitignorematcher(root, gitignorepaths)
def prefix(self): return True
def matchfn(self, f): # XXX: is_dir is set to True here for performance. # It should be set to whether "f" is actually a directory or not. return self._matcher.match_relative(f, True)
def prefix(self): return True
def explain(self, f): return self._matcher.explain(f, True)
def prefix(self): return True
def visitdir(self, dir): dir = normalizerootdir(dir, "visitdir") matched = self._matcher.match_relative(dir, True) if matched: # Everything in the directory is selected (ignored) return "all" else: # Not sure return True
def prefix(self): return True
def __repr__(self): return "<gitignorematcher>"
def prefix(self): return True
def __init__(self, root, cwd, badfn=None, rules=[]): super(treematcher, self).__init__(root, cwd, badfn) rules = list(rules) self._matcher = pathmatcher.treematcher(rules) self._rules = rules
def prefix(self): return True
def matchfn(self, f): return self._matcher.matches(f)
def prefix(self): return True
def visitdir(self, dir): matched = self._matcher.match_recursive(dir) if matched is None: return True elif matched is True: return "all" else: assert matched is False return False
def prefix(self): return True
def __repr__(self): return "<treematcher rules=%r>" % self._rules
def prefix(self): return True
def normalizerootdir(dir, funcname): if dir == ".": util.nouideprecwarn( "match.%s() no longer accepts '.', use '' instead." % funcname, "20190805" ) return "" return dir
def prefix(self): return True
def _kindpatstoglobs(kindpats, recursive=False): """Attempt to convert 'kindpats' to glob patterns that can be used in a treematcher. kindpats should be already normalized to be relative to repo root. If recursive is True, `glob:a*` will match both `a1/b` and `a1`, otherwise `glob:a*` will only match `a1` but not `a1/b`. Return None if there are unsupported patterns (ex. regular expressions). """ if not _usetreematcher: return None globs = [] for kindpat in kindpats: kind, pat = kindpat[0:2] if kind == "re": # Attempt to convert the re pat to globs reglobs = _convertretoglobs(pat) if reglobs is not None: globs += reglobs else: return None elif kind == "glob": # The treematcher (man gitignore) does not support csh-style # brackets (ex. "{a,b,c}"). Expand the brackets to patterns. for subpat in pathmatcher.expandcurlybrackets(pat): normalized = pathmatcher.normalizeglob(subpat) if recursive: normalized = _makeglobrecursive(normalized) globs.append(normalized) elif kind == "path": if pat == ".": # Special case. Comes from `util.normpath`. pat = "" else: pat = pathmatcher.plaintoglob(pat) pat = _makeglobrecursive(pat) globs.append(pat) else: return None return globs
def prefix(self): return True
def _makeglobrecursive(pat): """Make a glob pattern recursive by appending "/**" to it""" if pat.endswith("/") or not pat: return pat + "**" else: return pat + "/**"
def prefix(self): return True
def _convertretoglobs(repat): """Attempt to convert a regular expression pattern to glob patterns. A single regular expression pattern might be converted into multiple glob patterns. Return None if conversion is unsupported. >>> _convertretoglobs("abc*") is None True >>> _convertretoglobs("xx/yy/(?!zz/kk)") ['xx/yy/**', '!xx/yy/zz/kk/**'] >>> _convertretoglobs("x/y/(?:.*/)?BUCK") ['x/y/**/BUCK'] """ m = _repat1.match(repat) if m: prefix, excluded = m.groups() return ["%s/**" % prefix, "!%s/%s/**" % (prefix, excluded)] m = _repat2.match(repat) if m: prefix, name = m.groups() return ["%s/**/%s" % (prefix, name)] return None
def prefix(self): return True
def __init__(self, root, cwd, kindpats, ctx=None, badfn=None): super(patternmatcher, self).__init__(root, cwd, badfn) # kindpats are already normalized to be relative to repo-root. # Can we use tree matcher? rules = _kindpatstoglobs(kindpats, recursive=False) fallback = True if rules is not None: try: matcher = treematcher(root, cwd, badfn=badfn, rules=rules) # Replace self to 'matcher'. self.__dict__ = matcher.__dict__ self.__class__ = matcher.__class__ fallback = False except ValueError: # for example, Regex("Compiled regex exceeds size limit of 10485760 bytes.") pass if fallback: self._prefix = _prefix(kindpats) self._pats, self.matchfn = _buildmatch(ctx, kindpats, "$", root) self._files = _explicitfiles(kindpats)
def prefix(self): return True
def _dirs(self): return set(util.dirs(self._fileset))
def prefix(self): return True
def visitdir(self, dir): dir = normalizerootdir(dir, "visitdir") if self._prefix and dir in self._fileset: return "all" if not self._prefix: return True return ( dir in self._fileset or dir in self._dirs or any(parentdir in self._fileset for parentdir in util.finddirs(dir)) )
def prefix(self): return True
def prefix(self): return self._prefix
def prefix(self): return True
def __repr__(self): return "<patternmatcher patterns=%r>" % self._pats
def prefix(self): return True
def __init__(self, root, cwd, kindpats, ctx=None, badfn=None): super(includematcher, self).__init__(root, cwd, badfn) # Can we use tree matcher? rules = _kindpatstoglobs(kindpats, recursive=True) fallback = True if rules is not None: try: matcher = treematcher(root, cwd, badfn=badfn, rules=rules) # Replace self to 'matcher'. self.__dict__ = matcher.__dict__ self.__class__ = matcher.__class__ fallback = False except ValueError: # for example, Regex("Compiled regex exceeds size limit of 10485760 bytes.") pass if fallback: self._pats, self.matchfn = _buildmatch(ctx, kindpats, "(?:/|$)", root) # prefix is True if all patterns are recursive, so certain fast paths # can be enabled. Unfortunately, it's too easy to break it (ex. by # using "glob:*.c", "re:...", etc). self._prefix = _prefix(kindpats) roots, dirs = _rootsanddirs(kindpats) # roots are directories which are recursively included. # If self._prefix is True, then _roots can have a fast path for # visitdir to return "all", marking things included unconditionally. # If self._prefix is False, then that optimization is unsound because # "roots" might contain entries that is not recursive (ex. roots will # include "foo/bar" for pattern "glob:foo/bar/*.c"). self._roots = set(roots) # dirs are directories which are non-recursively included. # That is, files under that directory are included. But not # subdirectories. self._dirs = set(dirs) # Try to use a more efficient visitdir implementation visitdir = _buildvisitdir(kindpats) if visitdir: self.visitdir = visitdir
def prefix(self): return True
def visitdir(self, dir): dir = normalizerootdir(dir, "visitdir") if self._prefix and dir in self._roots: return "all" return ( dir in self._roots or dir in self._dirs or any(parentdir in self._roots for parentdir in util.finddirs(dir)) )
def prefix(self): return True
def __repr__(self): return "<includematcher includes=%r>" % self._pats
def prefix(self): return True
def __init__(self, root, cwd, files, badfn=None): super(exactmatcher, self).__init__(root, cwd, badfn) if isinstance(files, list): self._files = files else: self._files = list(files)
def prefix(self): return True
def visitdir(self, dir): dir = normalizerootdir(dir, "visitdir") return dir in self._dirs
def prefix(self): return True
def __repr__(self): return "<exactmatcher files=%r>" % self._files
def prefix(self): return True
def __init__(self, m1, m2): super(differencematcher, self).__init__(m1._root, m1._cwd) self._m1 = m1 self._m2 = m2 self.bad = m1.bad self.traversedir = m1.traversedir