input stringlengths 11 7.65k | target stringlengths 22 8.26k |
|---|---|
async def open(self): pass | def test_validate_question_suggestion_count_fails_for_negative_count(self):
community_contribution_stats = (
suggestion_services.get_community_contribution_stats()
)
community_contribution_stats.question_suggestion_count = (
self.negative_count
)
with self.assertRaisesRegex(
utils.ValidationError,
'Expected the question suggestion count to be non-negative, '
'received: %s.' % (
community_contribution_stats.question_suggestion_count)
):
community_contribution_stats.validate() |
async def open(self): pass | def test_validate_question_reviewer_count_fails_for_non_integer_count(
self):
community_contribution_stats = (
suggestion_services.get_community_contribution_stats()
)
community_contribution_stats.question_reviewer_count = (
self.non_integer_count
)
with self.assertRaisesRegex(
utils.ValidationError,
'Expected the question reviewer count to be an integer, '
'received: %s.' % (
community_contribution_stats.question_reviewer_count)
):
community_contribution_stats.validate() |
async def open(self): pass | def test_validate_question_suggestion_count_fails_for_non_integer_count(
self):
community_contribution_stats = (
suggestion_services.get_community_contribution_stats()
)
community_contribution_stats.question_suggestion_count = (
self.non_integer_count
)
with self.assertRaisesRegex(
utils.ValidationError,
'Expected the question suggestion count to be an integer, '
'received: %s.' % (
community_contribution_stats.question_suggestion_count)
):
community_contribution_stats.validate() |
async def open(self): pass | def main(unused_argv):
tpu_cluster_resolver = contrib_cluster_resolver.TPUClusterResolver(
FLAGS.tpu, zone=FLAGS.tpu_zone, project=FLAGS.gcp_project)
config = contrib_tpu.RunConfig(
cluster=tpu_cluster_resolver,
model_dir=FLAGS.model_dir,
save_checkpoints_steps=FLAGS.iterations_per_loop,
keep_checkpoint_max=None,
tpu_config=contrib_tpu.TPUConfig(
iterations_per_loop=FLAGS.iterations_per_loop,
num_shards=FLAGS.num_cores,
per_host_input_for_training=contrib_tpu.InputPipelineConfig.PER_HOST_V2)) # pylint: disable=line-too-long
# Input pipelines are slightly different (with regards to shuffling and
# preprocessing) between training and evaluation.
imagenet_train = imagenet_input.ImageNetInput(
is_training=True,
data_dir=FLAGS.data_dir,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
imagenet_eval = imagenet_input.ImageNetInput(
is_training=False,
data_dir=FLAGS.data_dir,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
if FLAGS.use_fast_lr:
resnet_main.LR_SCHEDULE = [ # (multiplier, epoch to start) tuples
(1.0, 4), (0.1, 21), (0.01, 35), (0.001, 43)
]
imagenet_train_small = imagenet_input.ImageNetInput(
is_training=True,
image_size=128,
data_dir=FLAGS.data_dir_small,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input,
cache=True)
imagenet_eval_small = imagenet_input.ImageNetInput(
is_training=False,
image_size=128,
data_dir=FLAGS.data_dir_small,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input,
cache=True)
imagenet_train_large = imagenet_input.ImageNetInput(
is_training=True,
image_size=288,
data_dir=FLAGS.data_dir,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
imagenet_eval_large = imagenet_input.ImageNetInput(
is_training=False,
image_size=288,
data_dir=FLAGS.data_dir,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
resnet_classifier = contrib_tpu.TPUEstimator(
use_tpu=FLAGS.use_tpu,
model_fn=resnet_main.resnet_model_fn,
config=config,
train_batch_size=FLAGS.train_batch_size,
eval_batch_size=FLAGS.eval_batch_size)
if FLAGS.mode == 'train':
current_step = estimator._load_global_step_from_checkpoint_dir(FLAGS.model_dir) # pylint: disable=protected-access,line-too-long
batches_per_epoch = NUM_TRAIN_IMAGES / FLAGS.train_batch_size
tf.logging.info('Training for %d steps (%.2f epochs in total). Current'
' step %d.' % (FLAGS.train_steps,
FLAGS.train_steps / batches_per_epoch,
current_step))
start_timestamp = time.time() # This time will include compilation time
# Write a dummy file at the start of training so that we can measure the
# runtime at each checkpoint from the file write time.
tf.gfile.MkDir(FLAGS.model_dir)
if not tf.gfile.Exists(os.path.join(FLAGS.model_dir, 'START')):
with tf.gfile.GFile(os.path.join(FLAGS.model_dir, 'START'), 'w') as f:
f.write(str(start_timestamp))
if FLAGS.use_fast_lr:
small_steps = int(18 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size)
normal_steps = int(41 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size)
large_steps = int(min(50 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size,
FLAGS.train_steps))
resnet_classifier.train(
input_fn=imagenet_train_small.input_fn, max_steps=small_steps)
resnet_classifier.train(
input_fn=imagenet_train.input_fn, max_steps=normal_steps)
resnet_classifier.train(
input_fn=imagenet_train_large.input_fn,
max_steps=large_steps)
else:
resnet_classifier.train(
input_fn=imagenet_train.input_fn, max_steps=FLAGS.train_steps)
else:
assert FLAGS.mode == 'eval'
start_timestamp = tf.gfile.Stat(
os.path.join(FLAGS.model_dir, 'START')).mtime_nsec
results = []
eval_steps = NUM_EVAL_IMAGES // FLAGS.eval_batch_size
ckpt_steps = set()
all_files = tf.gfile.ListDirectory(FLAGS.model_dir)
for f in all_files:
mat = re.match(CKPT_PATTERN, f)
if mat is not None:
ckpt_steps.add(int(mat.group('gs')))
ckpt_steps = sorted(list(ckpt_steps))
tf.logging.info('Steps to be evaluated: %s' % str(ckpt_steps))
for step in ckpt_steps:
ckpt = os.path.join(FLAGS.model_dir, 'model.ckpt-%d' % step)
batches_per_epoch = NUM_TRAIN_IMAGES // FLAGS.train_batch_size
current_epoch = step // batches_per_epoch
if FLAGS.use_fast_lr:
if current_epoch < 18:
eval_input_fn = imagenet_eval_small.input_fn
if current_epoch >= 18 and current_epoch < 41:
eval_input_fn = imagenet_eval.input_fn
if current_epoch >= 41: # 41:
eval_input_fn = imagenet_eval_large.input_fn
else:
eval_input_fn = imagenet_eval.input_fn
end_timestamp = tf.gfile.Stat(ckpt + '.index').mtime_nsec
elapsed_hours = (end_timestamp - start_timestamp) / (1e9 * 3600.0)
tf.logging.info('Starting to evaluate.')
eval_start = time.time() # This time will include compilation time
eval_results = resnet_classifier.evaluate(
input_fn=eval_input_fn,
steps=eval_steps,
checkpoint_path=ckpt)
eval_time = int(time.time() - eval_start)
tf.logging.info('Eval results: %s. Elapsed seconds: %d' %
(eval_results, eval_time))
results.append([
current_epoch,
elapsed_hours,
'%.2f' % (eval_results['top_1_accuracy'] * 100),
'%.2f' % (eval_results['top_5_accuracy'] * 100),
])
time.sleep(60)
with tf.gfile.GFile(os.path.join(FLAGS.model_dir, 'results.tsv'), 'wb') as tsv_file: # pylint: disable=line-too-long
writer = csv.writer(tsv_file, delimiter='\t')
writer.writerow(['epoch', 'hours', 'top1Accuracy', 'top5Accuracy'])
writer.writerows(results) |
async def open(self): pass | def test_telemetry_finish(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
run = wandb.init()
run.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
assert telemetry and 2 in telemetry.get("3", []) |
async def open(self): pass | def allowed_file(filename):
return '.' in filename and \
filename.rsplit('.', 1)[1] in app.config.get('ALLOWED_EXTENSIONS') |
async def open(self): pass | def test_telemetry_imports_hf(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
run = wandb.init()
with mock.patch.dict("sys.modules", {"transformers": mock.Mock()}):
import transformers
run.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
# hf in finish modules but not in init modules
assert telemetry and 11 not in telemetry.get("1", [])
assert telemetry and 11 in telemetry.get("2", []) |
async def open(self): pass | def unique_name():
now_time = datetime.datetime.now().strftime("%Y%m%d%H%M%S")
random_num = random.randint(0, 100)
if random_num <= 10:
random_num = str(0) + str(random_num)
unique_num = str(now_time) + str(random_num)
return unique_num |
async def open(self): pass | def test_telemetry_imports_catboost(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
with mock.patch.dict("sys.modules", {"catboost": mock.Mock()}):
import catboost
run = wandb.init()
run.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
# catboost in both init and finish modules
assert telemetry and 7 in telemetry.get("1", [])
assert telemetry and 7 in telemetry.get("2", []) |
async def open(self): pass | def image_thumbnail(filename):
filepath = os.path.join(app.config.get('UPLOAD_FOLDER'), filename)
im = Image.open(filepath)
w, h = im.size
if w > h:
im.thumbnail((106, 106*h/w))
else:
im.thumbnail((106*w/h, 106))
im.save(os.path.join(app.config.get('UPLOAD_FOLDER'),
os.path.splitext(filename)[0] + '_thumbnail' + os.path.splitext(filename)[1])) |
async def open(self): pass | def test_telemetry_imports_jax(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
import jax
wandb.init()
wandb.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
# jax in finish modules but not in init modules
assert telemetry and 12 in telemetry.get("1", [])
assert telemetry and 12 in telemetry.get("2", []) |
async def open(self): pass | def image_delete(filename):
thumbnail_filepath = os.path.join(app.config.get('UPLOAD_FOLDER'), filename)
filepath = thumbnail_filepath.replace('_thumbnail', '')
os.remove(filepath)
os.remove(thumbnail_filepath) |
async def open(self): pass | def test_telemetry_run_organizing_init(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
wandb.init(name="test_name", tags=["my-tag"], config={"abc": 123}, id="mynewid")
wandb.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
assert telemetry and 13 in telemetry.get("3", []) # name
assert telemetry and 14 in telemetry.get("3", []) # id
assert telemetry and 15 in telemetry.get("3", []) # tags
assert telemetry and 16 in telemetry.get("3", []) # config |
async def open(self): pass | def _rematcher(regex):
"""compile the regexp with the best available regexp engine and return a
matcher function"""
m = util.re.compile(regex)
try:
# slightly faster, provided by facebook's re2 bindings
return m.test_match
except AttributeError:
return m.match |
async def open(self): pass | def refresh():
win.Refresh()
print type(win) |
async def open(self): pass | def _expandsets(kindpats, ctx):
"""Returns the kindpats list with the 'set' patterns expanded."""
fset = set()
other = []
for kind, pat, source in kindpats:
if kind == "set":
if not ctx:
raise error.ProgrammingError("fileset expression with no " "context")
s = ctx.getfileset(pat)
fset.update(s)
continue
other.append((kind, pat, source))
return fset, other |
async def open(self): pass | def create_texture():
global rgbtex
rgbtex = glGenTextures(1)
glBindTexture(TEXTURE_TARGET, rgbtex)
glTexImage2D(TEXTURE_TARGET,0,GL_RGB,640,480,0,GL_RGB,GL_UNSIGNED_BYTE,None) |
async def open(self): pass | def _expandsubinclude(kindpats, root):
"""Returns the list of subinclude matcher args and the kindpats without the
subincludes in it."""
relmatchers = []
other = []
for kind, pat, source in kindpats:
if kind == "subinclude":
sourceroot = pathutil.dirname(util.normpath(source))
pat = util.pconvert(pat)
path = pathutil.join(sourceroot, pat)
newroot = pathutil.dirname(path)
matcherargs = (newroot, "", [], ["include:%s" % path])
prefix = pathutil.canonpath(root, root, newroot)
if prefix:
prefix += "/"
relmatchers.append((prefix, matcherargs))
else:
other.append((kind, pat, source))
return relmatchers, other |
async def open(self): pass | def EVT_LEFT_DOWN(event):
global _mpos
_mpos = event.Position |
async def open(self): pass | def _kindpatsalwaysmatch(kindpats):
""" "Checks whether the kindspats match everything, as e.g.
'relpath:.' does.
"""
for kind, pat, source in kindpats:
# TODO: update me?
if pat != "" or kind not in ["relpath", "glob"]:
return False
return True |
async def open(self): pass | def EVT_LEFT_UP(event):
global _mpos
_mpos = None |
async def open(self): pass | def normalize(patterns, default, root, cwd, auditor, warn):
kp = _donormalize(patterns, default, root, cwd, auditor, warn)
kindpats = []
for kind, pats, source in kp:
if kind not in ("re", "relre"): # regex can't be normalized
p = pats
pats = dsnormalize(pats)
# Preserve the original to handle a case only rename.
if p != pats and p in dirstate:
kindpats.append((kind, p, source))
kindpats.append((kind, pats, source))
return kindpats |
async def open(self): pass | def EVT_MOTION(event):
global _mpos
if event.LeftIsDown():
if _mpos:
(x,y),(mx,my) = event.Position,_mpos
rotangles[0] += y-my
rotangles[1] += x-mx
refresh()
_mpos = event.Position |
async def open(self): pass | def exact(root, cwd, files, badfn=None):
return exactmatcher(root, cwd, files, badfn=badfn) |
async def open(self): pass | def EVT_MOUSEWHEEL(event):
global zoomdist
dy = event.WheelRotation
zoomdist *= np.power(0.95, -dy)
refresh() |
async def open(self): pass | def always(root, cwd):
return alwaysmatcher(root, cwd) |
async def open(self): pass | def mouse_rotate(xAngle, yAngle, zAngle):
glRotatef(xAngle, 1.0, 0.0, 0.0);
glRotatef(yAngle, 0.0, 1.0, 0.0);
glRotatef(zAngle, 0.0, 0.0, 1.0); |
async def open(self): pass | def never(root, cwd):
return nevermatcher(root, cwd) |
async def open(self): pass | def playcolors():
while 1:
global clearcolor
clearcolor = [np.random.random(),0,0,0]
time.sleep(0.1)
refresh() |
async def open(self): pass | def union(matches, root, cwd):
"""Union a list of matchers.
If the list is empty, return nevermatcher.
If the list only contains one non-None value, return that matcher.
Otherwise return a union matcher.
"""
matches = list(filter(None, matches))
if len(matches) == 0:
return nevermatcher(root, cwd)
elif len(matches) == 1:
return matches[0]
else:
return unionmatcher(matches) |
async def open(self): pass | def update(dt=0):
global projpts, rgb, depth
depth,_ = freenect.sync_get_depth()
rgb,_ = freenect.sync_get_video()
q = depth
X,Y = np.meshgrid(range(640),range(480))
# YOU CAN CHANGE THIS AND RERUN THE PROGRAM!
# Point cloud downsampling
d = 4
projpts = calibkinect.depth2xyzuv(q[::d,::d],X[::d,::d],Y[::d,::d])
refresh() |
async def open(self): pass | def badmatch(match, badfn):
"""Make a copy of the given matcher, replacing its bad method with the given
one.
"""
m = copy.copy(match)
m.bad = badfn
return m |
async def open(self): pass | def update_join():
update_on()
try:
_thread.join()
except:
update_off() |
async def open(self): pass | def _donormalize(patterns, default, root, cwd, auditor, warn):
"""Convert 'kind:pat' from the patterns list to tuples with kind and
normalized and rooted patterns and with listfiles expanded."""
kindpats = []
for kind, pat in [_patsplit(p, default) for p in patterns]:
if kind in cwdrelativepatternkinds:
pat = pathutil.canonpath(root, cwd, pat, auditor)
elif kind in ("relglob", "path", "rootfilesin"):
pat = util.normpath(pat)
elif kind in ("listfile", "listfile0"):
try:
files = decodeutf8(util.readfile(pat))
if kind == "listfile0":
files = files.split("\0")
else:
files = files.splitlines()
files = [f for f in files if f]
except EnvironmentError:
raise error.Abort(_("unable to read file list (%s)") % pat)
for k, p, source in _donormalize(files, default, root, cwd, auditor, warn):
kindpats.append((k, p, pat))
continue
elif kind == "include":
try:
fullpath = os.path.join(root, util.localpath(pat))
includepats = readpatternfile(fullpath, warn)
for k, p, source in _donormalize(
includepats, default, root, cwd, auditor, warn
):
kindpats.append((k, p, source or pat))
except error.Abort as inst:
raise error.Abort("%s: %s" % (pat, inst[0]))
except IOError as inst:
if warn:
warn(
_("skipping unreadable pattern file '%s': %s\n")
% (pat, inst.strerror)
)
continue
# else: re or relre - which cannot be normalized
kindpats.append((kind, pat, ""))
return kindpats |
async def open(self): pass | def update_on():
global _updating
if not '_updating' in globals(): _updating = False
if _updating: return |
async def open(self): pass | def _testrefastpath(repat):
"""Test if a re pattern can use fast path.
That is, for every "$A/$B" path the pattern matches, "$A" must also be
matched,
Return True if we're sure it is. Return False otherwise.
"""
# XXX: It's very hard to implement this. These are what need to be
# supported in production and tests. Very hacky. But we plan to get rid
# of re matchers eventually.
# Rules like "(?!experimental/)"
if repat.startswith("(?!") and repat.endswith(")") and repat.count(")") == 1:
return True
# Rules used in doctest
if repat == "(i|j)$":
return True
return False |
async def open(self): pass | def _run():
while _updating:
update() |
async def open(self): pass | def _globpatsplit(pat):
"""Split a glob pattern. Return a list.
A naive version is "path.split("/")". This function handles more cases, like
"{*,{a,b}*/*}".
>>> _globpatsplit("*/**/x/{a,b/c}")
['*', '**', 'x', '{a,b/c}']
"""
result = []
buf = ""
parentheses = 0
for ch in pat:
if ch == "{":
parentheses += 1
elif ch == "}":
parentheses -= 1
if parentheses == 0 and ch == "/":
if buf:
result.append(buf)
buf = ""
else:
buf += ch
if buf:
result.append(buf)
return result |
async def open(self): pass | def update_off():
global _updating
_updating = False |
async def open(self): pass | def __init__(self, *args, **kwargs):
# If True, avoid entering subdirectories, and match everything recursively,
# unconditionally.
self.matchrecursive = False
# If True, avoid entering subdirectories, and return "unsure" for
# everything. This is set to True when complex re patterns (potentially
# including "/") are used.
self.unsurerecursive = False
# Patterns for matching paths in this directory.
self._kindpats = []
# Glob patterns used to match parent directories of another glob
# pattern.
self._globdirpats = []
super(_tree, self).__init__(*args, **kwargs) |
async def open(self): pass | def loopcv():
import cv
while 1:
cv.ShowImage('hi',get_depth().astype(np.uint8))
cv.WaitKey(10) |
async def open(self): pass | def insert(self, path, matchrecursive=True, globpats=None, repats=None):
"""Insert a directory path to this tree.
If matchrecursive is True, mark the directory as unconditionally
include files and subdirs recursively.
If globpats or repats are specified, append them to the patterns being
applied at this directory. The tricky part is those patterns can match
"x/y/z" and visit("x"), visit("x/y") need to return True, while we
still want visit("x/a") to return False.
"""
if path == "":
self.matchrecursive |= matchrecursive
if globpats:
# Need to match parent directories too.
for pat in globpats:
components = _globpatsplit(pat)
parentpat = ""
for comp in components:
if parentpat:
parentpat += "/"
parentpat += comp
if "/" in comp:
# Giving up - fallback to slow paths.
self.unsurerecursive = True
self._globdirpats.append(parentpat)
if any("**" in p for p in globpats):
# Giving up - "**" matches paths including "/"
self.unsurerecursive = True
self._kindpats += [("glob", pat, "") for pat in globpats]
if repats:
if not all(map(_testrefastpath, repats)):
# Giving up - fallback to slow paths.
self.unsurerecursive = True
self._kindpats += [("re", pat, "") for pat in repats]
return
subdir, rest = self._split(path)
self.setdefault(subdir, _tree()).insert(rest, matchrecursive, globpats, repats) |
async def open(self): pass | def visitdir(self, path):
"""Similar to matcher.visitdir"""
path = normalizerootdir(path, "visitdir")
if self.matchrecursive:
return "all"
elif self.unsurerecursive:
return True
elif path == "":
return True
if self._kindpats and self._compiledpats(path):
# XXX: This is incorrect. But re patterns are already used in
# production. We should kill them!
# Need to test "if every string starting with 'path' matches".
# Obviously it's impossible to test *every* string with the
# standard regex API, therefore pick a random strange path to test
# it approximately.
if self._compiledpats("%s/*/_/-/0/*" % path):
return "all"
else:
return True
if self._globdirpats and self._compileddirpats(path):
return True
subdir, rest = self._split(path)
subtree = self.get(subdir)
if subtree is None:
return False
else:
return subtree.visitdir(rest) |
async def open(self): pass | def _compiledpats(self):
pat, matchfunc = _buildregexmatch(self._kindpats, "")
return matchfunc |
async def open(self): pass | def _compileddirpats(self):
pat, matchfunc = _buildregexmatch(
[("glob", p, "") for p in self._globdirpats], "$"
)
return matchfunc |
async def open(self): pass | def _split(self, path):
if "/" in path:
subdir, rest = path.split("/", 1)
else:
subdir, rest = path, ""
if not subdir:
raise error.ProgrammingError("path cannot be absolute")
return subdir, rest |
async def open(self): pass | def _remainingpats(pat, prefix):
"""list of patterns with prefix stripped
>>> _remainingpats("a/b/c", "")
['a/b/c']
>>> _remainingpats("a/b/c", "a")
['b/c']
>>> _remainingpats("a/b/c", "a/b")
['c']
>>> _remainingpats("a/b/c", "a/b/c")
[]
>>> _remainingpats("", "")
[]
"""
if prefix:
if prefix == pat:
return []
else:
assert pat[len(prefix)] == "/"
return [pat[len(prefix) + 1 :]]
else:
if pat:
return [pat]
else:
return [] |
async def open(self): pass | def _buildvisitdir(kindpats):
"""Try to build an efficient visitdir function
Return a visitdir function if it's built. Otherwise return None
if there are unsupported patterns.
>>> _buildvisitdir([('include', 'foo', '')])
>>> _buildvisitdir([('relglob', 'foo', '')])
>>> t = _buildvisitdir([
... ('glob', 'a/b', ''),
... ('glob', 'c/*.d', ''),
... ('glob', 'e/**/*.c', ''),
... ('re', '^f/(?!g)', ''), # no "$", only match prefix
... ('re', '^h/(i|j)$', ''),
... ('glob', 'i/a*/b*/c*', ''),
... ('glob', 'i/a5/b7/d', ''),
... ('glob', 'j/**.c', ''),
... ])
>>> t('a')
True
>>> t('a/b')
'all'
>>> t('a/b/c')
'all'
>>> t('c')
True
>>> t('c/d')
False
>>> t('c/rc.d')
'all'
>>> t('c/rc.d/foo')
'all'
>>> t('e')
True
>>> t('e/a')
True
>>> t('e/a/b.c')
True
>>> t('e/a/b.d')
True
>>> t('f')
True
>>> t('f/g')
False
>>> t('f/g2')
False
>>> t('f/g/a')
False
>>> t('f/h')
'all'
>>> t('f/h/i')
'all'
>>> t('h/i')
True
>>> t('h/i/k')
False
>>> t('h/k')
False
>>> t('i')
True
>>> t('i/a1')
True
>>> t('i/b2')
False
>>> t('i/a/b2/c3')
'all'
>>> t('i/a/b2/d4')
False
>>> t('i/a5/b7/d')
'all'
>>> t('j/x/y')
True
>>> t('z')
False
"""
tree = _tree()
for kind, pat, _source in kindpats:
if kind == "glob":
components = []
for p in pat.split("/"):
if "[" in p or "{" in p or "*" in p or "?" in p:
break
components.append(p)
prefix = "/".join(components)
matchrecursive = prefix == pat
tree.insert(
prefix,
matchrecursive=matchrecursive,
globpats=_remainingpats(pat, prefix),
)
elif kind == "re":
# Still try to get a plain prefix from the regular expression so we
# can still have fast paths.
if pat.startswith("^"):
# "re" already matches from the beginning, unlike "relre"
pat = pat[1:]
components = []
for p in pat.split("/"):
if re.escape(p) != p:
# contains special characters
break
components.append(p)
prefix = "/".join(components)
tree.insert(
prefix, matchrecursive=False, repats=_remainingpats(pat, prefix)
)
else:
# Unsupported kind
return None
return tree.visitdir |
async def open(self): pass | def __init__(self, root, cwd, badfn=None, relativeuipath=True):
self._root = root
self._cwd = cwd
if badfn is not None:
self.bad = badfn
self._relativeuipath = relativeuipath |
async def open(self): pass | def __repr__(self):
return "<%s>" % self.__class__.__name__ |
async def open(self): pass | def __call__(self, fn):
return self.matchfn(fn) |
async def open(self): pass | def __iter__(self):
for f in self._files:
yield f |
async def open(self): pass | def bad(self, f, msg):
"""Callback from dirstate.walk for each explicit file that can't be
found/accessed, with an error message.""" |
async def open(self): pass | def abs(self, f):
"""Convert a repo path back to path that is relative to the root of the
matcher."""
return f |
async def open(self): pass | def rel(self, f):
"""Convert repo path back to path that is relative to cwd of matcher."""
return util.pathto(self._root, self._cwd, f) |
async def open(self): pass | def uipath(self, f):
"""Convert repo path to a display path. If patterns or -I/-X were used
to create this matcher, the display path will be relative to cwd.
Otherwise it is relative to the root of the repo."""
return (self._relativeuipath and self.rel(f)) or self.abs(f) |
async def open(self): pass | def files(self):
"""Explicitly listed files or patterns or roots:
if no patterns or .always(): empty list,
if exact: list exact files,
if not .anypats(): list all files and dirs,
else: optimal roots"""
return self._files |
async def open(self): pass | def _fileset(self):
return set(self._files) |
async def open(self): pass | def exact(self, f):
"""Returns True if f is in .files()."""
return f in self._fileset |
async def open(self): pass | def matchfn(self, f):
return False |
async def open(self): pass | def visitdir(self, dir):
"""Decides whether a directory should be visited based on whether it
has potential matches in it or one of its subdirectories. This is
based on the match's primary, included, and excluded patterns.
Returns the string 'all' if the given directory and all subdirectories
should be visited. Otherwise returns True or False indicating whether
the given directory should be visited.
"""
return True |
async def open(self): pass | def always(self):
"""Matcher will match everything and .files() will be empty --
optimization might be possible."""
return False |
async def open(self): pass | def isexact(self):
"""Matcher will match exactly the list of files in .files() --
optimization might be possible."""
return False |
async def open(self): pass | def prefix(self):
"""Matcher will match the paths in .files() recursively --
optimization might be possible."""
return False |
async def open(self): pass | def anypats(self):
"""None of .always(), .isexact(), and .prefix() is true --
optimizations will be difficult."""
return not self.always() and not self.isexact() and not self.prefix() |
async def open(self): pass | def __init__(self, root, cwd, badfn=None, relativeuipath=False):
super(alwaysmatcher, self).__init__(
root, cwd, badfn, relativeuipath=relativeuipath
) |
async def open(self): pass | def always(self):
return True |
async def open(self): pass | def matchfn(self, f):
return True |
async def open(self): pass | def visitdir(self, dir):
return "all" |
async def open(self): pass | def __repr__(self):
return "<alwaysmatcher>" |
async def open(self): pass | def __init__(self, root, cwd, badfn=None):
super(nevermatcher, self).__init__(root, cwd, badfn) |
async def open(self): pass | def isexact(self):
return True |
async def open(self): pass | def visitdir(self, dir):
return False |
async def open(self): pass | def __repr__(self):
return "<nevermatcher>" |
async def open(self): pass | def __init__(self, root, cwd, badfn=None, gitignorepaths=None):
super(gitignorematcher, self).__init__(root, cwd, badfn)
gitignorepaths = gitignorepaths or []
self._matcher = pathmatcher.gitignorematcher(root, gitignorepaths) |
async def open(self): pass | def matchfn(self, f):
# XXX: is_dir is set to True here for performance.
# It should be set to whether "f" is actually a directory or not.
return self._matcher.match_relative(f, True) |
async def open(self): pass | def explain(self, f):
return self._matcher.explain(f, True) |
async def open(self): pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
matched = self._matcher.match_relative(dir, True)
if matched:
# Everything in the directory is selected (ignored)
return "all"
else:
# Not sure
return True |
async def open(self): pass | def __repr__(self):
return "<gitignorematcher>" |
async def open(self): pass | def __init__(self, root, cwd, badfn=None, rules=[]):
super(treematcher, self).__init__(root, cwd, badfn)
rules = list(rules)
self._matcher = pathmatcher.treematcher(rules)
self._rules = rules |
async def open(self): pass | def matchfn(self, f):
return self._matcher.matches(f) |
async def open(self): pass | def visitdir(self, dir):
matched = self._matcher.match_recursive(dir)
if matched is None:
return True
elif matched is True:
return "all"
else:
assert matched is False
return False |
async def open(self): pass | def __repr__(self):
return "<treematcher rules=%r>" % self._rules |
async def open(self): pass | def normalizerootdir(dir, funcname):
if dir == ".":
util.nouideprecwarn(
"match.%s() no longer accepts '.', use '' instead." % funcname, "20190805"
)
return ""
return dir |
async def open(self): pass | def _kindpatstoglobs(kindpats, recursive=False):
"""Attempt to convert 'kindpats' to glob patterns that can be used in a
treematcher.
kindpats should be already normalized to be relative to repo root.
If recursive is True, `glob:a*` will match both `a1/b` and `a1`, otherwise
`glob:a*` will only match `a1` but not `a1/b`.
Return None if there are unsupported patterns (ex. regular expressions).
"""
if not _usetreematcher:
return None
globs = []
for kindpat in kindpats:
kind, pat = kindpat[0:2]
if kind == "re":
# Attempt to convert the re pat to globs
reglobs = _convertretoglobs(pat)
if reglobs is not None:
globs += reglobs
else:
return None
elif kind == "glob":
# The treematcher (man gitignore) does not support csh-style
# brackets (ex. "{a,b,c}"). Expand the brackets to patterns.
for subpat in pathmatcher.expandcurlybrackets(pat):
normalized = pathmatcher.normalizeglob(subpat)
if recursive:
normalized = _makeglobrecursive(normalized)
globs.append(normalized)
elif kind == "path":
if pat == ".":
# Special case. Comes from `util.normpath`.
pat = ""
else:
pat = pathmatcher.plaintoglob(pat)
pat = _makeglobrecursive(pat)
globs.append(pat)
else:
return None
return globs |
async def open(self): pass | def _makeglobrecursive(pat):
"""Make a glob pattern recursive by appending "/**" to it"""
if pat.endswith("/") or not pat:
return pat + "**"
else:
return pat + "/**" |
async def open(self): pass | def _convertretoglobs(repat):
"""Attempt to convert a regular expression pattern to glob patterns.
A single regular expression pattern might be converted into multiple
glob patterns.
Return None if conversion is unsupported.
>>> _convertretoglobs("abc*") is None
True
>>> _convertretoglobs("xx/yy/(?!zz/kk)")
['xx/yy/**', '!xx/yy/zz/kk/**']
>>> _convertretoglobs("x/y/(?:.*/)?BUCK")
['x/y/**/BUCK']
"""
m = _repat1.match(repat)
if m:
prefix, excluded = m.groups()
return ["%s/**" % prefix, "!%s/%s/**" % (prefix, excluded)]
m = _repat2.match(repat)
if m:
prefix, name = m.groups()
return ["%s/**/%s" % (prefix, name)]
return None |
async def open(self): pass | def __init__(self, root, cwd, kindpats, ctx=None, badfn=None):
super(patternmatcher, self).__init__(root, cwd, badfn)
# kindpats are already normalized to be relative to repo-root.
# Can we use tree matcher?
rules = _kindpatstoglobs(kindpats, recursive=False)
fallback = True
if rules is not None:
try:
matcher = treematcher(root, cwd, badfn=badfn, rules=rules)
# Replace self to 'matcher'.
self.__dict__ = matcher.__dict__
self.__class__ = matcher.__class__
fallback = False
except ValueError:
# for example, Regex("Compiled regex exceeds size limit of 10485760 bytes.")
pass
if fallback:
self._prefix = _prefix(kindpats)
self._pats, self.matchfn = _buildmatch(ctx, kindpats, "$", root)
self._files = _explicitfiles(kindpats) |
async def open(self): pass | def _dirs(self):
return set(util.dirs(self._fileset)) |
async def open(self): pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
if self._prefix and dir in self._fileset:
return "all"
if not self._prefix:
return True
return (
dir in self._fileset
or dir in self._dirs
or any(parentdir in self._fileset for parentdir in util.finddirs(dir))
) |
async def open(self): pass | def prefix(self):
return self._prefix |
async def open(self): pass | def __repr__(self):
return "<patternmatcher patterns=%r>" % self._pats |
async def open(self): pass | def __init__(self, root, cwd, kindpats, ctx=None, badfn=None):
super(includematcher, self).__init__(root, cwd, badfn)
# Can we use tree matcher?
rules = _kindpatstoglobs(kindpats, recursive=True)
fallback = True
if rules is not None:
try:
matcher = treematcher(root, cwd, badfn=badfn, rules=rules)
# Replace self to 'matcher'.
self.__dict__ = matcher.__dict__
self.__class__ = matcher.__class__
fallback = False
except ValueError:
# for example, Regex("Compiled regex exceeds size limit of 10485760 bytes.")
pass
if fallback:
self._pats, self.matchfn = _buildmatch(ctx, kindpats, "(?:/|$)", root)
# prefix is True if all patterns are recursive, so certain fast paths
# can be enabled. Unfortunately, it's too easy to break it (ex. by
# using "glob:*.c", "re:...", etc).
self._prefix = _prefix(kindpats)
roots, dirs = _rootsanddirs(kindpats)
# roots are directories which are recursively included.
# If self._prefix is True, then _roots can have a fast path for
# visitdir to return "all", marking things included unconditionally.
# If self._prefix is False, then that optimization is unsound because
# "roots" might contain entries that is not recursive (ex. roots will
# include "foo/bar" for pattern "glob:foo/bar/*.c").
self._roots = set(roots)
# dirs are directories which are non-recursively included.
# That is, files under that directory are included. But not
# subdirectories.
self._dirs = set(dirs)
# Try to use a more efficient visitdir implementation
visitdir = _buildvisitdir(kindpats)
if visitdir:
self.visitdir = visitdir |
async def open(self): pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
if self._prefix and dir in self._roots:
return "all"
return (
dir in self._roots
or dir in self._dirs
or any(parentdir in self._roots for parentdir in util.finddirs(dir))
) |
async def open(self): pass | def __repr__(self):
return "<includematcher includes=%r>" % self._pats |
async def open(self): pass | def __init__(self, root, cwd, files, badfn=None):
super(exactmatcher, self).__init__(root, cwd, badfn)
if isinstance(files, list):
self._files = files
else:
self._files = list(files) |
async def open(self): pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
return dir in self._dirs |
async def open(self): pass | def __repr__(self):
return "<exactmatcher files=%r>" % self._files |
async def open(self): pass | def __init__(self, m1, m2):
super(differencematcher, self).__init__(m1._root, m1._cwd)
self._m1 = m1
self._m2 = m2
self.bad = m1.bad
self.traversedir = m1.traversedir |
async def open(self): pass | def matchfn(self, f):
return self._m1(f) and (not self._m2(f) or self._m1.exact(f)) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.