input stringlengths 11 7.65k | target stringlengths 22 8.26k |
|---|---|
def _save_obj(self, *args, **kwargs):
pass | def test_validate_question_suggestion_count_fails_for_non_integer_count(
self):
community_contribution_stats = (
suggestion_services.get_community_contribution_stats()
)
community_contribution_stats.question_suggestion_count = (
self.non_integer_count
)
with self.assertRaisesRegex(
utils.ValidationError,
'Expected the question suggestion count to be an integer, '
'received: %s.' % (
community_contribution_stats.question_suggestion_count)
):
community_contribution_stats.validate() |
def _save_obj(self, *args, **kwargs):
pass | def main(unused_argv):
tpu_cluster_resolver = contrib_cluster_resolver.TPUClusterResolver(
FLAGS.tpu, zone=FLAGS.tpu_zone, project=FLAGS.gcp_project)
config = contrib_tpu.RunConfig(
cluster=tpu_cluster_resolver,
model_dir=FLAGS.model_dir,
save_checkpoints_steps=FLAGS.iterations_per_loop,
keep_checkpoint_max=None,
tpu_config=contrib_tpu.TPUConfig(
iterations_per_loop=FLAGS.iterations_per_loop,
num_shards=FLAGS.num_cores,
per_host_input_for_training=contrib_tpu.InputPipelineConfig.PER_HOST_V2)) # pylint: disable=line-too-long
# Input pipelines are slightly different (with regards to shuffling and
# preprocessing) between training and evaluation.
imagenet_train = imagenet_input.ImageNetInput(
is_training=True,
data_dir=FLAGS.data_dir,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
imagenet_eval = imagenet_input.ImageNetInput(
is_training=False,
data_dir=FLAGS.data_dir,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
if FLAGS.use_fast_lr:
resnet_main.LR_SCHEDULE = [ # (multiplier, epoch to start) tuples
(1.0, 4), (0.1, 21), (0.01, 35), (0.001, 43)
]
imagenet_train_small = imagenet_input.ImageNetInput(
is_training=True,
image_size=128,
data_dir=FLAGS.data_dir_small,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input,
cache=True)
imagenet_eval_small = imagenet_input.ImageNetInput(
is_training=False,
image_size=128,
data_dir=FLAGS.data_dir_small,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input,
cache=True)
imagenet_train_large = imagenet_input.ImageNetInput(
is_training=True,
image_size=288,
data_dir=FLAGS.data_dir,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
imagenet_eval_large = imagenet_input.ImageNetInput(
is_training=False,
image_size=288,
data_dir=FLAGS.data_dir,
num_parallel_calls=FLAGS.num_parallel_calls,
use_bfloat16=True,
transpose_input=FLAGS.transpose_input)
resnet_classifier = contrib_tpu.TPUEstimator(
use_tpu=FLAGS.use_tpu,
model_fn=resnet_main.resnet_model_fn,
config=config,
train_batch_size=FLAGS.train_batch_size,
eval_batch_size=FLAGS.eval_batch_size)
if FLAGS.mode == 'train':
current_step = estimator._load_global_step_from_checkpoint_dir(FLAGS.model_dir) # pylint: disable=protected-access,line-too-long
batches_per_epoch = NUM_TRAIN_IMAGES / FLAGS.train_batch_size
tf.logging.info('Training for %d steps (%.2f epochs in total). Current'
' step %d.' % (FLAGS.train_steps,
FLAGS.train_steps / batches_per_epoch,
current_step))
start_timestamp = time.time() # This time will include compilation time
# Write a dummy file at the start of training so that we can measure the
# runtime at each checkpoint from the file write time.
tf.gfile.MkDir(FLAGS.model_dir)
if not tf.gfile.Exists(os.path.join(FLAGS.model_dir, 'START')):
with tf.gfile.GFile(os.path.join(FLAGS.model_dir, 'START'), 'w') as f:
f.write(str(start_timestamp))
if FLAGS.use_fast_lr:
small_steps = int(18 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size)
normal_steps = int(41 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size)
large_steps = int(min(50 * NUM_TRAIN_IMAGES / FLAGS.train_batch_size,
FLAGS.train_steps))
resnet_classifier.train(
input_fn=imagenet_train_small.input_fn, max_steps=small_steps)
resnet_classifier.train(
input_fn=imagenet_train.input_fn, max_steps=normal_steps)
resnet_classifier.train(
input_fn=imagenet_train_large.input_fn,
max_steps=large_steps)
else:
resnet_classifier.train(
input_fn=imagenet_train.input_fn, max_steps=FLAGS.train_steps)
else:
assert FLAGS.mode == 'eval'
start_timestamp = tf.gfile.Stat(
os.path.join(FLAGS.model_dir, 'START')).mtime_nsec
results = []
eval_steps = NUM_EVAL_IMAGES // FLAGS.eval_batch_size
ckpt_steps = set()
all_files = tf.gfile.ListDirectory(FLAGS.model_dir)
for f in all_files:
mat = re.match(CKPT_PATTERN, f)
if mat is not None:
ckpt_steps.add(int(mat.group('gs')))
ckpt_steps = sorted(list(ckpt_steps))
tf.logging.info('Steps to be evaluated: %s' % str(ckpt_steps))
for step in ckpt_steps:
ckpt = os.path.join(FLAGS.model_dir, 'model.ckpt-%d' % step)
batches_per_epoch = NUM_TRAIN_IMAGES // FLAGS.train_batch_size
current_epoch = step // batches_per_epoch
if FLAGS.use_fast_lr:
if current_epoch < 18:
eval_input_fn = imagenet_eval_small.input_fn
if current_epoch >= 18 and current_epoch < 41:
eval_input_fn = imagenet_eval.input_fn
if current_epoch >= 41: # 41:
eval_input_fn = imagenet_eval_large.input_fn
else:
eval_input_fn = imagenet_eval.input_fn
end_timestamp = tf.gfile.Stat(ckpt + '.index').mtime_nsec
elapsed_hours = (end_timestamp - start_timestamp) / (1e9 * 3600.0)
tf.logging.info('Starting to evaluate.')
eval_start = time.time() # This time will include compilation time
eval_results = resnet_classifier.evaluate(
input_fn=eval_input_fn,
steps=eval_steps,
checkpoint_path=ckpt)
eval_time = int(time.time() - eval_start)
tf.logging.info('Eval results: %s. Elapsed seconds: %d' %
(eval_results, eval_time))
results.append([
current_epoch,
elapsed_hours,
'%.2f' % (eval_results['top_1_accuracy'] * 100),
'%.2f' % (eval_results['top_5_accuracy'] * 100),
])
time.sleep(60)
with tf.gfile.GFile(os.path.join(FLAGS.model_dir, 'results.tsv'), 'wb') as tsv_file: # pylint: disable=line-too-long
writer = csv.writer(tsv_file, delimiter='\t')
writer.writerow(['epoch', 'hours', 'top1Accuracy', 'top5Accuracy'])
writer.writerows(results) |
def _save_obj(self, *args, **kwargs):
pass | def test_telemetry_finish(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
run = wandb.init()
run.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
assert telemetry and 2 in telemetry.get("3", []) |
def _save_obj(self, *args, **kwargs):
pass | def allowed_file(filename):
return '.' in filename and \
filename.rsplit('.', 1)[1] in app.config.get('ALLOWED_EXTENSIONS') |
def _save_obj(self, *args, **kwargs):
pass | def test_telemetry_imports_hf(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
run = wandb.init()
with mock.patch.dict("sys.modules", {"transformers": mock.Mock()}):
import transformers
run.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
# hf in finish modules but not in init modules
assert telemetry and 11 not in telemetry.get("1", [])
assert telemetry and 11 in telemetry.get("2", []) |
def _save_obj(self, *args, **kwargs):
pass | def unique_name():
now_time = datetime.datetime.now().strftime("%Y%m%d%H%M%S")
random_num = random.randint(0, 100)
if random_num <= 10:
random_num = str(0) + str(random_num)
unique_num = str(now_time) + str(random_num)
return unique_num |
def _save_obj(self, *args, **kwargs):
pass | def test_telemetry_imports_catboost(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
with mock.patch.dict("sys.modules", {"catboost": mock.Mock()}):
import catboost
run = wandb.init()
run.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
# catboost in both init and finish modules
assert telemetry and 7 in telemetry.get("1", [])
assert telemetry and 7 in telemetry.get("2", []) |
def _save_obj(self, *args, **kwargs):
pass | def image_thumbnail(filename):
filepath = os.path.join(app.config.get('UPLOAD_FOLDER'), filename)
im = Image.open(filepath)
w, h = im.size
if w > h:
im.thumbnail((106, 106*h/w))
else:
im.thumbnail((106*w/h, 106))
im.save(os.path.join(app.config.get('UPLOAD_FOLDER'),
os.path.splitext(filename)[0] + '_thumbnail' + os.path.splitext(filename)[1])) |
def _save_obj(self, *args, **kwargs):
pass | def test_telemetry_imports_jax(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
import jax
wandb.init()
wandb.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
# jax in finish modules but not in init modules
assert telemetry and 12 in telemetry.get("1", [])
assert telemetry and 12 in telemetry.get("2", []) |
def _save_obj(self, *args, **kwargs):
pass | def image_delete(filename):
thumbnail_filepath = os.path.join(app.config.get('UPLOAD_FOLDER'), filename)
filepath = thumbnail_filepath.replace('_thumbnail', '')
os.remove(filepath)
os.remove(thumbnail_filepath) |
def _save_obj(self, *args, **kwargs):
pass | def test_telemetry_run_organizing_init(runner, live_mock_server, parse_ctx):
with runner.isolated_filesystem():
wandb.init(name="test_name", tags=["my-tag"], config={"abc": 123}, id="mynewid")
wandb.finish()
ctx_util = parse_ctx(live_mock_server.get_ctx())
telemetry = ctx_util.telemetry
assert telemetry and 13 in telemetry.get("3", []) # name
assert telemetry and 14 in telemetry.get("3", []) # id
assert telemetry and 15 in telemetry.get("3", []) # tags
assert telemetry and 16 in telemetry.get("3", []) # config |
def _save_obj(self, *args, **kwargs):
pass | def _rematcher(regex):
"""compile the regexp with the best available regexp engine and return a
matcher function"""
m = util.re.compile(regex)
try:
# slightly faster, provided by facebook's re2 bindings
return m.test_match
except AttributeError:
return m.match |
def _save_obj(self, *args, **kwargs):
pass | def refresh():
win.Refresh()
print type(win) |
def _save_obj(self, *args, **kwargs):
pass | def _expandsets(kindpats, ctx):
"""Returns the kindpats list with the 'set' patterns expanded."""
fset = set()
other = []
for kind, pat, source in kindpats:
if kind == "set":
if not ctx:
raise error.ProgrammingError("fileset expression with no " "context")
s = ctx.getfileset(pat)
fset.update(s)
continue
other.append((kind, pat, source))
return fset, other |
def _save_obj(self, *args, **kwargs):
pass | def create_texture():
global rgbtex
rgbtex = glGenTextures(1)
glBindTexture(TEXTURE_TARGET, rgbtex)
glTexImage2D(TEXTURE_TARGET,0,GL_RGB,640,480,0,GL_RGB,GL_UNSIGNED_BYTE,None) |
def _save_obj(self, *args, **kwargs):
pass | def _expandsubinclude(kindpats, root):
"""Returns the list of subinclude matcher args and the kindpats without the
subincludes in it."""
relmatchers = []
other = []
for kind, pat, source in kindpats:
if kind == "subinclude":
sourceroot = pathutil.dirname(util.normpath(source))
pat = util.pconvert(pat)
path = pathutil.join(sourceroot, pat)
newroot = pathutil.dirname(path)
matcherargs = (newroot, "", [], ["include:%s" % path])
prefix = pathutil.canonpath(root, root, newroot)
if prefix:
prefix += "/"
relmatchers.append((prefix, matcherargs))
else:
other.append((kind, pat, source))
return relmatchers, other |
def _save_obj(self, *args, **kwargs):
pass | def EVT_LEFT_DOWN(event):
global _mpos
_mpos = event.Position |
def _save_obj(self, *args, **kwargs):
pass | def _kindpatsalwaysmatch(kindpats):
""" "Checks whether the kindspats match everything, as e.g.
'relpath:.' does.
"""
for kind, pat, source in kindpats:
# TODO: update me?
if pat != "" or kind not in ["relpath", "glob"]:
return False
return True |
def _save_obj(self, *args, **kwargs):
pass | def EVT_LEFT_UP(event):
global _mpos
_mpos = None |
def _save_obj(self, *args, **kwargs):
pass | def normalize(patterns, default, root, cwd, auditor, warn):
kp = _donormalize(patterns, default, root, cwd, auditor, warn)
kindpats = []
for kind, pats, source in kp:
if kind not in ("re", "relre"): # regex can't be normalized
p = pats
pats = dsnormalize(pats)
# Preserve the original to handle a case only rename.
if p != pats and p in dirstate:
kindpats.append((kind, p, source))
kindpats.append((kind, pats, source))
return kindpats |
def _save_obj(self, *args, **kwargs):
pass | def EVT_MOTION(event):
global _mpos
if event.LeftIsDown():
if _mpos:
(x,y),(mx,my) = event.Position,_mpos
rotangles[0] += y-my
rotangles[1] += x-mx
refresh()
_mpos = event.Position |
def _save_obj(self, *args, **kwargs):
pass | def exact(root, cwd, files, badfn=None):
return exactmatcher(root, cwd, files, badfn=badfn) |
def _save_obj(self, *args, **kwargs):
pass | def EVT_MOUSEWHEEL(event):
global zoomdist
dy = event.WheelRotation
zoomdist *= np.power(0.95, -dy)
refresh() |
def _save_obj(self, *args, **kwargs):
pass | def always(root, cwd):
return alwaysmatcher(root, cwd) |
def _save_obj(self, *args, **kwargs):
pass | def mouse_rotate(xAngle, yAngle, zAngle):
glRotatef(xAngle, 1.0, 0.0, 0.0);
glRotatef(yAngle, 0.0, 1.0, 0.0);
glRotatef(zAngle, 0.0, 0.0, 1.0); |
def _save_obj(self, *args, **kwargs):
pass | def never(root, cwd):
return nevermatcher(root, cwd) |
def _save_obj(self, *args, **kwargs):
pass | def playcolors():
while 1:
global clearcolor
clearcolor = [np.random.random(),0,0,0]
time.sleep(0.1)
refresh() |
def _save_obj(self, *args, **kwargs):
pass | def union(matches, root, cwd):
"""Union a list of matchers.
If the list is empty, return nevermatcher.
If the list only contains one non-None value, return that matcher.
Otherwise return a union matcher.
"""
matches = list(filter(None, matches))
if len(matches) == 0:
return nevermatcher(root, cwd)
elif len(matches) == 1:
return matches[0]
else:
return unionmatcher(matches) |
def _save_obj(self, *args, **kwargs):
pass | def update(dt=0):
global projpts, rgb, depth
depth,_ = freenect.sync_get_depth()
rgb,_ = freenect.sync_get_video()
q = depth
X,Y = np.meshgrid(range(640),range(480))
# YOU CAN CHANGE THIS AND RERUN THE PROGRAM!
# Point cloud downsampling
d = 4
projpts = calibkinect.depth2xyzuv(q[::d,::d],X[::d,::d],Y[::d,::d])
refresh() |
def _save_obj(self, *args, **kwargs):
pass | def badmatch(match, badfn):
"""Make a copy of the given matcher, replacing its bad method with the given
one.
"""
m = copy.copy(match)
m.bad = badfn
return m |
def _save_obj(self, *args, **kwargs):
pass | def update_join():
update_on()
try:
_thread.join()
except:
update_off() |
def _save_obj(self, *args, **kwargs):
pass | def _donormalize(patterns, default, root, cwd, auditor, warn):
"""Convert 'kind:pat' from the patterns list to tuples with kind and
normalized and rooted patterns and with listfiles expanded."""
kindpats = []
for kind, pat in [_patsplit(p, default) for p in patterns]:
if kind in cwdrelativepatternkinds:
pat = pathutil.canonpath(root, cwd, pat, auditor)
elif kind in ("relglob", "path", "rootfilesin"):
pat = util.normpath(pat)
elif kind in ("listfile", "listfile0"):
try:
files = decodeutf8(util.readfile(pat))
if kind == "listfile0":
files = files.split("\0")
else:
files = files.splitlines()
files = [f for f in files if f]
except EnvironmentError:
raise error.Abort(_("unable to read file list (%s)") % pat)
for k, p, source in _donormalize(files, default, root, cwd, auditor, warn):
kindpats.append((k, p, pat))
continue
elif kind == "include":
try:
fullpath = os.path.join(root, util.localpath(pat))
includepats = readpatternfile(fullpath, warn)
for k, p, source in _donormalize(
includepats, default, root, cwd, auditor, warn
):
kindpats.append((k, p, source or pat))
except error.Abort as inst:
raise error.Abort("%s: %s" % (pat, inst[0]))
except IOError as inst:
if warn:
warn(
_("skipping unreadable pattern file '%s': %s\n")
% (pat, inst.strerror)
)
continue
# else: re or relre - which cannot be normalized
kindpats.append((kind, pat, ""))
return kindpats |
def _save_obj(self, *args, **kwargs):
pass | def update_on():
global _updating
if not '_updating' in globals(): _updating = False
if _updating: return |
def _save_obj(self, *args, **kwargs):
pass | def _testrefastpath(repat):
"""Test if a re pattern can use fast path.
That is, for every "$A/$B" path the pattern matches, "$A" must also be
matched,
Return True if we're sure it is. Return False otherwise.
"""
# XXX: It's very hard to implement this. These are what need to be
# supported in production and tests. Very hacky. But we plan to get rid
# of re matchers eventually.
# Rules like "(?!experimental/)"
if repat.startswith("(?!") and repat.endswith(")") and repat.count(")") == 1:
return True
# Rules used in doctest
if repat == "(i|j)$":
return True
return False |
def _save_obj(self, *args, **kwargs):
pass | def _run():
while _updating:
update() |
def _save_obj(self, *args, **kwargs):
pass | def _globpatsplit(pat):
"""Split a glob pattern. Return a list.
A naive version is "path.split("/")". This function handles more cases, like
"{*,{a,b}*/*}".
>>> _globpatsplit("*/**/x/{a,b/c}")
['*', '**', 'x', '{a,b/c}']
"""
result = []
buf = ""
parentheses = 0
for ch in pat:
if ch == "{":
parentheses += 1
elif ch == "}":
parentheses -= 1
if parentheses == 0 and ch == "/":
if buf:
result.append(buf)
buf = ""
else:
buf += ch
if buf:
result.append(buf)
return result |
def _save_obj(self, *args, **kwargs):
pass | def update_off():
global _updating
_updating = False |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, *args, **kwargs):
# If True, avoid entering subdirectories, and match everything recursively,
# unconditionally.
self.matchrecursive = False
# If True, avoid entering subdirectories, and return "unsure" for
# everything. This is set to True when complex re patterns (potentially
# including "/") are used.
self.unsurerecursive = False
# Patterns for matching paths in this directory.
self._kindpats = []
# Glob patterns used to match parent directories of another glob
# pattern.
self._globdirpats = []
super(_tree, self).__init__(*args, **kwargs) |
def _save_obj(self, *args, **kwargs):
pass | def loopcv():
import cv
while 1:
cv.ShowImage('hi',get_depth().astype(np.uint8))
cv.WaitKey(10) |
def _save_obj(self, *args, **kwargs):
pass | def insert(self, path, matchrecursive=True, globpats=None, repats=None):
"""Insert a directory path to this tree.
If matchrecursive is True, mark the directory as unconditionally
include files and subdirs recursively.
If globpats or repats are specified, append them to the patterns being
applied at this directory. The tricky part is those patterns can match
"x/y/z" and visit("x"), visit("x/y") need to return True, while we
still want visit("x/a") to return False.
"""
if path == "":
self.matchrecursive |= matchrecursive
if globpats:
# Need to match parent directories too.
for pat in globpats:
components = _globpatsplit(pat)
parentpat = ""
for comp in components:
if parentpat:
parentpat += "/"
parentpat += comp
if "/" in comp:
# Giving up - fallback to slow paths.
self.unsurerecursive = True
self._globdirpats.append(parentpat)
if any("**" in p for p in globpats):
# Giving up - "**" matches paths including "/"
self.unsurerecursive = True
self._kindpats += [("glob", pat, "") for pat in globpats]
if repats:
if not all(map(_testrefastpath, repats)):
# Giving up - fallback to slow paths.
self.unsurerecursive = True
self._kindpats += [("re", pat, "") for pat in repats]
return
subdir, rest = self._split(path)
self.setdefault(subdir, _tree()).insert(rest, matchrecursive, globpats, repats) |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, path):
"""Similar to matcher.visitdir"""
path = normalizerootdir(path, "visitdir")
if self.matchrecursive:
return "all"
elif self.unsurerecursive:
return True
elif path == "":
return True
if self._kindpats and self._compiledpats(path):
# XXX: This is incorrect. But re patterns are already used in
# production. We should kill them!
# Need to test "if every string starting with 'path' matches".
# Obviously it's impossible to test *every* string with the
# standard regex API, therefore pick a random strange path to test
# it approximately.
if self._compiledpats("%s/*/_/-/0/*" % path):
return "all"
else:
return True
if self._globdirpats and self._compileddirpats(path):
return True
subdir, rest = self._split(path)
subtree = self.get(subdir)
if subtree is None:
return False
else:
return subtree.visitdir(rest) |
def _save_obj(self, *args, **kwargs):
pass | def _compiledpats(self):
pat, matchfunc = _buildregexmatch(self._kindpats, "")
return matchfunc |
def _save_obj(self, *args, **kwargs):
pass | def _compileddirpats(self):
pat, matchfunc = _buildregexmatch(
[("glob", p, "") for p in self._globdirpats], "$"
)
return matchfunc |
def _save_obj(self, *args, **kwargs):
pass | def _split(self, path):
if "/" in path:
subdir, rest = path.split("/", 1)
else:
subdir, rest = path, ""
if not subdir:
raise error.ProgrammingError("path cannot be absolute")
return subdir, rest |
def _save_obj(self, *args, **kwargs):
pass | def _remainingpats(pat, prefix):
"""list of patterns with prefix stripped
>>> _remainingpats("a/b/c", "")
['a/b/c']
>>> _remainingpats("a/b/c", "a")
['b/c']
>>> _remainingpats("a/b/c", "a/b")
['c']
>>> _remainingpats("a/b/c", "a/b/c")
[]
>>> _remainingpats("", "")
[]
"""
if prefix:
if prefix == pat:
return []
else:
assert pat[len(prefix)] == "/"
return [pat[len(prefix) + 1 :]]
else:
if pat:
return [pat]
else:
return [] |
def _save_obj(self, *args, **kwargs):
pass | def _buildvisitdir(kindpats):
"""Try to build an efficient visitdir function
Return a visitdir function if it's built. Otherwise return None
if there are unsupported patterns.
>>> _buildvisitdir([('include', 'foo', '')])
>>> _buildvisitdir([('relglob', 'foo', '')])
>>> t = _buildvisitdir([
... ('glob', 'a/b', ''),
... ('glob', 'c/*.d', ''),
... ('glob', 'e/**/*.c', ''),
... ('re', '^f/(?!g)', ''), # no "$", only match prefix
... ('re', '^h/(i|j)$', ''),
... ('glob', 'i/a*/b*/c*', ''),
... ('glob', 'i/a5/b7/d', ''),
... ('glob', 'j/**.c', ''),
... ])
>>> t('a')
True
>>> t('a/b')
'all'
>>> t('a/b/c')
'all'
>>> t('c')
True
>>> t('c/d')
False
>>> t('c/rc.d')
'all'
>>> t('c/rc.d/foo')
'all'
>>> t('e')
True
>>> t('e/a')
True
>>> t('e/a/b.c')
True
>>> t('e/a/b.d')
True
>>> t('f')
True
>>> t('f/g')
False
>>> t('f/g2')
False
>>> t('f/g/a')
False
>>> t('f/h')
'all'
>>> t('f/h/i')
'all'
>>> t('h/i')
True
>>> t('h/i/k')
False
>>> t('h/k')
False
>>> t('i')
True
>>> t('i/a1')
True
>>> t('i/b2')
False
>>> t('i/a/b2/c3')
'all'
>>> t('i/a/b2/d4')
False
>>> t('i/a5/b7/d')
'all'
>>> t('j/x/y')
True
>>> t('z')
False
"""
tree = _tree()
for kind, pat, _source in kindpats:
if kind == "glob":
components = []
for p in pat.split("/"):
if "[" in p or "{" in p or "*" in p or "?" in p:
break
components.append(p)
prefix = "/".join(components)
matchrecursive = prefix == pat
tree.insert(
prefix,
matchrecursive=matchrecursive,
globpats=_remainingpats(pat, prefix),
)
elif kind == "re":
# Still try to get a plain prefix from the regular expression so we
# can still have fast paths.
if pat.startswith("^"):
# "re" already matches from the beginning, unlike "relre"
pat = pat[1:]
components = []
for p in pat.split("/"):
if re.escape(p) != p:
# contains special characters
break
components.append(p)
prefix = "/".join(components)
tree.insert(
prefix, matchrecursive=False, repats=_remainingpats(pat, prefix)
)
else:
# Unsupported kind
return None
return tree.visitdir |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, badfn=None, relativeuipath=True):
self._root = root
self._cwd = cwd
if badfn is not None:
self.bad = badfn
self._relativeuipath = relativeuipath |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<%s>" % self.__class__.__name__ |
def _save_obj(self, *args, **kwargs):
pass | def __call__(self, fn):
return self.matchfn(fn) |
def _save_obj(self, *args, **kwargs):
pass | def __iter__(self):
for f in self._files:
yield f |
def _save_obj(self, *args, **kwargs):
pass | def bad(self, f, msg):
"""Callback from dirstate.walk for each explicit file that can't be
found/accessed, with an error message.""" |
def _save_obj(self, *args, **kwargs):
pass | def abs(self, f):
"""Convert a repo path back to path that is relative to the root of the
matcher."""
return f |
def _save_obj(self, *args, **kwargs):
pass | def rel(self, f):
"""Convert repo path back to path that is relative to cwd of matcher."""
return util.pathto(self._root, self._cwd, f) |
def _save_obj(self, *args, **kwargs):
pass | def uipath(self, f):
"""Convert repo path to a display path. If patterns or -I/-X were used
to create this matcher, the display path will be relative to cwd.
Otherwise it is relative to the root of the repo."""
return (self._relativeuipath and self.rel(f)) or self.abs(f) |
def _save_obj(self, *args, **kwargs):
pass | def files(self):
"""Explicitly listed files or patterns or roots:
if no patterns or .always(): empty list,
if exact: list exact files,
if not .anypats(): list all files and dirs,
else: optimal roots"""
return self._files |
def _save_obj(self, *args, **kwargs):
pass | def _fileset(self):
return set(self._files) |
def _save_obj(self, *args, **kwargs):
pass | def exact(self, f):
"""Returns True if f is in .files()."""
return f in self._fileset |
def _save_obj(self, *args, **kwargs):
pass | def matchfn(self, f):
return False |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
"""Decides whether a directory should be visited based on whether it
has potential matches in it or one of its subdirectories. This is
based on the match's primary, included, and excluded patterns.
Returns the string 'all' if the given directory and all subdirectories
should be visited. Otherwise returns True or False indicating whether
the given directory should be visited.
"""
return True |
def _save_obj(self, *args, **kwargs):
pass | def always(self):
"""Matcher will match everything and .files() will be empty --
optimization might be possible."""
return False |
def _save_obj(self, *args, **kwargs):
pass | def isexact(self):
"""Matcher will match exactly the list of files in .files() --
optimization might be possible."""
return False |
def _save_obj(self, *args, **kwargs):
pass | def prefix(self):
"""Matcher will match the paths in .files() recursively --
optimization might be possible."""
return False |
def _save_obj(self, *args, **kwargs):
pass | def anypats(self):
"""None of .always(), .isexact(), and .prefix() is true --
optimizations will be difficult."""
return not self.always() and not self.isexact() and not self.prefix() |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, badfn=None, relativeuipath=False):
super(alwaysmatcher, self).__init__(
root, cwd, badfn, relativeuipath=relativeuipath
) |
def _save_obj(self, *args, **kwargs):
pass | def always(self):
return True |
def _save_obj(self, *args, **kwargs):
pass | def matchfn(self, f):
return True |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
return "all" |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<alwaysmatcher>" |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, badfn=None):
super(nevermatcher, self).__init__(root, cwd, badfn) |
def _save_obj(self, *args, **kwargs):
pass | def isexact(self):
return True |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
return False |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<nevermatcher>" |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, badfn=None, gitignorepaths=None):
super(gitignorematcher, self).__init__(root, cwd, badfn)
gitignorepaths = gitignorepaths or []
self._matcher = pathmatcher.gitignorematcher(root, gitignorepaths) |
def _save_obj(self, *args, **kwargs):
pass | def matchfn(self, f):
# XXX: is_dir is set to True here for performance.
# It should be set to whether "f" is actually a directory or not.
return self._matcher.match_relative(f, True) |
def _save_obj(self, *args, **kwargs):
pass | def explain(self, f):
return self._matcher.explain(f, True) |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
matched = self._matcher.match_relative(dir, True)
if matched:
# Everything in the directory is selected (ignored)
return "all"
else:
# Not sure
return True |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<gitignorematcher>" |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, badfn=None, rules=[]):
super(treematcher, self).__init__(root, cwd, badfn)
rules = list(rules)
self._matcher = pathmatcher.treematcher(rules)
self._rules = rules |
def _save_obj(self, *args, **kwargs):
pass | def matchfn(self, f):
return self._matcher.matches(f) |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
matched = self._matcher.match_recursive(dir)
if matched is None:
return True
elif matched is True:
return "all"
else:
assert matched is False
return False |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<treematcher rules=%r>" % self._rules |
def _save_obj(self, *args, **kwargs):
pass | def normalizerootdir(dir, funcname):
if dir == ".":
util.nouideprecwarn(
"match.%s() no longer accepts '.', use '' instead." % funcname, "20190805"
)
return ""
return dir |
def _save_obj(self, *args, **kwargs):
pass | def _kindpatstoglobs(kindpats, recursive=False):
"""Attempt to convert 'kindpats' to glob patterns that can be used in a
treematcher.
kindpats should be already normalized to be relative to repo root.
If recursive is True, `glob:a*` will match both `a1/b` and `a1`, otherwise
`glob:a*` will only match `a1` but not `a1/b`.
Return None if there are unsupported patterns (ex. regular expressions).
"""
if not _usetreematcher:
return None
globs = []
for kindpat in kindpats:
kind, pat = kindpat[0:2]
if kind == "re":
# Attempt to convert the re pat to globs
reglobs = _convertretoglobs(pat)
if reglobs is not None:
globs += reglobs
else:
return None
elif kind == "glob":
# The treematcher (man gitignore) does not support csh-style
# brackets (ex. "{a,b,c}"). Expand the brackets to patterns.
for subpat in pathmatcher.expandcurlybrackets(pat):
normalized = pathmatcher.normalizeglob(subpat)
if recursive:
normalized = _makeglobrecursive(normalized)
globs.append(normalized)
elif kind == "path":
if pat == ".":
# Special case. Comes from `util.normpath`.
pat = ""
else:
pat = pathmatcher.plaintoglob(pat)
pat = _makeglobrecursive(pat)
globs.append(pat)
else:
return None
return globs |
def _save_obj(self, *args, **kwargs):
pass | def _makeglobrecursive(pat):
"""Make a glob pattern recursive by appending "/**" to it"""
if pat.endswith("/") or not pat:
return pat + "**"
else:
return pat + "/**" |
def _save_obj(self, *args, **kwargs):
pass | def _convertretoglobs(repat):
"""Attempt to convert a regular expression pattern to glob patterns.
A single regular expression pattern might be converted into multiple
glob patterns.
Return None if conversion is unsupported.
>>> _convertretoglobs("abc*") is None
True
>>> _convertretoglobs("xx/yy/(?!zz/kk)")
['xx/yy/**', '!xx/yy/zz/kk/**']
>>> _convertretoglobs("x/y/(?:.*/)?BUCK")
['x/y/**/BUCK']
"""
m = _repat1.match(repat)
if m:
prefix, excluded = m.groups()
return ["%s/**" % prefix, "!%s/%s/**" % (prefix, excluded)]
m = _repat2.match(repat)
if m:
prefix, name = m.groups()
return ["%s/**/%s" % (prefix, name)]
return None |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, kindpats, ctx=None, badfn=None):
super(patternmatcher, self).__init__(root, cwd, badfn)
# kindpats are already normalized to be relative to repo-root.
# Can we use tree matcher?
rules = _kindpatstoglobs(kindpats, recursive=False)
fallback = True
if rules is not None:
try:
matcher = treematcher(root, cwd, badfn=badfn, rules=rules)
# Replace self to 'matcher'.
self.__dict__ = matcher.__dict__
self.__class__ = matcher.__class__
fallback = False
except ValueError:
# for example, Regex("Compiled regex exceeds size limit of 10485760 bytes.")
pass
if fallback:
self._prefix = _prefix(kindpats)
self._pats, self.matchfn = _buildmatch(ctx, kindpats, "$", root)
self._files = _explicitfiles(kindpats) |
def _save_obj(self, *args, **kwargs):
pass | def _dirs(self):
return set(util.dirs(self._fileset)) |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
if self._prefix and dir in self._fileset:
return "all"
if not self._prefix:
return True
return (
dir in self._fileset
or dir in self._dirs
or any(parentdir in self._fileset for parentdir in util.finddirs(dir))
) |
def _save_obj(self, *args, **kwargs):
pass | def prefix(self):
return self._prefix |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<patternmatcher patterns=%r>" % self._pats |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, kindpats, ctx=None, badfn=None):
super(includematcher, self).__init__(root, cwd, badfn)
# Can we use tree matcher?
rules = _kindpatstoglobs(kindpats, recursive=True)
fallback = True
if rules is not None:
try:
matcher = treematcher(root, cwd, badfn=badfn, rules=rules)
# Replace self to 'matcher'.
self.__dict__ = matcher.__dict__
self.__class__ = matcher.__class__
fallback = False
except ValueError:
# for example, Regex("Compiled regex exceeds size limit of 10485760 bytes.")
pass
if fallback:
self._pats, self.matchfn = _buildmatch(ctx, kindpats, "(?:/|$)", root)
# prefix is True if all patterns are recursive, so certain fast paths
# can be enabled. Unfortunately, it's too easy to break it (ex. by
# using "glob:*.c", "re:...", etc).
self._prefix = _prefix(kindpats)
roots, dirs = _rootsanddirs(kindpats)
# roots are directories which are recursively included.
# If self._prefix is True, then _roots can have a fast path for
# visitdir to return "all", marking things included unconditionally.
# If self._prefix is False, then that optimization is unsound because
# "roots" might contain entries that is not recursive (ex. roots will
# include "foo/bar" for pattern "glob:foo/bar/*.c").
self._roots = set(roots)
# dirs are directories which are non-recursively included.
# That is, files under that directory are included. But not
# subdirectories.
self._dirs = set(dirs)
# Try to use a more efficient visitdir implementation
visitdir = _buildvisitdir(kindpats)
if visitdir:
self.visitdir = visitdir |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
if self._prefix and dir in self._roots:
return "all"
return (
dir in self._roots
or dir in self._dirs
or any(parentdir in self._roots for parentdir in util.finddirs(dir))
) |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<includematcher includes=%r>" % self._pats |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, root, cwd, files, badfn=None):
super(exactmatcher, self).__init__(root, cwd, badfn)
if isinstance(files, list):
self._files = files
else:
self._files = list(files) |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
return dir in self._dirs |
def _save_obj(self, *args, **kwargs):
pass | def __repr__(self):
return "<exactmatcher files=%r>" % self._files |
def _save_obj(self, *args, **kwargs):
pass | def __init__(self, m1, m2):
super(differencematcher, self).__init__(m1._root, m1._cwd)
self._m1 = m1
self._m2 = m2
self.bad = m1.bad
self.traversedir = m1.traversedir |
def _save_obj(self, *args, **kwargs):
pass | def matchfn(self, f):
return self._m1(f) and (not self._m2(f) or self._m1.exact(f)) |
def _save_obj(self, *args, **kwargs):
pass | def _files(self):
if self.isexact():
return [f for f in self._m1.files() if self(f)]
# If m1 is not an exact matcher, we can't easily figure out the set of
# files, because its files() are not always files. For example, if
# m1 is "path:dir" and m2 is "rootfileins:.", we don't
# want to remove "dir" from the set even though it would match m2,
# because the "dir" in m1 may not be a file.
return self._m1.files() |
def _save_obj(self, *args, **kwargs):
pass | def visitdir(self, dir):
dir = normalizerootdir(dir, "visitdir")
if not self._m2.visitdir(dir):
return self._m1.visitdir(dir)
if self._m2.visitdir(dir) == "all":
# There's a bug here: If m1 matches file 'dir/file' and m2 excludes
# 'dir' (recursively), we should still visit 'dir' due to the
# exception we have for exact matches.
return False
return bool(self._m1.visitdir(dir)) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.