code stringlengths 3 6.57k |
|---|
kwargs.pop("params", {}) |
_SERIALIZER.query("api_version", api_version, 'str') |
kwargs.pop("headers", {}) |
_SERIALIZER.header("accept", accept, 'str') |
PrivateLinkResourcesOperations(object) |
__init__(self, client, config, serializer, deserializer) |
cls(response) |
kwargs.pop('cls', None) |
error_map.update(kwargs.pop('error_map', {}) |
kwargs.pop('api_version', "2022-04-01") |
_convert_request(request) |
self._client.format_url(request.url) |
map_error(status_code=response.status_code, response=response, error_map=error_map) |
HttpResponseError(response=response, error_format=ARMErrorFormat) |
self._deserialize('PrivateLinkResourcesListResult', pipeline_response) |
cls(pipeline_response, deserialized, {}) |
Migration(migrations.Migration) |
models.CharField(max_length=255) |
default_data_setup(sender, **kwargs) |
User.objects.get(username='ANONYMOUS_USER') |
print('Adding ANONYMOUS_USER') |
User.objects.create_user('ANONYMOUS_USER', 'anonymous_user@example.com') |
anon.set_unusable_password() |
anon.save() |
RadioConfig(AppConfig) |
ready(self) |
post_migrate.connect(default_data_setup, sender=self) |
or_list(booleans) |
get_ngrams(D) |
ngrams (aka a token containing a dollar sign ($) |
set() |
ngrams.add(w) |
list(ngrams) |
get_frequent_ngrams(text, n, stopword_list, threshold) |
ngrams(text, n) |
Counter(bigrams) |
bigram_freq.most_common() |
not (or_list([i in stopword_list for i in bigram]) |
frequent_bigrams.append('{}${}'.format(bigram[0], bigram[1]) |
ngrammize_text(text, ngrams) |
len(text) |
len(text) |
bigrammized_text.append(term) |
format(term, next_term) |
bigrammized_text.append(test_bigram) |
bigrammized_text.append(term) |
get_dataset_ngrams(docs, min_freq=1000, sw=None, extra_bigrams=None, extra_ngrams=None) |
stopwords.words('english') |
get_pp_pipeline(remove_stopwords=False) |
sw_pp.clean_document(sw) |
full_text.extend(doc) |
get_frequent_ngrams(full_text, 2, sw, min_freq) |
frequent_bigrams.extend(extra_bigrams) |
ngrammize_text(full_text, frequent_bigrams) |
get_frequent_ngrams(bigrammized_text, 2, sw, min_freq) |
frequent_ngrams.extend(extra_ngrams) |
insert_ngrams_flat_from_lists(docs, frequent_bigrams, frequent_ngrams) |
range(0, len(docs) |
ngrammize_text(doc, frequent_bigrams) |
ngrammize_text(doc, frequent_ngrams) |
insert_ngrams_flat(docs, min_freq=1000, sw=None, extra_bigrams=None, extra_ngrams=None) |
get_dataset_ngrams(docs, min_freq, sw, extra_bigrams, extra_ngrams) |
insert_ngrams_flat_from_lists(docs, fb, fn) |
insert_ngrams_from_lists(date_doc_tuples, frequent_bigrams, frequent_ngrams) |
range(0, len(date_doc_tuples) |
ngrammize_text(doc, frequent_bigrams) |
ngrammize_text(doc, frequent_ngrams) |
insert_ngrams(date_docs, min_freq=1000, sw=None, extra_bigrams=None, extra_ngrams=None) |
get_dataset_ngrams([x[1] for x in date_docs], min_freq, sw, extra_bigrams, extra_ngrams) |
insert_ngrams_from_lists(date_docs, fb, fn) |
collections.namedtuple('Results', 'model reward config time metadata') |
logging.getLogger(__name__) |
logger.setLevel(logging.WARNING) |
create_scheduler(train_fn, scheduler, scheduler_options) |
isinstance(scheduler, str) |
scheduler.lower() |
callable(scheduler) |
copy.copy(scheduler_options) |
scheduler_cls(train_fn, **scheduler_options) |
BaseTask(object) |
Dataset() |
try_import_mxnet() |
time.time() |
create_scheduler(train_fn, search_strategy, scheduler_options) |
scheduler.run() |
scheduler.join_jobs() |
scheduler.get_best_reward() |
scheduler.get_best_config() |
hasattr(args, 'epochs') |
hasattr(args, 'final_fit_epochs') |
train_fn.args.update({'final_fit':True}) |
train_fn.kwvars.update({'final_fit':True}) |
create_scheduler(train_fn, search_strategy, scheduler_options) |
scheduler_final.run_with_config(best_config) |
time.time() |
in_ipynb() |
replace('exp1.ag', 'plot_training_curves.png') |
scheduler.get_training_curves(filename=plot_training_curves, plot=True, use_legend=False) |
copy.deepcopy(args) |
logger.warning('No valid results obtained with best config, the result may not be useful...') |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.