text
stringlengths
5
22M
id
stringlengths
12
177
metadata
dict
__index_level_0__
int64
0
1.37k
<?xml version='1.0' encoding='UTF-8'?> <glyph name="view-comfy" format="2"> <advance width="1200"/> <unicode hex="F0E6A"/> <note> view-comfy </note> <outline> <contour> <point x="20" y="1137" type="line"/> <point x="20" y="893" type="line"/> <point x="263" y="893" type="line"/> <poin...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-comfy.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-comfy.glif", "repo_id": "cascadia-code", "token_count": 1270 }
1,333
<?xml version='1.0' encoding='UTF-8'?> <glyph name="view-grid" format="2"> <advance width="1200"/> <unicode hex="F0570"/> <note> view-grid </note> <outline> <contour> <point x="20" y="1290" type="line"/> <point x="20" y="773" type="line"/> <point x="537" y="773" type="line"/> <point ...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-grid.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-grid.glif", "repo_id": "cascadia-code", "token_count": 472 }
1,334
<?xml version='1.0' encoding='UTF-8'?> <glyph name="wall-sconce-flat-variant-outline" format="2"> <advance width="1200"/> <unicode hex="F17CA"/> <note> wall-sconce-flat-variant-outline </note> <outline> <contour> <point x="1077" y="643" type="line"/> <point x="123" y="643" type="line"/> <p...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wall-sconce-flat-variant-outline.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wall-sconce-flat-variant-outline.glif", "repo_id": "cascadia-code", "token_count": 596 }
1,335
<?xml version='1.0' encoding='UTF-8'?> <glyph name="wallet" format="2"> <advance width="1200"/> <unicode hex="F0584"/> <note> wallet </note> <outline> <contour> <point x="570" y="343" type="line" smooth="yes"/> <point x="518" y="343"/> <point x="447" y="415"/> <point x="447" y="467" ...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wallet.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wallet.glif", "repo_id": "cascadia-code", "token_count": 917 }
1,336
<?xml version='1.0' encoding='UTF-8'?> <glyph name="weather-hurricane" format="2"> <advance width="1200"/> <unicode hex="F0898"/> <note> weather-hurricane </note> <outline> <contour> <point x="787" y="1093"/> <point x="770" y="1150"/> <point x="783" y="1177" type="qcurve" smooth="yes"/> ...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/weather-hurricane.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/weather-hurricane.glif", "repo_id": "cascadia-code", "token_count": 1035 }
1,337
<?xml version='1.0' encoding='UTF-8'?> <glyph name="window-minimize" format="2"> <advance width="1200"/> <unicode hex="F05B0"/> <note> window-minimize </note> <outline> <contour> <point x="1180" y="856" type="line"/> <point x="20" y="856" type="line"/> <point x="20" y="564" type="line"/> ...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/window-minimize.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/window-minimize.glif", "repo_id": "cascadia-code", "token_count": 185 }
1,338
<?xml version='1.0' encoding='UTF-8'?> <glyph name="zodiac-scorpio" format="2"> <advance width="1200"/> <unicode hex="F0A86"/> <note> zodiac-scorpio </note> <outline> <contour> <point x="1180" y="323" type="line"/> <point x="947" y="557" type="line"/> <point x="858" y="468" type="line"/> ...
cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/zodiac-scorpio.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/zodiac-scorpio.glif", "repo_id": "cascadia-code", "token_count": 1042 }
1,339
<?xml version='1.0' encoding='UTF-8'?> <glyph name="POMODORO_TICKING" format="2"> <advance width="1200"/> <unicode hex="E003"/> <note> POMODORO_TICKING </note> <outline> <contour> <point x="644" y="660" type="line"/> <point x="644" y="297" type="line"/> <point x="546" y="297" type="line"/>...
cascadia-code/sources/nerdfonts/full/processed/Pomicons.ufo/glyphs/P_O_M_O_D_O_R_O__T_I_C_K_I_N_G_.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/Pomicons.ufo/glyphs/P_O_M_O_D_O_R_O__T_I_C_K_I_N_G_.glif", "repo_id": "cascadia-code", "token_count": 1304 }
1,340
<?xml version='1.0' encoding='UTF-8'?> <glyph name="uniE0B2" format="2"> <advance width="1200"/> <unicode hex="E0B2"/> <note> uniE0B2 </note> <outline> <contour> <point x="-122" y="942" type="line"/> <point x="1322" y="-412" type="line"/> <point x="1322" y="2294" type="line"/> </contou...
cascadia-code/sources/nerdfonts/full/processed/PowerlineExtraSymbols.ufo/glyphs/uniE_0B_2.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/PowerlineExtraSymbols.ufo/glyphs/uniE_0B_2.glif", "repo_id": "cascadia-code", "token_count": 167 }
1,341
<?xml version='1.0' encoding='UTF-8'?> <glyph name="activate-breakpoints" format="2"> <advance width="1200"/> <unicode hex="EA97"/> <note> activate-breakpoints </note> <outline> <contour> <point x="1180" y="993"/> <point x="1126" y="1130"/> <point x="1020" y="1237"/> <point x="883" y...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/activate-breakpoints.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/activate-breakpoints.glif", "repo_id": "cascadia-code", "token_count": 873 }
1,342
<?xml version='1.0' encoding='UTF-8'?> <glyph name="check-all" format="2"> <advance width="1200"/> <unicode hex="EBB1"/> <note> check-all </note> <outline> <contour> <point x="1129" y="1074" type="line"/> <point x="578" y="425" type="line"/> <point x="372" y="714" type="line"/> <poin...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/check-all.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/check-all.glif", "repo_id": "cascadia-code", "token_count": 552 }
1,343
<?xml version='1.0' encoding='UTF-8'?> <glyph name="debug-breakpoint-log-unverified" format="2"> <advance width="1200"/> <unicode hex="EAAA"/> <note> debug-breakpoint-log-unverified </note> <outline> <contour> <point x="600" y="127" type="line"/> <point x="1180" y="707" type="line"/> <poin...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-breakpoint-log-unverified.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-breakpoint-log-unverified.glif", "repo_id": "cascadia-code", "token_count": 287 }
1,344
<?xml version='1.0' encoding='UTF-8'?> <glyph name="debug-stackframe" format="2"> <advance width="1200"/> <unicode hex="EB8B"/> <note> debug-stackframe </note> <outline> <contour> <point x="1180" y="627" type="line"/> <point x="1180" y="793" type="line"/> <point x="750" y="1271" type="line...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-stackframe.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-stackframe.glif", "repo_id": "cascadia-code", "token_count": 423 }
1,345
<?xml version='1.0' encoding='UTF-8'?> <glyph name="diff-renamed" format="2"> <advance width="1200"/> <unicode hex="EAE0"/> <note> diff-renamed </note> <outline> <contour> <point x="20" y="1250" type="line"/> <point x="20" y="170" type="line"/> <point x="60" y="130" type="line"/> <po...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/diff-renamed.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/diff-renamed.glif", "repo_id": "cascadia-code", "token_count": 522 }
1,346
<?xml version='1.0' encoding='UTF-8'?> <glyph name="file-code" format="2"> <advance width="1200"/> <unicode hex="EAE9"/> <unicode hex="EAFB"/> <note> file-code </note> <outline> <contour> <point x="814" y="1385" type="line"/> <point x="66" y="1385" type="line"/> <point x="20" y="1339" ty...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/file-code.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/file-code.glif", "repo_id": "cascadia-code", "token_count": 875 }
1,347
<?xml version='1.0' encoding='UTF-8'?> <glyph name="folder-library" format="2"> <advance width="1200"/> <unicode hex="EBDF"/> <note> folder-library </note> <outline> <contour> <point x="507" y="1195" type="line"/> <point x="476" y="1208" type="line"/> <point x="60" y="1208" type="line"/> ...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/folder-library.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/folder-library.glif", "repo_id": "cascadia-code", "token_count": 1005 }
1,348
<?xml version='1.0' encoding='UTF-8'?> <glyph name="mail-read" format="2"> <advance width="1200"/> <unicode hex="EB1B"/> <note> mail-read </note> <outline> <contour> <point x="1180" y="894" type="line"/> <point x="622" y="1226" type="line"/> <point x="578" y="1226" type="line"/> <poi...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/mail-read.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/mail-read.glif", "repo_id": "cascadia-code", "token_count": 565 }
1,349
<?xml version='1.0' encoding='UTF-8'?> <glyph name="redo" format="2"> <advance width="1200"/> <unicode hex="EBB0"/> <note> redo </note> <outline> <contour> <point x="1013" y="1420" type="line"/> <point x="1013" y="1193" type="line"/> <point x="938" y="1269" type="line" smooth="yes"/> ...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/redo.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/redo.glif", "repo_id": "cascadia-code", "token_count": 613 }
1,350
<?xml version='1.0' encoding='UTF-8'?> <glyph name="report" format="2"> <advance width="1200"/> <unicode hex="EB42"/> <note> report </note> <outline> <contour> <point x="20" y="1246" type="line"/> <point x="20" y="413" type="line"/> <point x="60" y="374" type="line"/> <point x="268" ...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/report.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/report.glif", "repo_id": "cascadia-code", "token_count": 735 }
1,351
<?xml version='1.0' encoding='UTF-8'?> <glyph name="star-empty" format="2"> <advance width="1200"/> <unicode hex="EA6A"/> <note> star-empty </note> <outline> <contour> <point x="1180" y="856" type="line"/> <point x="733" y="856" type="line"/> <point x="600" y="1290" type="line"/> <po...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/star-empty.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/star-empty.glif", "repo_id": "cascadia-code", "token_count": 740 }
1,352
<?xml version='1.0' encoding='UTF-8'?> <glyph name="symbol-keyword" format="2"> <advance width="1200"/> <unicode hex="EB62"/> <note> symbol-keyword </note> <outline> <contour> <point x="1180" y="1124" type="line"/> <point x="768" y="1124" type="line"/> <point x="768" y="1040" type="line"/>...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/symbol-keyword.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/symbol-keyword.glif", "repo_id": "cascadia-code", "token_count": 869 }
1,353
<?xml version='1.0' encoding='UTF-8'?> <glyph name="verified" format="2"> <advance width="1200"/> <unicode hex="EB77"/> <note> verified </note> <outline> <contour> <point x="427" y="276" type="line"/> <point x="570" y="128" type="line"/> <point x="630" y="128" type="line"/> <point x=...
cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/verified.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/verified.glif", "repo_id": "cascadia-code", "token_count": 1261 }
1,354
<?xml version='1.0' encoding='UTF-8'?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>ascender</key> <integer>410</integer> <key>capHeight</key> <integer>0</integer> <key>copyright</key> <string>Copyright...
cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/fontinfo.plist/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/fontinfo.plist", "repo_id": "cascadia-code", "token_count": 1282 }
1,355
<?xml version='1.0' encoding='UTF-8'?> <glyph name="KDE Neon" format="2"> <advance width="1200"/> <unicode hex="F331"/> <note> KDE Neon </note> <outline> <contour> <point x="568" y="1267" type="qcurve"/> <point x="373" y="1256"/> <point x="228" y="1127" type="qcurve"/> <point x="224"...
cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/K_D_E_ N_eon.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/K_D_E_ N_eon.glif", "repo_id": "cascadia-code", "token_count": 3095 }
1,356
<?xml version='1.0' encoding='UTF-8'?> <glyph name="Puppy Linux" format="2"> <advance width="1200"/> <unicode hex="F341"/> <note> Puppy Linux </note> <outline> <contour> <point x="753" y="1287"/> <point x="716" y="1277" type="qcurve"/> <point x="518" y="1317"/> <point x="341" y="1227...
cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/P_uppy L_inux.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/P_uppy L_inux.glif", "repo_id": "cascadia-code", "token_count": 3748 }
1,357
<?xml version='1.0' encoding='UTF-8'?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>AOSC OS</key> <string>A_O_S_C_ O_S_.glif</string> <key>Alma Linux</key> <string>A_lma L_inux.glif</string> <key>Alpine</ke...
cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/contents.plist/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/contents.plist", "repo_id": "cascadia-code", "token_count": 3778 }
1,358
<?xml version='1.0' encoding='UTF-8'?> <glyph name="i_seti_csv" format="2"> <advance width="1200"/> <unicode hex="E64A"/> <note> i_seti_csv </note> <outline> <contour> <point x="1180" y="1284" type="line"/> <point x="20" y="1284" type="line"/> <point x="20" y="136" type="line"/> <poi...
cascadia-code/sources/nerdfonts/full/processed/original-source.ufo/glyphs/i_seti_csv.glif/0
{ "file_path": "cascadia-code/sources/nerdfonts/full/processed/original-source.ufo/glyphs/i_seti_csv.glif", "repo_id": "cascadia-code", "token_count": 1042 }
1,359
{ "name": "Causica Devcontainer", "build": { "context": "..", "dockerfile": "../Dockerfile", "target": "dev" }, "runArgs": [ // give headspace for pytorch dataloaders passing tensors across processes "--shm-size=1gb" ], "containerEnv": { "DOCKER_BUILDKIT": "1" }, "mounts": [ "source=command_histor...
causica/.devcontainer/devcontainer.json/0
{ "file_path": "causica/.devcontainer/devcontainer.json", "repo_id": "causica", "token_count": 806 }
1,360
# Microsoft Open Source Code of Conduct This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). Resources: - [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/) - [Microsoft Code of Conduct FAQ](https://opensource.mic...
causica/CODE_OF_CONDUCT.md/0
{ "file_path": "causica/CODE_OF_CONDUCT.md", "repo_id": "causica", "token_count": 115 }
1,361
import torch from tensordict import TensorDict def check_temporal_tensordict_shapes(num_timepoints: int, tds: TensorDict) -> bool: """Check that the shapes within the TensorDict are consistent with the number of time points.""" return all(val.shape[len(tds.batch_size)] == num_timepoints for val in tds.values(...
causica/research_experiments/scotch/src/scotch/scotch_utils/scotch_utils.py/0
{ "file_path": "causica/research_experiments/scotch/src/scotch/scotch_utils/scotch_utils.py", "repo_id": "causica", "token_count": 192 }
1,362
import math import numpy as np import torch import torch.distributions as td from causica.distributions.adjacency.adjacency_distributions import AdjacencyDistribution from causica.triangular_transformations import fill_triangular class ErdosRenyiDAGDistribution(AdjacencyDistribution): """ An adjacency distr...
causica/src/causica/distributions/adjacency/directed_acyclic.py/0
{ "file_path": "causica/src/causica/distributions/adjacency/directed_acyclic.py", "repo_id": "causica", "token_count": 1716 }
1,363
from typing import Optional import torch import torch.distributions as td from torch import nn from causica.distributions.noise.noise import IndependentNoise, Noise, NoiseModule class UnivariateCauchyNoise(td.Cauchy, Noise[torch.Tensor]): def sample_to_noise(self, samples: torch.Tensor) -> torch.Tensor: ...
causica/src/causica/distributions/noise/univariate_cauchy.py/0
{ "file_path": "causica/src/causica/distributions/noise/univariate_cauchy.py", "repo_id": "causica", "token_count": 878 }
1,364
import pytest import torch import torch.distributions as td from causica.data_generation.samplers.functional_relationships_sampler import LinearRelationshipsSampler from causica.data_generation.samplers.noise_dist_sampler import ( BernoulliNoiseModuleSampler, CategoricalNoiseModuleSampler, JointNoiseModule...
causica/test/datasets/test_synthetic_dataset.py/0
{ "file_path": "causica/test/datasets/test_synthetic_dataset.py", "repo_id": "causica", "token_count": 1393 }
1,365
import pytest import torch from causica.distributions.noise import UnivariateLaplaceNoise @pytest.mark.parametrize(("batch", "dimension"), [(1, 10), (2, 5)]) def test_init(batch, dimension): mean = torch.randn((batch, dimension)) scale = torch.ones((batch, dimension)) noise_model = UnivariateLaplaceNoise...
causica/test/distributions/noise/test_univariate_laplace.py/0
{ "file_path": "causica/test/distributions/noise/test_univariate_laplace.py", "repo_id": "causica", "token_count": 843 }
1,366
import os from pathlib import Path import mlflow import pytest import yaml from jsonargparse import Namespace from mlflow import ActiveRun from pytorch_lightning import LightningModule, Trainer from pytorch_lightning.cli import LightningArgumentParser from pytorch_lightning.trainer.states import TrainerFn from causic...
causica/test/lightning/test_callbacks.py/0
{ "file_path": "causica/test/lightning/test_callbacks.py", "repo_id": "causica", "token_count": 503 }
1,367
# Copyright (c) Microsoft Corporation. # Licensed under the MIT license. import sys if sys.version_info >= (3, 8): from importlib import metadata else: import importlib_metadata as metadata __version__ = metadata.version("cliffordlayers")
cliffordlayers/cliffordlayers/__init__.py/0
{ "file_path": "cliffordlayers/cliffordlayers/__init__.py", "repo_id": "cliffordlayers", "token_count": 74 }
1,368
import torch import torch.nn.functional as F from torch.nn.modules.utils import _single, _pair, _triple def get_clifford_g3_kernel(weights): assert len(weights) == 6 scale = weights[4] square_1 = weights[0] * weights[0] square_2 = weights[1] * weights[1] square_3 = weights[2] * weights[2] sq...
cliffordlayers/cliffordlayers/nn/functional/cliffordg3conv.py/0
{ "file_path": "cliffordlayers/cliffordlayers/nn/functional/cliffordg3conv.py", "repo_id": "cliffordlayers", "token_count": 1316 }
1,369
[project] name = "cliffordlayers" dynamic = ["version"] description = "A PyTorch library for Clifford layers" readme = "README.md" requires-python = ">=3.8" license = {text = "MIT"} authors = [{name = "Jayesh K. Gupta", email = "mail@rejuvyesh.com"}, {name="Johannes Brandstetter", email="johannesb@microsoft.com"}, {nam...
cliffordlayers/pyproject.toml/0
{ "file_path": "cliffordlayers/pyproject.toml", "repo_id": "cliffordlayers", "token_count": 420 }
1,370
# Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License.
cookiecutter-spacy-fastapi/{{cookiecutter.project_slug}}/app/tests/__init__.py/0
{ "file_path": "cookiecutter-spacy-fastapi/{{cookiecutter.project_slug}}/app/tests/__init__.py", "repo_id": "cookiecutter-spacy-fastapi", "token_count": 21 }
1,371
build/ node_modules/ dist/ coverage/ # Client client/build/ client/node_modules/ client/dist/ client/coverage/
0xDeCA10B/demo/.eslintignore/0
{ "file_path": "0xDeCA10B/demo/.eslintignore", "repo_id": "0xDeCA10B", "token_count": 42 }
0
{ "name": "decai-demo-client", "version": "1.4.5", "license": "MIT", "private": true, "proxy": "http://localhost:5387/", "dependencies": { "@drizzle-utils/get-web3": "^0.2.2-alpha.0", "@material-ui/core": "^4.7.1", "@material-ui/icons": "^4.5.1", "@tensorflow-models/mobilenet": "^2.0.4", ...
0xDeCA10B/demo/client/package.json/0
{ "file_path": "0xDeCA10B/demo/client/package.json", "repo_id": "0xDeCA10B", "token_count": 933 }
1
pragma solidity ^0.6; pragma experimental ABIEncoderV2; import "../../../lib/SafeMath.sol"; import "../../../lib/SignedSafeMath.sol"; import {Ownable} from "../ownership/Ownable.sol"; /** * A classifier that can take a data sample as input and return a predict classification/label for the data. */ contract Classif...
0xDeCA10B/demo/client/src/contracts/classification/Classifier.sol/0
{ "file_path": "0xDeCA10B/demo/client/src/contracts/classification/Classifier.sol", "repo_id": "0xDeCA10B", "token_count": 1033 }
2
import * as _getWeb3 from '@drizzle-utils/get-web3' import Web3 from 'web3' // Only required for custom/fallback provider option. export async function getWeb3() { if (typeof window !== "undefined" && window.ethereum) { // Get rid of a warning about network refreshing. window.ethereum.autoRefreshOnNetworkChange =...
0xDeCA10B/demo/client/src/getWeb3.js/0
{ "file_path": "0xDeCA10B/demo/client/src/getWeb3.js", "repo_id": "0xDeCA10B", "token_count": 350 }
3
// Basically the same as tensor-utils.js but made for back-end. // There are fancier ways to share code for front-end and back-end // but they all seemed too complicated and unreliable in all cases. const tf = require('@tensorflow/tfjs-node') exports.normalize1d = function (x) { return tf.tidy(_ => { if (!(x insta...
0xDeCA10B/demo/client/src/ml-models/tensor-utils-node.js/0
{ "file_path": "0xDeCA10B/demo/client/src/ml-models/tensor-utils-node.js", "repo_id": "0xDeCA10B", "token_count": 311 }
4
const util = require('util') exports.setTimeoutPromise = util.promisify(setTimeout)
0xDeCA10B/demo/client/src/utils.js/0
{ "file_path": "0xDeCA10B/demo/client/src/utils.js", "repo_id": "0xDeCA10B", "token_count": 27 }
5
{ "compilerOptions": { "target": "es2020", "lib": [ "dom", "dom.iterable", "esnext" ], "allowJs": true, "skipLibCheck": true, "esModuleInterop": true, "allowSyntheticDefaultImports": true, "strict": true, "forceConsistentCasingInFileNames": true, "module": "es...
0xDeCA10B/demo/client/tsconfig.json/0
{ "file_path": "0xDeCA10B/demo/client/tsconfig.json", "repo_id": "0xDeCA10B", "token_count": 231 }
6
from collections import defaultdict from dataclasses import dataclass, field from typing import Dict import numpy as np from injector import inject, singleton from decai.simulation.contract.objects import Address, RejectException, SmartContract, TimeMock @dataclass class StoredData: # Storing the data is not ne...
0xDeCA10B/simulation/decai/simulation/contract/data/data_handler.py/0
{ "file_path": "0xDeCA10B/simulation/decai/simulation/contract/data/data_handler.py", "repo_id": "0xDeCA10B", "token_count": 1965 }
7
import unittest from decai.simulation.data.featuremapping.hashing.murmurhash3 import MurmurHash3 class TestMurmurHash3(unittest.TestCase): @classmethod def setUpClass(cls): cls.h = MurmurHash3() def test_classifications(self): h = self.h.hash("hey") assert type(h) == int ...
0xDeCA10B/simulation/decai/simulation/data/featuremapping/hashing/tests/test_murmurhash3.py/0
{ "file_path": "0xDeCA10B/simulation/decai/simulation/data/featuremapping/hashing/tests/test_murmurhash3.py", "repo_id": "0xDeCA10B", "token_count": 168 }
8
import logging from dataclasses import dataclass, field from logging import Logger from injector import Module, provider, singleton @dataclass class LoggingModule(Module): _log_level: int = field(default=logging.INFO) @provider @singleton def provide_logger(self) -> Logger: result = logging....
0xDeCA10B/simulation/decai/simulation/logging_module.py/0
{ "file_path": "0xDeCA10B/simulation/decai/simulation/logging_module.py", "repo_id": "0xDeCA10B", "token_count": 246 }
9
# Lab 3 - CUDA实现和优化 ## 实验目的 1. 理解DNN框架中的张量运算在GPU加速器上加速原理。 2. 通过CUDA实现和优化一个定制化张量运算。 ## 实验环境 * PyTorch==1.5.0 * CUDA 10.0 ## 实验原理 1. 矩阵运算与计算机体系结构 2. GPU加速器的加速原理 ## 实验内容 ### 实验流程图 ![](/imgs/Lab3-flow.png "Lab3 flow chat") ### 具体步骤 1. 理解PyTorch中Linear张量运算的计算过程,推导计算公式 2. 了解GPU端加速的原理,CUDA内核编程和实现一个kernel的原理 3. 实...
AI-System/Labs/BasicLabs/Lab3/README.md/0
{ "file_path": "AI-System/Labs/BasicLabs/Lab3/README.md", "repo_id": "AI-System", "token_count": 1565 }
10
# What image do you want to start building on? # Make a folder in your image where your app's source code can live # Tell your container where your app's source code will live # What source code do you what to copy, and where to put it? # Does your app have any dependencies that should be installed? # What po...
AI-System/Labs/BasicLabs/Lab5/Dockerfile.template/0
{ "file_path": "AI-System/Labs/BasicLabs/Lab5/Dockerfile.template", "repo_id": "AI-System", "token_count": 102 }
11
from urllib.request import urlretrieve import time import tarfile # TODO URL_path = r"" filename = r"Data.tar.gz" print("Please input the local folder path:") local = input() from_path = URL_path + "/" + filename to_path = local + "/" + filename print("Downloading...") try: urlretrieve(from_path, to_path) pr...
AI-System/Textbook/src/DataDownload.py/0
{ "file_path": "AI-System/Textbook/src/DataDownload.py", "repo_id": "AI-System", "token_count": 259 }
12
<!--Copyright © Microsoft Corporation. All rights reserved. 适用于[License](https://github.com/microsoft/AI-System/blob/main/LICENSE)版权许可--> # 人工智能系统概述 (System Perspective of System for AI) # 简介 目前应用最为广泛,工程实践最丰富,研究最为火热的人工智能(Artificial Intelligence)方向就是深度学习(Deep Learning)。深度学习算法目前在计算机视觉,自然语言处理,语音识别等典型场景取得了超越传统算法的效果与突...
AI-System/Textbook/第1章-人工智能系统概述/1-前言.md/0
{ "file_path": "AI-System/Textbook/第1章-人工智能系统概述/1-前言.md", "repo_id": "AI-System", "token_count": 2578 }
13
<!--Copyright © Microsoft Corporation. All rights reserved. 适用于[License](https://github.com/microsoft/AI-System/blob/main/LICENSE)版权许可--> ## 2.6 梯度下降优化算法 本小节主要围绕梯度下降优化算法,随机梯度下降 SGD,动量算法 Momentum,梯度加速算法 NAG,AdaGrad,AdaDelta,均方根反向传播 RMSProp,Adam展开介绍。 - [2.6 梯度下降优化算法](#26-梯度下降优化算法) - [随机梯度下降 SGD](#随机梯度下降-sgd) ...
AI-System/Textbook/第2章-神经网络基础/2.6-梯度下降优化算法.md/0
{ "file_path": "AI-System/Textbook/第2章-神经网络基础/2.6-梯度下降优化算法.md", "repo_id": "AI-System", "token_count": 6684 }
14
<!--Copyright © Microsoft Corporation. All rights reserved. 适用于[License](https://github.com/microsoft/AI-System/blob/main/LICENSE)版权许可--> # 8.1 推理系统简介 本章将围绕推理系统的应用场景,对比推理和训练场景的不同点,进而介绍推理系统的设计目标与约束,为后面章节内容的展开做好铺垫。 - [8.1 推理系统简介](#81-推理系统简介) - [8.1.1 对比推理与训练过程](#811-对比推理与训练过程) - [8.1.2 推理系统的优化目标与约束](#812-推理系统的优化...
AI-System/Textbook/第8章-深度学习推理系统/8.1-推理系统简介.md/0
{ "file_path": "AI-System/Textbook/第8章-深度学习推理系统/8.1-推理系统简介.md", "repo_id": "AI-System", "token_count": 15350 }
15
name: master channels: - conda-forge dependencies: - python=3.6 - pip - ipykernel - ipywidgets==7.3.0 - pandas>=0.23.3 - nb_conda_kernels==2.1.0 - scikit-learn>=0.19.1 - tensorflow-gpu==1.14.0 - tornado<6 - urllib3==1.23 - pip: - aiohttp>=3.3.2 - azure-cli==2.0.77 - azureml-sdk[note...
AI/.ci/environment.yml/0
{ "file_path": "AI/.ci/environment.yml", "repo_id": "AI", "token_count": 445 }
16
parameters: Agent: Hosted Ubuntu 1604 Demands: "python3" stageName: 'defaultStageName' jobDisplayName: 'defaultDisplayName' jobTimeoutInMinutes: 180 TridentWorkloadTypeShort: # DeployLocation: # TestPostfix: # "" | "-release" | "-preview" Deploy_Location_Short: # DefaultWorkingDirectory: # Templat...
AI/.ci/stage/deploy_notebooks_stage_v2.yml/0
{ "file_path": "AI/.ci/stage/deploy_notebooks_stage_v2.yml", "repo_id": "AI", "token_count": 1219 }
17
parameters: azureSubscription: '' azure_subscription: '' location: submodules/DeployMLModelPipelines azureresourcegroup: dcibhpdl workspacename: dcibhpwsdl azureregion: westus2 aksimagename: dcibhpaksdl aks_name: dcibhpaksdl aks_service_name: dcibhpaksdlapi conda: batchscoringdl_aml doCleanup: tru...
AI/.ci/steps/DLBatchDeployAMLJob.yml/0
{ "file_path": "AI/.ci/steps/DLBatchDeployAMLJob.yml", "repo_id": "AI", "token_count": 814 }
18
# AI Architecture # # A Github Service Connection must also be created with the name "AIArchitecturesAndPractices-GitHub" # https://docs.microsoft.com/en-us/azure/devops/pipelines/process/demands?view=azure-devops&tabs=yaml # # An Agent_Name Variable must be creating in the Azure DevOps UI. # https://docs.microsoft.com...
AI/.ci/steps/conda_pytest_pylint.yml/0
{ "file_path": "AI/.ci/steps/conda_pytest_pylint.yml", "repo_id": "AI", "token_count": 693 }
19
steps: - script: | docker stop $(docker ps -a -q) docker rm $(docker ps -a -q) docker system prune -a -f displayName: 'Docker Clean'
AI/.ci/steps/docker_clean.yml/0
{ "file_path": "AI/.ci/steps/docker_clean.yml", "repo_id": "AI", "token_count": 67 }
20
variables: TridentWorkloadTypeShort: aidlscore DeployLocation: eastus2 ProjectLocation: "." PythonPath: "." Template: DLTrainDeployAMLJob.yml
AI/.ci/vars/dlhyperparametertuning_vars.yml/0
{ "file_path": "AI/.ci/vars/dlhyperparametertuning_vars.yml", "repo_id": "AI", "token_count": 52 }
21
#!/bin/bash # # This script is for generate ann data for a model in training # # For the overall design of the ann driver, check run_train.sh # # This script continuously generate ann data using latest model from model_dir # For training, run this script after initial ann data is created from run_train.sh # Make sure p...
ANCE/commands/run_ann_data_gen.sh/0
{ "file_path": "ANCE/commands/run_ann_data_gen.sh", "repo_id": "ANCE", "token_count": 661 }
22
from .tokenization_seed_encoder import * from .configuration_seed_encoder import * from .modeling_seed_encoder import *
ANCE/model/SEED_Encoder/__init__.py/0
{ "file_path": "ANCE/model/SEED_Encoder/__init__.py", "repo_id": "ANCE", "token_count": 35 }
23
import sys sys.path += ['../'] import pandas as pd from sklearn.metrics import roc_curve, auc import gzip import copy import torch from torch import nn import torch.distributed as dist from tqdm import tqdm, trange import os from os import listdir from os.path import isfile, join import json import logging import rando...
ANCE/utils/util.py/0
{ "file_path": "ANCE/utils/util.py", "repo_id": "ANCE", "token_count": 5747 }
24
""" Code for self-training with weak supervision. Author: Giannis Karamanolakis (gkaraman@cs.columbia.edu) """ import numpy as np # Weak Source Classes # Here is the place to define heuristic rules (labeling functions) # Note: most rules are already provided in benchmarks as pre-processed files (for efficiency). cla...
ASTRA/astra/weaksource/ReGeXRules.py/0
{ "file_path": "ASTRA/astra/weaksource/ReGeXRules.py", "repo_id": "ASTRA", "token_count": 1248 }
25
# Contributor Covenant Code of Conduct ## Our Pledge We as members, contributors, and leaders pledge to make participation in our community a harassment-free experience for everyone, regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, level o...
AdaMix/CODE_OF_CONDUCT.md/0
{ "file_path": "AdaMix/CODE_OF_CONDUCT.md", "repo_id": "AdaMix", "token_count": 1107 }
26
{ "entries": [ { "1": { "category": "Airport", "dbpedialinks": [], "lexicalisations": [ { "comment": "good", "lang": "", "lex": "Abilene, Texas is served by...
AdaMix/NLG/data/webnlg_challenge_2017/test.json/0
{ "file_path": "AdaMix/NLG/data/webnlg_challenge_2017/test.json", "repo_id": "AdaMix", "token_count": 3935629 }
27
. ./venv/bin/activate sudo apt install default-jre -y seed=110 vv=lora_only while [[ $# -gt 0 ]] do key="$1" case $key in --seed) seed=$2 shift shift ;; --vv) vv=$2 shift shift ;; esac done python -m torch.distributed.launch --nproc_per_node=16 src/gpt2_beam.py \ --data ./...
AdaMix/NLG/run_eval_webnlg_lora_only.sh/0
{ "file_path": "AdaMix/NLG/run_eval_webnlg_lora_only.sh", "repo_id": "AdaMix", "token_count": 650 }
28
# ------------------------------------------------------------------------------------------ # Copyright (c). All rights reserved. # Licensed under the MIT License (MIT). See LICENSE in the repo root for license information. # -----------------------------------------------------------------------------------------...
AdaMix/NLG/src/gpt2_beam.py/0
{ "file_path": "AdaMix/NLG/src/gpt2_beam.py", "repo_id": "AdaMix", "token_count": 7930 }
29
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/glossary.rst/0
{ "file_path": "AdaMix/docs/source/glossary.rst", "repo_id": "AdaMix", "token_count": 4043 }
30
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/internal/modeling_utils.rst/0
{ "file_path": "AdaMix/docs/source/internal/modeling_utils.rst", "repo_id": "AdaMix", "token_count": 926 }
31
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/model_doc/albert.rst/0
{ "file_path": "AdaMix/docs/source/model_doc/albert.rst", "repo_id": "AdaMix", "token_count": 1444 }
32
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/model_doc/distilbert.rst/0
{ "file_path": "AdaMix/docs/source/model_doc/distilbert.rst", "repo_id": "AdaMix", "token_count": 1471 }
33
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/model_doc/marian.rst/0
{ "file_path": "AdaMix/docs/source/model_doc/marian.rst", "repo_id": "AdaMix", "token_count": 3273 }
34
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/model_doc/transformerxl.rst/0
{ "file_path": "AdaMix/docs/source/model_doc/transformerxl.rst", "repo_id": "AdaMix", "token_count": 1282 }
35
.. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by ...
AdaMix/docs/source/serialization.rst/0
{ "file_path": "AdaMix/docs/source/serialization.rst", "repo_id": "AdaMix", "token_count": 3385 }
36
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
AdaMix/examples/legacy/question-answering/run_squad_trainer.py/0
{ "file_path": "AdaMix/examples/legacy/question-answering/run_squad_trainer.py", "repo_id": "AdaMix", "token_count": 2553 }
37
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
AdaMix/examples/legacy/seq2seq/old_test_datasets.py/0
{ "file_path": "AdaMix/examples/legacy/seq2seq/old_test_datasets.py", "repo_id": "AdaMix", "token_count": 5148 }
38
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
AdaMix/examples/legacy/seq2seq/seq2seq_training_args.py/0
{ "file_path": "AdaMix/examples/legacy/seq2seq/seq2seq_training_args.py", "repo_id": "AdaMix", "token_count": 889 }
39
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
AdaMix/examples/legacy/seq2seq/xla_spawn.py/0
{ "file_path": "AdaMix/examples/legacy/seq2seq/xla_spawn.py", "repo_id": "AdaMix", "token_count": 916 }
40
# coding=utf-8 # Copyright 2020 The Google AI Language Team Authors, The HuggingFace Inc. team and Microsoft Corporation. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License....
AdaMix/examples/research_projects/bert-loses-patience/run_glue_with_pabee.py/0
{ "file_path": "AdaMix/examples/research_projects/bert-loses-patience/run_glue_with_pabee.py", "repo_id": "AdaMix", "token_count": 12948 }
41
#!/bin/bash export CUDA_VISIBLE_DEVICES=0 PATH_TO_DATA=/h/xinji/projects/GLUE MODEL_TYPE=bert # bert or roberta MODEL_SIZE=base # base or large DATASET=MRPC # SST-2, MRPC, RTE, QNLI, QQP, or MNLI MODEL_NAME=${MODEL_TYPE}-${MODEL_SIZE} if [ $MODEL_TYPE = 'bert' ] then MODEL_NAME=${MODEL_NAME}-uncased fi python...
AdaMix/examples/research_projects/deebert/eval_deebert.sh/0
{ "file_path": "AdaMix/examples/research_projects/deebert/eval_deebert.sh", "repo_id": "AdaMix", "token_count": 360 }
42
# coding=utf-8 # Copyright 2019-present, the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
AdaMix/examples/research_projects/distillation/scripts/extract_distilbert.py/0
{ "file_path": "AdaMix/examples/research_projects/distillation/scripts/extract_distilbert.py", "repo_id": "AdaMix", "token_count": 1828 }
43
""" coding=utf-8 Copyright 2018, Antonio Mendoza Hao Tan, Mohit Bansal Adapted From Facebook Inc, Detectron2 && Huggingface Co. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www...
AdaMix/examples/research_projects/lxmert/modeling_frcnn.py/0
{ "file_path": "AdaMix/examples/research_projects/lxmert/modeling_frcnn.py", "repo_id": "AdaMix", "token_count": 34761 }
44
# flake8: noqa from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
AdaMix/examples/research_projects/movement-pruning/emmental/__init__.py/0
{ "file_path": "AdaMix/examples/research_projects/movement-pruning/emmental/__init__.py", "repo_id": "AdaMix", "token_count": 107 }
45
#!/usr/bin/env bash export PYTHONPATH="../":"${PYTHONPATH}" export WANDB_PROJECT=dmar export MAX_LEN=128 python finetune.py \ --learning_rate=3e-4 \ --do_train \ --do_predict \ --fp16 \ --val_check_interval 0.25 \ --data_dir $ENRO_DIR \ --max_source_length $MAX_LEN --max_target_length $MAX_LEN --val_max_t...
AdaMix/examples/research_projects/seq2seq-distillation/distil_marian_no_teacher.sh/0
{ "file_path": "AdaMix/examples/research_projects/seq2seq-distillation/distil_marian_no_teacher.sh", "repo_id": "AdaMix", "token_count": 274 }
46
#!/usr/bin/env bash export PYTHONPATH="../":"${PYTHONPATH}" python finetune.py \ --learning_rate=3e-5 \ --fp16 \ --do_train \ --val_check_interval=0.25 \ --adam_eps 1e-06 \ --num_train_epochs 6 --src_lang en_XX --tgt_lang ro_RO \ --data_dir $ENRO_DIR \ --max_source_length $MAX_LEN --max...
AdaMix/examples/research_projects/seq2seq-distillation/train_mbart_cc25_enro.sh/0
{ "file_path": "AdaMix/examples/research_projects/seq2seq-distillation/train_mbart_cc25_enro.sh", "repo_id": "AdaMix", "token_count": 273 }
47
# coding=utf-8 # Copyright 2018 HuggingFace Inc.. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or a...
AdaMix/examples/test_examples.py/0
{ "file_path": "AdaMix/examples/test_examples.py", "repo_id": "AdaMix", "token_count": 5464 }
48
<jupyter_start><jupyter_text>Tokenization doesn't have to be slow ! IntroductionBefore going deep into any Machine Learning or Deep Learning Natural Language Processing models, every practitionershould find a way to map raw input strings to a representation understandable by a trainable model.One very simple approach w...
AdaMix/notebooks/01-training-tokenizers.ipynb/0
{ "file_path": "AdaMix/notebooks/01-training-tokenizers.ipynb", "repo_id": "AdaMix", "token_count": 2371 }
49
#!/usr/bin/env python # coding: utf-8 # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENS...
AdaMix/scripts/fsmt/fsmt-make-tiny-model.py/0
{ "file_path": "AdaMix/scripts/fsmt/fsmt-make-tiny-model.py", "repo_id": "AdaMix", "token_count": 698 }
50
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.a...
AdaMix/src/transformers/benchmark/benchmark_args.py/0
{ "file_path": "AdaMix/src/transformers/benchmark/benchmark_args.py", "repo_id": "AdaMix", "token_count": 1591 }
51
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
AdaMix/src/transformers/configuration_utils.py/0
{ "file_path": "AdaMix/src/transformers/configuration_utils.py", "repo_id": "AdaMix", "token_count": 13394 }
52
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
AdaMix/src/transformers/data/metrics/squad_metrics.py/0
{ "file_path": "AdaMix/src/transformers/data/metrics/squad_metrics.py", "repo_id": "AdaMix", "token_count": 13805 }
53
# Copyright 2020 The HuggingFace Team, the AllenNLP library authors. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # ...
AdaMix/src/transformers/file_utils.py/0
{ "file_path": "AdaMix/src/transformers/file_utils.py", "repo_id": "AdaMix", "token_count": 24853 }
54
# coding=utf-8 # Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
AdaMix/src/transformers/models/albert/tokenization_albert_fast.py/0
{ "file_path": "AdaMix/src/transformers/models/albert/tokenization_albert_fast.py", "repo_id": "AdaMix", "token_count": 5145 }
55
# coding=utf-8 # Copyright 2020, The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
AdaMix/src/transformers/models/bort/convert_bort_original_gluonnlp_checkpoint_to_pytorch.py/0
{ "file_path": "AdaMix/src/transformers/models/bort/convert_bort_original_gluonnlp_checkpoint_to_pytorch.py", "repo_id": "AdaMix", "token_count": 6174 }
56
# coding=utf-8 # Copyright 2018 Salesforce and The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless require...
AdaMix/src/transformers/models/ctrl/tokenization_ctrl.py/0
{ "file_path": "AdaMix/src/transformers/models/ctrl/tokenization_ctrl.py", "repo_id": "AdaMix", "token_count": 3919 }
57
# coding=utf-8 # Copyright 2020 Microsoft and the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required...
AdaMix/src/transformers/models/deberta_v2/tokenization_deberta_v2.py/0
{ "file_path": "AdaMix/src/transformers/models/deberta_v2/tokenization_deberta_v2.py", "repo_id": "AdaMix", "token_count": 8468 }
58
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors, The HuggingFace Inc. team, and the # Lxmert Authors. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License....
AdaMix/src/transformers/models/lxmert/modeling_tf_lxmert.py/0
{ "file_path": "AdaMix/src/transformers/models/lxmert/modeling_tf_lxmert.py", "repo_id": "AdaMix", "token_count": 29062 }
59
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
AdaMix/src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py/0
{ "file_path": "AdaMix/src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py", "repo_id": "AdaMix", "token_count": 1117 }
60