text stringlengths 5 22M | id stringlengths 12 177 | metadata dict | __index_level_0__ int64 0 1.37k |
|---|---|---|---|
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="view-comfy" format="2">
<advance width="1200"/>
<unicode hex="F0E6A"/>
<note>
view-comfy
</note>
<outline>
<contour>
<point x="20" y="1137" type="line"/>
<point x="20" y="893" type="line"/>
<point x="263" y="893" type="line"/>
<poin... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-comfy.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-comfy.glif",
"repo_id": "cascadia-code",
"token_count": 1270
} | 1,333 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="view-grid" format="2">
<advance width="1200"/>
<unicode hex="F0570"/>
<note>
view-grid
</note>
<outline>
<contour>
<point x="20" y="1290" type="line"/>
<point x="20" y="773" type="line"/>
<point x="537" y="773" type="line"/>
<point ... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-grid.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/view-grid.glif",
"repo_id": "cascadia-code",
"token_count": 472
} | 1,334 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="wall-sconce-flat-variant-outline" format="2">
<advance width="1200"/>
<unicode hex="F17CA"/>
<note>
wall-sconce-flat-variant-outline
</note>
<outline>
<contour>
<point x="1077" y="643" type="line"/>
<point x="123" y="643" type="line"/>
<p... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wall-sconce-flat-variant-outline.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wall-sconce-flat-variant-outline.glif",
"repo_id": "cascadia-code",
"token_count": 596
} | 1,335 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="wallet" format="2">
<advance width="1200"/>
<unicode hex="F0584"/>
<note>
wallet
</note>
<outline>
<contour>
<point x="570" y="343" type="line" smooth="yes"/>
<point x="518" y="343"/>
<point x="447" y="415"/>
<point x="447" y="467" ... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wallet.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/wallet.glif",
"repo_id": "cascadia-code",
"token_count": 917
} | 1,336 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="weather-hurricane" format="2">
<advance width="1200"/>
<unicode hex="F0898"/>
<note>
weather-hurricane
</note>
<outline>
<contour>
<point x="787" y="1093"/>
<point x="770" y="1150"/>
<point x="783" y="1177" type="qcurve" smooth="yes"/>
... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/weather-hurricane.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/weather-hurricane.glif",
"repo_id": "cascadia-code",
"token_count": 1035
} | 1,337 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="window-minimize" format="2">
<advance width="1200"/>
<unicode hex="F05B0"/>
<note>
window-minimize
</note>
<outline>
<contour>
<point x="1180" y="856" type="line"/>
<point x="20" y="856" type="line"/>
<point x="20" y="564" type="line"/>
... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/window-minimize.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/window-minimize.glif",
"repo_id": "cascadia-code",
"token_count": 185
} | 1,338 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="zodiac-scorpio" format="2">
<advance width="1200"/>
<unicode hex="F0A86"/>
<note>
zodiac-scorpio
</note>
<outline>
<contour>
<point x="1180" y="323" type="line"/>
<point x="947" y="557" type="line"/>
<point x="858" y="468" type="line"/>
... | cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/zodiac-scorpio.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/MaterialDesignIconsDesktop.ufo/glyphs/zodiac-scorpio.glif",
"repo_id": "cascadia-code",
"token_count": 1042
} | 1,339 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="POMODORO_TICKING" format="2">
<advance width="1200"/>
<unicode hex="E003"/>
<note>
POMODORO_TICKING
</note>
<outline>
<contour>
<point x="644" y="660" type="line"/>
<point x="644" y="297" type="line"/>
<point x="546" y="297" type="line"/>... | cascadia-code/sources/nerdfonts/full/processed/Pomicons.ufo/glyphs/P_O_M_O_D_O_R_O__T_I_C_K_I_N_G_.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/Pomicons.ufo/glyphs/P_O_M_O_D_O_R_O__T_I_C_K_I_N_G_.glif",
"repo_id": "cascadia-code",
"token_count": 1304
} | 1,340 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="uniE0B2" format="2">
<advance width="1200"/>
<unicode hex="E0B2"/>
<note>
uniE0B2
</note>
<outline>
<contour>
<point x="-122" y="942" type="line"/>
<point x="1322" y="-412" type="line"/>
<point x="1322" y="2294" type="line"/>
</contou... | cascadia-code/sources/nerdfonts/full/processed/PowerlineExtraSymbols.ufo/glyphs/uniE_0B_2.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/PowerlineExtraSymbols.ufo/glyphs/uniE_0B_2.glif",
"repo_id": "cascadia-code",
"token_count": 167
} | 1,341 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="activate-breakpoints" format="2">
<advance width="1200"/>
<unicode hex="EA97"/>
<note>
activate-breakpoints
</note>
<outline>
<contour>
<point x="1180" y="993"/>
<point x="1126" y="1130"/>
<point x="1020" y="1237"/>
<point x="883" y... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/activate-breakpoints.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/activate-breakpoints.glif",
"repo_id": "cascadia-code",
"token_count": 873
} | 1,342 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="check-all" format="2">
<advance width="1200"/>
<unicode hex="EBB1"/>
<note>
check-all
</note>
<outline>
<contour>
<point x="1129" y="1074" type="line"/>
<point x="578" y="425" type="line"/>
<point x="372" y="714" type="line"/>
<poin... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/check-all.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/check-all.glif",
"repo_id": "cascadia-code",
"token_count": 552
} | 1,343 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="debug-breakpoint-log-unverified" format="2">
<advance width="1200"/>
<unicode hex="EAAA"/>
<note>
debug-breakpoint-log-unverified
</note>
<outline>
<contour>
<point x="600" y="127" type="line"/>
<point x="1180" y="707" type="line"/>
<poin... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-breakpoint-log-unverified.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-breakpoint-log-unverified.glif",
"repo_id": "cascadia-code",
"token_count": 287
} | 1,344 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="debug-stackframe" format="2">
<advance width="1200"/>
<unicode hex="EB8B"/>
<note>
debug-stackframe
</note>
<outline>
<contour>
<point x="1180" y="627" type="line"/>
<point x="1180" y="793" type="line"/>
<point x="750" y="1271" type="line... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-stackframe.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/debug-stackframe.glif",
"repo_id": "cascadia-code",
"token_count": 423
} | 1,345 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="diff-renamed" format="2">
<advance width="1200"/>
<unicode hex="EAE0"/>
<note>
diff-renamed
</note>
<outline>
<contour>
<point x="20" y="1250" type="line"/>
<point x="20" y="170" type="line"/>
<point x="60" y="130" type="line"/>
<po... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/diff-renamed.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/diff-renamed.glif",
"repo_id": "cascadia-code",
"token_count": 522
} | 1,346 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="file-code" format="2">
<advance width="1200"/>
<unicode hex="EAE9"/>
<unicode hex="EAFB"/>
<note>
file-code
</note>
<outline>
<contour>
<point x="814" y="1385" type="line"/>
<point x="66" y="1385" type="line"/>
<point x="20" y="1339" ty... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/file-code.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/file-code.glif",
"repo_id": "cascadia-code",
"token_count": 875
} | 1,347 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="folder-library" format="2">
<advance width="1200"/>
<unicode hex="EBDF"/>
<note>
folder-library
</note>
<outline>
<contour>
<point x="507" y="1195" type="line"/>
<point x="476" y="1208" type="line"/>
<point x="60" y="1208" type="line"/>
... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/folder-library.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/folder-library.glif",
"repo_id": "cascadia-code",
"token_count": 1005
} | 1,348 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="mail-read" format="2">
<advance width="1200"/>
<unicode hex="EB1B"/>
<note>
mail-read
</note>
<outline>
<contour>
<point x="1180" y="894" type="line"/>
<point x="622" y="1226" type="line"/>
<point x="578" y="1226" type="line"/>
<poi... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/mail-read.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/mail-read.glif",
"repo_id": "cascadia-code",
"token_count": 565
} | 1,349 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="redo" format="2">
<advance width="1200"/>
<unicode hex="EBB0"/>
<note>
redo
</note>
<outline>
<contour>
<point x="1013" y="1420" type="line"/>
<point x="1013" y="1193" type="line"/>
<point x="938" y="1269" type="line" smooth="yes"/>
... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/redo.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/redo.glif",
"repo_id": "cascadia-code",
"token_count": 613
} | 1,350 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="report" format="2">
<advance width="1200"/>
<unicode hex="EB42"/>
<note>
report
</note>
<outline>
<contour>
<point x="20" y="1246" type="line"/>
<point x="20" y="413" type="line"/>
<point x="60" y="374" type="line"/>
<point x="268" ... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/report.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/report.glif",
"repo_id": "cascadia-code",
"token_count": 735
} | 1,351 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="star-empty" format="2">
<advance width="1200"/>
<unicode hex="EA6A"/>
<note>
star-empty
</note>
<outline>
<contour>
<point x="1180" y="856" type="line"/>
<point x="733" y="856" type="line"/>
<point x="600" y="1290" type="line"/>
<po... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/star-empty.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/star-empty.glif",
"repo_id": "cascadia-code",
"token_count": 740
} | 1,352 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="symbol-keyword" format="2">
<advance width="1200"/>
<unicode hex="EB62"/>
<note>
symbol-keyword
</note>
<outline>
<contour>
<point x="1180" y="1124" type="line"/>
<point x="768" y="1124" type="line"/>
<point x="768" y="1040" type="line"/>... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/symbol-keyword.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/symbol-keyword.glif",
"repo_id": "cascadia-code",
"token_count": 869
} | 1,353 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="verified" format="2">
<advance width="1200"/>
<unicode hex="EB77"/>
<note>
verified
</note>
<outline>
<contour>
<point x="427" y="276" type="line"/>
<point x="570" y="128" type="line"/>
<point x="630" y="128" type="line"/>
<point x=... | cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/verified.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/codicon.ufo/glyphs/verified.glif",
"repo_id": "cascadia-code",
"token_count": 1261
} | 1,354 |
<?xml version='1.0' encoding='UTF-8'?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>ascender</key>
<integer>410</integer>
<key>capHeight</key>
<integer>0</integer>
<key>copyright</key>
<string>Copyright... | cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/fontinfo.plist/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/fontinfo.plist",
"repo_id": "cascadia-code",
"token_count": 1282
} | 1,355 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="KDE Neon" format="2">
<advance width="1200"/>
<unicode hex="F331"/>
<note>
KDE Neon
</note>
<outline>
<contour>
<point x="568" y="1267" type="qcurve"/>
<point x="373" y="1256"/>
<point x="228" y="1127" type="qcurve"/>
<point x="224"... | cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/K_D_E_ N_eon.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/K_D_E_ N_eon.glif",
"repo_id": "cascadia-code",
"token_count": 3095
} | 1,356 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="Puppy Linux" format="2">
<advance width="1200"/>
<unicode hex="F341"/>
<note>
Puppy Linux
</note>
<outline>
<contour>
<point x="753" y="1287"/>
<point x="716" y="1277" type="qcurve"/>
<point x="518" y="1317"/>
<point x="341" y="1227... | cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/P_uppy L_inux.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/P_uppy L_inux.glif",
"repo_id": "cascadia-code",
"token_count": 3748
} | 1,357 |
<?xml version='1.0' encoding='UTF-8'?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>AOSC OS</key>
<string>A_O_S_C_ O_S_.glif</string>
<key>Alma Linux</key>
<string>A_lma L_inux.glif</string>
<key>Alpine</ke... | cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/contents.plist/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/font-logos.ufo/glyphs/contents.plist",
"repo_id": "cascadia-code",
"token_count": 3778
} | 1,358 |
<?xml version='1.0' encoding='UTF-8'?>
<glyph name="i_seti_csv" format="2">
<advance width="1200"/>
<unicode hex="E64A"/>
<note>
i_seti_csv
</note>
<outline>
<contour>
<point x="1180" y="1284" type="line"/>
<point x="20" y="1284" type="line"/>
<point x="20" y="136" type="line"/>
<poi... | cascadia-code/sources/nerdfonts/full/processed/original-source.ufo/glyphs/i_seti_csv.glif/0 | {
"file_path": "cascadia-code/sources/nerdfonts/full/processed/original-source.ufo/glyphs/i_seti_csv.glif",
"repo_id": "cascadia-code",
"token_count": 1042
} | 1,359 |
{
"name": "Causica Devcontainer",
"build": {
"context": "..",
"dockerfile": "../Dockerfile",
"target": "dev"
},
"runArgs": [
// give headspace for pytorch dataloaders passing tensors across processes
"--shm-size=1gb"
],
"containerEnv": {
"DOCKER_BUILDKIT": "1"
},
"mounts": [
"source=command_histor... | causica/.devcontainer/devcontainer.json/0 | {
"file_path": "causica/.devcontainer/devcontainer.json",
"repo_id": "causica",
"token_count": 806
} | 1,360 |
# Microsoft Open Source Code of Conduct
This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/).
Resources:
- [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/)
- [Microsoft Code of Conduct FAQ](https://opensource.mic... | causica/CODE_OF_CONDUCT.md/0 | {
"file_path": "causica/CODE_OF_CONDUCT.md",
"repo_id": "causica",
"token_count": 115
} | 1,361 |
import torch
from tensordict import TensorDict
def check_temporal_tensordict_shapes(num_timepoints: int, tds: TensorDict) -> bool:
"""Check that the shapes within the TensorDict are consistent with the number of time points."""
return all(val.shape[len(tds.batch_size)] == num_timepoints for val in tds.values(... | causica/research_experiments/scotch/src/scotch/scotch_utils/scotch_utils.py/0 | {
"file_path": "causica/research_experiments/scotch/src/scotch/scotch_utils/scotch_utils.py",
"repo_id": "causica",
"token_count": 192
} | 1,362 |
import math
import numpy as np
import torch
import torch.distributions as td
from causica.distributions.adjacency.adjacency_distributions import AdjacencyDistribution
from causica.triangular_transformations import fill_triangular
class ErdosRenyiDAGDistribution(AdjacencyDistribution):
"""
An adjacency distr... | causica/src/causica/distributions/adjacency/directed_acyclic.py/0 | {
"file_path": "causica/src/causica/distributions/adjacency/directed_acyclic.py",
"repo_id": "causica",
"token_count": 1716
} | 1,363 |
from typing import Optional
import torch
import torch.distributions as td
from torch import nn
from causica.distributions.noise.noise import IndependentNoise, Noise, NoiseModule
class UnivariateCauchyNoise(td.Cauchy, Noise[torch.Tensor]):
def sample_to_noise(self, samples: torch.Tensor) -> torch.Tensor:
... | causica/src/causica/distributions/noise/univariate_cauchy.py/0 | {
"file_path": "causica/src/causica/distributions/noise/univariate_cauchy.py",
"repo_id": "causica",
"token_count": 878
} | 1,364 |
import pytest
import torch
import torch.distributions as td
from causica.data_generation.samplers.functional_relationships_sampler import LinearRelationshipsSampler
from causica.data_generation.samplers.noise_dist_sampler import (
BernoulliNoiseModuleSampler,
CategoricalNoiseModuleSampler,
JointNoiseModule... | causica/test/datasets/test_synthetic_dataset.py/0 | {
"file_path": "causica/test/datasets/test_synthetic_dataset.py",
"repo_id": "causica",
"token_count": 1393
} | 1,365 |
import pytest
import torch
from causica.distributions.noise import UnivariateLaplaceNoise
@pytest.mark.parametrize(("batch", "dimension"), [(1, 10), (2, 5)])
def test_init(batch, dimension):
mean = torch.randn((batch, dimension))
scale = torch.ones((batch, dimension))
noise_model = UnivariateLaplaceNoise... | causica/test/distributions/noise/test_univariate_laplace.py/0 | {
"file_path": "causica/test/distributions/noise/test_univariate_laplace.py",
"repo_id": "causica",
"token_count": 843
} | 1,366 |
import os
from pathlib import Path
import mlflow
import pytest
import yaml
from jsonargparse import Namespace
from mlflow import ActiveRun
from pytorch_lightning import LightningModule, Trainer
from pytorch_lightning.cli import LightningArgumentParser
from pytorch_lightning.trainer.states import TrainerFn
from causic... | causica/test/lightning/test_callbacks.py/0 | {
"file_path": "causica/test/lightning/test_callbacks.py",
"repo_id": "causica",
"token_count": 503
} | 1,367 |
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
import sys
if sys.version_info >= (3, 8):
from importlib import metadata
else:
import importlib_metadata as metadata
__version__ = metadata.version("cliffordlayers")
| cliffordlayers/cliffordlayers/__init__.py/0 | {
"file_path": "cliffordlayers/cliffordlayers/__init__.py",
"repo_id": "cliffordlayers",
"token_count": 74
} | 1,368 |
import torch
import torch.nn.functional as F
from torch.nn.modules.utils import _single, _pair, _triple
def get_clifford_g3_kernel(weights):
assert len(weights) == 6
scale = weights[4]
square_1 = weights[0] * weights[0]
square_2 = weights[1] * weights[1]
square_3 = weights[2] * weights[2]
sq... | cliffordlayers/cliffordlayers/nn/functional/cliffordg3conv.py/0 | {
"file_path": "cliffordlayers/cliffordlayers/nn/functional/cliffordg3conv.py",
"repo_id": "cliffordlayers",
"token_count": 1316
} | 1,369 |
[project]
name = "cliffordlayers"
dynamic = ["version"]
description = "A PyTorch library for Clifford layers"
readme = "README.md"
requires-python = ">=3.8"
license = {text = "MIT"}
authors = [{name = "Jayesh K. Gupta", email = "mail@rejuvyesh.com"}, {name="Johannes Brandstetter", email="johannesb@microsoft.com"}, {nam... | cliffordlayers/pyproject.toml/0 | {
"file_path": "cliffordlayers/pyproject.toml",
"repo_id": "cliffordlayers",
"token_count": 420
} | 1,370 |
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
| cookiecutter-spacy-fastapi/{{cookiecutter.project_slug}}/app/tests/__init__.py/0 | {
"file_path": "cookiecutter-spacy-fastapi/{{cookiecutter.project_slug}}/app/tests/__init__.py",
"repo_id": "cookiecutter-spacy-fastapi",
"token_count": 21
} | 1,371 |
build/
node_modules/
dist/
coverage/
# Client
client/build/
client/node_modules/
client/dist/
client/coverage/
| 0xDeCA10B/demo/.eslintignore/0 | {
"file_path": "0xDeCA10B/demo/.eslintignore",
"repo_id": "0xDeCA10B",
"token_count": 42
} | 0 |
{
"name": "decai-demo-client",
"version": "1.4.5",
"license": "MIT",
"private": true,
"proxy": "http://localhost:5387/",
"dependencies": {
"@drizzle-utils/get-web3": "^0.2.2-alpha.0",
"@material-ui/core": "^4.7.1",
"@material-ui/icons": "^4.5.1",
"@tensorflow-models/mobilenet": "^2.0.4",
... | 0xDeCA10B/demo/client/package.json/0 | {
"file_path": "0xDeCA10B/demo/client/package.json",
"repo_id": "0xDeCA10B",
"token_count": 933
} | 1 |
pragma solidity ^0.6;
pragma experimental ABIEncoderV2;
import "../../../lib/SafeMath.sol";
import "../../../lib/SignedSafeMath.sol";
import {Ownable} from "../ownership/Ownable.sol";
/**
* A classifier that can take a data sample as input and return a predict classification/label for the data.
*/
contract Classif... | 0xDeCA10B/demo/client/src/contracts/classification/Classifier.sol/0 | {
"file_path": "0xDeCA10B/demo/client/src/contracts/classification/Classifier.sol",
"repo_id": "0xDeCA10B",
"token_count": 1033
} | 2 |
import * as _getWeb3 from '@drizzle-utils/get-web3'
import Web3 from 'web3' // Only required for custom/fallback provider option.
export async function getWeb3() {
if (typeof window !== "undefined" && window.ethereum) {
// Get rid of a warning about network refreshing.
window.ethereum.autoRefreshOnNetworkChange =... | 0xDeCA10B/demo/client/src/getWeb3.js/0 | {
"file_path": "0xDeCA10B/demo/client/src/getWeb3.js",
"repo_id": "0xDeCA10B",
"token_count": 350
} | 3 |
// Basically the same as tensor-utils.js but made for back-end.
// There are fancier ways to share code for front-end and back-end
// but they all seemed too complicated and unreliable in all cases.
const tf = require('@tensorflow/tfjs-node')
exports.normalize1d = function (x) {
return tf.tidy(_ => {
if (!(x insta... | 0xDeCA10B/demo/client/src/ml-models/tensor-utils-node.js/0 | {
"file_path": "0xDeCA10B/demo/client/src/ml-models/tensor-utils-node.js",
"repo_id": "0xDeCA10B",
"token_count": 311
} | 4 |
const util = require('util')
exports.setTimeoutPromise = util.promisify(setTimeout)
| 0xDeCA10B/demo/client/src/utils.js/0 | {
"file_path": "0xDeCA10B/demo/client/src/utils.js",
"repo_id": "0xDeCA10B",
"token_count": 27
} | 5 |
{
"compilerOptions": {
"target": "es2020",
"lib": [
"dom",
"dom.iterable",
"esnext"
],
"allowJs": true,
"skipLibCheck": true,
"esModuleInterop": true,
"allowSyntheticDefaultImports": true,
"strict": true,
"forceConsistentCasingInFileNames": true,
"module": "es... | 0xDeCA10B/demo/client/tsconfig.json/0 | {
"file_path": "0xDeCA10B/demo/client/tsconfig.json",
"repo_id": "0xDeCA10B",
"token_count": 231
} | 6 |
from collections import defaultdict
from dataclasses import dataclass, field
from typing import Dict
import numpy as np
from injector import inject, singleton
from decai.simulation.contract.objects import Address, RejectException, SmartContract, TimeMock
@dataclass
class StoredData:
# Storing the data is not ne... | 0xDeCA10B/simulation/decai/simulation/contract/data/data_handler.py/0 | {
"file_path": "0xDeCA10B/simulation/decai/simulation/contract/data/data_handler.py",
"repo_id": "0xDeCA10B",
"token_count": 1965
} | 7 |
import unittest
from decai.simulation.data.featuremapping.hashing.murmurhash3 import MurmurHash3
class TestMurmurHash3(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.h = MurmurHash3()
def test_classifications(self):
h = self.h.hash("hey")
assert type(h) == int
... | 0xDeCA10B/simulation/decai/simulation/data/featuremapping/hashing/tests/test_murmurhash3.py/0 | {
"file_path": "0xDeCA10B/simulation/decai/simulation/data/featuremapping/hashing/tests/test_murmurhash3.py",
"repo_id": "0xDeCA10B",
"token_count": 168
} | 8 |
import logging
from dataclasses import dataclass, field
from logging import Logger
from injector import Module, provider, singleton
@dataclass
class LoggingModule(Module):
_log_level: int = field(default=logging.INFO)
@provider
@singleton
def provide_logger(self) -> Logger:
result = logging.... | 0xDeCA10B/simulation/decai/simulation/logging_module.py/0 | {
"file_path": "0xDeCA10B/simulation/decai/simulation/logging_module.py",
"repo_id": "0xDeCA10B",
"token_count": 246
} | 9 |
# Lab 3 - CUDA实现和优化
## 实验目的
1. 理解DNN框架中的张量运算在GPU加速器上加速原理。
2. 通过CUDA实现和优化一个定制化张量运算。
## 实验环境
* PyTorch==1.5.0
* CUDA 10.0
## 实验原理
1. 矩阵运算与计算机体系结构
2. GPU加速器的加速原理
## 实验内容
### 实验流程图

### 具体步骤
1. 理解PyTorch中Linear张量运算的计算过程,推导计算公式
2. 了解GPU端加速的原理,CUDA内核编程和实现一个kernel的原理
3. 实... | AI-System/Labs/BasicLabs/Lab3/README.md/0 | {
"file_path": "AI-System/Labs/BasicLabs/Lab3/README.md",
"repo_id": "AI-System",
"token_count": 1565
} | 10 |
# What image do you want to start building on?
# Make a folder in your image where your app's source code can live
# Tell your container where your app's source code will live
# What source code do you what to copy, and where to put it?
# Does your app have any dependencies that should be installed?
# What po... | AI-System/Labs/BasicLabs/Lab5/Dockerfile.template/0 | {
"file_path": "AI-System/Labs/BasicLabs/Lab5/Dockerfile.template",
"repo_id": "AI-System",
"token_count": 102
} | 11 |
from urllib.request import urlretrieve
import time
import tarfile
# TODO
URL_path = r""
filename = r"Data.tar.gz"
print("Please input the local folder path:")
local = input()
from_path = URL_path + "/" + filename
to_path = local + "/" + filename
print("Downloading...")
try:
urlretrieve(from_path, to_path)
pr... | AI-System/Textbook/src/DataDownload.py/0 | {
"file_path": "AI-System/Textbook/src/DataDownload.py",
"repo_id": "AI-System",
"token_count": 259
} | 12 |
<!--Copyright © Microsoft Corporation. All rights reserved.
适用于[License](https://github.com/microsoft/AI-System/blob/main/LICENSE)版权许可-->
# 人工智能系统概述 (System Perspective of System for AI)
# 简介
目前应用最为广泛,工程实践最丰富,研究最为火热的人工智能(Artificial Intelligence)方向就是深度学习(Deep Learning)。深度学习算法目前在计算机视觉,自然语言处理,语音识别等典型场景取得了超越传统算法的效果与突... | AI-System/Textbook/第1章-人工智能系统概述/1-前言.md/0 | {
"file_path": "AI-System/Textbook/第1章-人工智能系统概述/1-前言.md",
"repo_id": "AI-System",
"token_count": 2578
} | 13 |
<!--Copyright © Microsoft Corporation. All rights reserved.
适用于[License](https://github.com/microsoft/AI-System/blob/main/LICENSE)版权许可-->
## 2.6 梯度下降优化算法
本小节主要围绕梯度下降优化算法,随机梯度下降 SGD,动量算法 Momentum,梯度加速算法 NAG,AdaGrad,AdaDelta,均方根反向传播 RMSProp,Adam展开介绍。
- [2.6 梯度下降优化算法](#26-梯度下降优化算法)
- [随机梯度下降 SGD](#随机梯度下降-sgd)
... | AI-System/Textbook/第2章-神经网络基础/2.6-梯度下降优化算法.md/0 | {
"file_path": "AI-System/Textbook/第2章-神经网络基础/2.6-梯度下降优化算法.md",
"repo_id": "AI-System",
"token_count": 6684
} | 14 |
<!--Copyright © Microsoft Corporation. All rights reserved.
适用于[License](https://github.com/microsoft/AI-System/blob/main/LICENSE)版权许可-->
# 8.1 推理系统简介
本章将围绕推理系统的应用场景,对比推理和训练场景的不同点,进而介绍推理系统的设计目标与约束,为后面章节内容的展开做好铺垫。
- [8.1 推理系统简介](#81-推理系统简介)
- [8.1.1 对比推理与训练过程](#811-对比推理与训练过程)
- [8.1.2 推理系统的优化目标与约束](#812-推理系统的优化... | AI-System/Textbook/第8章-深度学习推理系统/8.1-推理系统简介.md/0 | {
"file_path": "AI-System/Textbook/第8章-深度学习推理系统/8.1-推理系统简介.md",
"repo_id": "AI-System",
"token_count": 15350
} | 15 |
name: master
channels:
- conda-forge
dependencies:
- python=3.6
- pip
- ipykernel
- ipywidgets==7.3.0
- pandas>=0.23.3
- nb_conda_kernels==2.1.0
- scikit-learn>=0.19.1
- tensorflow-gpu==1.14.0
- tornado<6
- urllib3==1.23
- pip:
- aiohttp>=3.3.2
- azure-cli==2.0.77
- azureml-sdk[note... | AI/.ci/environment.yml/0 | {
"file_path": "AI/.ci/environment.yml",
"repo_id": "AI",
"token_count": 445
} | 16 |
parameters:
Agent: Hosted Ubuntu 1604
Demands: "python3"
stageName: 'defaultStageName'
jobDisplayName: 'defaultDisplayName'
jobTimeoutInMinutes: 180
TridentWorkloadTypeShort: #
DeployLocation: #
TestPostfix: # "" | "-release" | "-preview"
Deploy_Location_Short: #
DefaultWorkingDirectory: #
Templat... | AI/.ci/stage/deploy_notebooks_stage_v2.yml/0 | {
"file_path": "AI/.ci/stage/deploy_notebooks_stage_v2.yml",
"repo_id": "AI",
"token_count": 1219
} | 17 |
parameters:
azureSubscription: ''
azure_subscription: ''
location: submodules/DeployMLModelPipelines
azureresourcegroup: dcibhpdl
workspacename: dcibhpwsdl
azureregion: westus2
aksimagename: dcibhpaksdl
aks_name: dcibhpaksdl
aks_service_name: dcibhpaksdlapi
conda: batchscoringdl_aml
doCleanup: tru... | AI/.ci/steps/DLBatchDeployAMLJob.yml/0 | {
"file_path": "AI/.ci/steps/DLBatchDeployAMLJob.yml",
"repo_id": "AI",
"token_count": 814
} | 18 |
# AI Architecture
#
# A Github Service Connection must also be created with the name "AIArchitecturesAndPractices-GitHub"
# https://docs.microsoft.com/en-us/azure/devops/pipelines/process/demands?view=azure-devops&tabs=yaml
#
# An Agent_Name Variable must be creating in the Azure DevOps UI.
# https://docs.microsoft.com... | AI/.ci/steps/conda_pytest_pylint.yml/0 | {
"file_path": "AI/.ci/steps/conda_pytest_pylint.yml",
"repo_id": "AI",
"token_count": 693
} | 19 |
steps:
- script: |
docker stop $(docker ps -a -q)
docker rm $(docker ps -a -q)
docker system prune -a -f
displayName: 'Docker Clean'
| AI/.ci/steps/docker_clean.yml/0 | {
"file_path": "AI/.ci/steps/docker_clean.yml",
"repo_id": "AI",
"token_count": 67
} | 20 |
variables:
TridentWorkloadTypeShort: aidlscore
DeployLocation: eastus2
ProjectLocation: "."
PythonPath: "."
Template: DLTrainDeployAMLJob.yml
| AI/.ci/vars/dlhyperparametertuning_vars.yml/0 | {
"file_path": "AI/.ci/vars/dlhyperparametertuning_vars.yml",
"repo_id": "AI",
"token_count": 52
} | 21 |
#!/bin/bash
#
# This script is for generate ann data for a model in training
#
# For the overall design of the ann driver, check run_train.sh
#
# This script continuously generate ann data using latest model from model_dir
# For training, run this script after initial ann data is created from run_train.sh
# Make sure p... | ANCE/commands/run_ann_data_gen.sh/0 | {
"file_path": "ANCE/commands/run_ann_data_gen.sh",
"repo_id": "ANCE",
"token_count": 661
} | 22 |
from .tokenization_seed_encoder import *
from .configuration_seed_encoder import *
from .modeling_seed_encoder import * | ANCE/model/SEED_Encoder/__init__.py/0 | {
"file_path": "ANCE/model/SEED_Encoder/__init__.py",
"repo_id": "ANCE",
"token_count": 35
} | 23 |
import sys
sys.path += ['../']
import pandas as pd
from sklearn.metrics import roc_curve, auc
import gzip
import copy
import torch
from torch import nn
import torch.distributed as dist
from tqdm import tqdm, trange
import os
from os import listdir
from os.path import isfile, join
import json
import logging
import rando... | ANCE/utils/util.py/0 | {
"file_path": "ANCE/utils/util.py",
"repo_id": "ANCE",
"token_count": 5747
} | 24 |
"""
Code for self-training with weak supervision.
Author: Giannis Karamanolakis (gkaraman@cs.columbia.edu)
"""
import numpy as np
# Weak Source Classes
# Here is the place to define heuristic rules (labeling functions)
# Note: most rules are already provided in benchmarks as pre-processed files (for efficiency).
cla... | ASTRA/astra/weaksource/ReGeXRules.py/0 | {
"file_path": "ASTRA/astra/weaksource/ReGeXRules.py",
"repo_id": "ASTRA",
"token_count": 1248
} | 25 |
# Contributor Covenant Code of Conduct
## Our Pledge
We as members, contributors, and leaders pledge to make participation in our
community a harassment-free experience for everyone, regardless of age, body
size, visible or invisible disability, ethnicity, sex characteristics, gender
identity and expression, level o... | AdaMix/CODE_OF_CONDUCT.md/0 | {
"file_path": "AdaMix/CODE_OF_CONDUCT.md",
"repo_id": "AdaMix",
"token_count": 1107
} | 26 |
{
"entries": [
{
"1": {
"category": "Airport",
"dbpedialinks": [],
"lexicalisations": [
{
"comment": "good",
"lang": "",
"lex": "Abilene, Texas is served by... | AdaMix/NLG/data/webnlg_challenge_2017/test.json/0 | {
"file_path": "AdaMix/NLG/data/webnlg_challenge_2017/test.json",
"repo_id": "AdaMix",
"token_count": 3935629
} | 27 |
. ./venv/bin/activate
sudo apt install default-jre -y
seed=110
vv=lora_only
while [[ $# -gt 0 ]]
do
key="$1"
case $key in
--seed)
seed=$2
shift
shift
;;
--vv)
vv=$2
shift
shift
;;
esac
done
python -m torch.distributed.launch --nproc_per_node=16 src/gpt2_beam.py \
--data ./... | AdaMix/NLG/run_eval_webnlg_lora_only.sh/0 | {
"file_path": "AdaMix/NLG/run_eval_webnlg_lora_only.sh",
"repo_id": "AdaMix",
"token_count": 650
} | 28 |
# ------------------------------------------------------------------------------------------
# Copyright (c). All rights reserved.
# Licensed under the MIT License (MIT). See LICENSE in the repo root for license information.
# -----------------------------------------------------------------------------------------... | AdaMix/NLG/src/gpt2_beam.py/0 | {
"file_path": "AdaMix/NLG/src/gpt2_beam.py",
"repo_id": "AdaMix",
"token_count": 7930
} | 29 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/glossary.rst/0 | {
"file_path": "AdaMix/docs/source/glossary.rst",
"repo_id": "AdaMix",
"token_count": 4043
} | 30 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/internal/modeling_utils.rst/0 | {
"file_path": "AdaMix/docs/source/internal/modeling_utils.rst",
"repo_id": "AdaMix",
"token_count": 926
} | 31 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/model_doc/albert.rst/0 | {
"file_path": "AdaMix/docs/source/model_doc/albert.rst",
"repo_id": "AdaMix",
"token_count": 1444
} | 32 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/model_doc/distilbert.rst/0 | {
"file_path": "AdaMix/docs/source/model_doc/distilbert.rst",
"repo_id": "AdaMix",
"token_count": 1471
} | 33 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/model_doc/marian.rst/0 | {
"file_path": "AdaMix/docs/source/model_doc/marian.rst",
"repo_id": "AdaMix",
"token_count": 3273
} | 34 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/model_doc/transformerxl.rst/0 | {
"file_path": "AdaMix/docs/source/model_doc/transformerxl.rst",
"repo_id": "AdaMix",
"token_count": 1282
} | 35 |
..
Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by ... | AdaMix/docs/source/serialization.rst/0 | {
"file_path": "AdaMix/docs/source/serialization.rst",
"repo_id": "AdaMix",
"token_count": 3385
} | 36 |
# coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | AdaMix/examples/legacy/question-answering/run_squad_trainer.py/0 | {
"file_path": "AdaMix/examples/legacy/question-answering/run_squad_trainer.py",
"repo_id": "AdaMix",
"token_count": 2553
} | 37 |
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | AdaMix/examples/legacy/seq2seq/old_test_datasets.py/0 | {
"file_path": "AdaMix/examples/legacy/seq2seq/old_test_datasets.py",
"repo_id": "AdaMix",
"token_count": 5148
} | 38 |
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | AdaMix/examples/legacy/seq2seq/seq2seq_training_args.py/0 | {
"file_path": "AdaMix/examples/legacy/seq2seq/seq2seq_training_args.py",
"repo_id": "AdaMix",
"token_count": 889
} | 39 |
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | AdaMix/examples/legacy/seq2seq/xla_spawn.py/0 | {
"file_path": "AdaMix/examples/legacy/seq2seq/xla_spawn.py",
"repo_id": "AdaMix",
"token_count": 916
} | 40 |
# coding=utf-8
# Copyright 2020 The Google AI Language Team Authors, The HuggingFace Inc. team and Microsoft Corporation.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.... | AdaMix/examples/research_projects/bert-loses-patience/run_glue_with_pabee.py/0 | {
"file_path": "AdaMix/examples/research_projects/bert-loses-patience/run_glue_with_pabee.py",
"repo_id": "AdaMix",
"token_count": 12948
} | 41 |
#!/bin/bash
export CUDA_VISIBLE_DEVICES=0
PATH_TO_DATA=/h/xinji/projects/GLUE
MODEL_TYPE=bert # bert or roberta
MODEL_SIZE=base # base or large
DATASET=MRPC # SST-2, MRPC, RTE, QNLI, QQP, or MNLI
MODEL_NAME=${MODEL_TYPE}-${MODEL_SIZE}
if [ $MODEL_TYPE = 'bert' ]
then
MODEL_NAME=${MODEL_NAME}-uncased
fi
python... | AdaMix/examples/research_projects/deebert/eval_deebert.sh/0 | {
"file_path": "AdaMix/examples/research_projects/deebert/eval_deebert.sh",
"repo_id": "AdaMix",
"token_count": 360
} | 42 |
# coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by a... | AdaMix/examples/research_projects/distillation/scripts/extract_distilbert.py/0 | {
"file_path": "AdaMix/examples/research_projects/distillation/scripts/extract_distilbert.py",
"repo_id": "AdaMix",
"token_count": 1828
} | 43 |
"""
coding=utf-8
Copyright 2018, Antonio Mendoza Hao Tan, Mohit Bansal
Adapted From Facebook Inc, Detectron2 && Huggingface Co.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www... | AdaMix/examples/research_projects/lxmert/modeling_frcnn.py/0 | {
"file_path": "AdaMix/examples/research_projects/lxmert/modeling_frcnn.py",
"repo_id": "AdaMix",
"token_count": 34761
} | 44 |
# flake8: noqa
from .configuration_bert_masked import MaskedBertConfig
from .modeling_bert_masked import (
MaskedBertForMultipleChoice,
MaskedBertForQuestionAnswering,
MaskedBertForSequenceClassification,
MaskedBertForTokenClassification,
MaskedBertModel,
)
from .modules import *
| AdaMix/examples/research_projects/movement-pruning/emmental/__init__.py/0 | {
"file_path": "AdaMix/examples/research_projects/movement-pruning/emmental/__init__.py",
"repo_id": "AdaMix",
"token_count": 107
} | 45 |
#!/usr/bin/env bash
export PYTHONPATH="../":"${PYTHONPATH}"
export WANDB_PROJECT=dmar
export MAX_LEN=128
python finetune.py \
--learning_rate=3e-4 \
--do_train \
--do_predict \
--fp16 \
--val_check_interval 0.25 \
--data_dir $ENRO_DIR \
--max_source_length $MAX_LEN --max_target_length $MAX_LEN --val_max_t... | AdaMix/examples/research_projects/seq2seq-distillation/distil_marian_no_teacher.sh/0 | {
"file_path": "AdaMix/examples/research_projects/seq2seq-distillation/distil_marian_no_teacher.sh",
"repo_id": "AdaMix",
"token_count": 274
} | 46 |
#!/usr/bin/env bash
export PYTHONPATH="../":"${PYTHONPATH}"
python finetune.py \
--learning_rate=3e-5 \
--fp16 \
--do_train \
--val_check_interval=0.25 \
--adam_eps 1e-06 \
--num_train_epochs 6 --src_lang en_XX --tgt_lang ro_RO \
--data_dir $ENRO_DIR \
--max_source_length $MAX_LEN --max... | AdaMix/examples/research_projects/seq2seq-distillation/train_mbart_cc25_enro.sh/0 | {
"file_path": "AdaMix/examples/research_projects/seq2seq-distillation/train_mbart_cc25_enro.sh",
"repo_id": "AdaMix",
"token_count": 273
} | 47 |
# coding=utf-8
# Copyright 2018 HuggingFace Inc..
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or a... | AdaMix/examples/test_examples.py/0 | {
"file_path": "AdaMix/examples/test_examples.py",
"repo_id": "AdaMix",
"token_count": 5464
} | 48 |
<jupyter_start><jupyter_text>Tokenization doesn't have to be slow ! IntroductionBefore going deep into any Machine Learning or Deep Learning Natural Language Processing models, every practitionershould find a way to map raw input strings to a representation understandable by a trainable model.One very simple approach w... | AdaMix/notebooks/01-training-tokenizers.ipynb/0 | {
"file_path": "AdaMix/notebooks/01-training-tokenizers.ipynb",
"repo_id": "AdaMix",
"token_count": 2371
} | 49 |
#!/usr/bin/env python
# coding: utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENS... | AdaMix/scripts/fsmt/fsmt-make-tiny-model.py/0 | {
"file_path": "AdaMix/scripts/fsmt/fsmt-make-tiny-model.py",
"repo_id": "AdaMix",
"token_count": 698
} | 50 |
# coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.a... | AdaMix/src/transformers/benchmark/benchmark_args.py/0 | {
"file_path": "AdaMix/src/transformers/benchmark/benchmark_args.py",
"repo_id": "AdaMix",
"token_count": 1591
} | 51 |
# coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | AdaMix/src/transformers/configuration_utils.py/0 | {
"file_path": "AdaMix/src/transformers/configuration_utils.py",
"repo_id": "AdaMix",
"token_count": 13394
} | 52 |
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | AdaMix/src/transformers/data/metrics/squad_metrics.py/0 | {
"file_path": "AdaMix/src/transformers/data/metrics/squad_metrics.py",
"repo_id": "AdaMix",
"token_count": 13805
} | 53 |
# Copyright 2020 The HuggingFace Team, the AllenNLP library authors. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
... | AdaMix/src/transformers/file_utils.py/0 | {
"file_path": "AdaMix/src/transformers/file_utils.py",
"repo_id": "AdaMix",
"token_count": 24853
} | 54 |
# coding=utf-8
# Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | AdaMix/src/transformers/models/albert/tokenization_albert_fast.py/0 | {
"file_path": "AdaMix/src/transformers/models/albert/tokenization_albert_fast.py",
"repo_id": "AdaMix",
"token_count": 5145
} | 55 |
# coding=utf-8
# Copyright 2020, The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | AdaMix/src/transformers/models/bort/convert_bort_original_gluonnlp_checkpoint_to_pytorch.py/0 | {
"file_path": "AdaMix/src/transformers/models/bort/convert_bort_original_gluonnlp_checkpoint_to_pytorch.py",
"repo_id": "AdaMix",
"token_count": 6174
} | 56 |
# coding=utf-8
# Copyright 2018 Salesforce and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless require... | AdaMix/src/transformers/models/ctrl/tokenization_ctrl.py/0 | {
"file_path": "AdaMix/src/transformers/models/ctrl/tokenization_ctrl.py",
"repo_id": "AdaMix",
"token_count": 3919
} | 57 |
# coding=utf-8
# Copyright 2020 Microsoft and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required... | AdaMix/src/transformers/models/deberta_v2/tokenization_deberta_v2.py/0 | {
"file_path": "AdaMix/src/transformers/models/deberta_v2/tokenization_deberta_v2.py",
"repo_id": "AdaMix",
"token_count": 8468
} | 58 |
# coding=utf-8
# Copyright 2018 The Google AI Language Team Authors, The HuggingFace Inc. team, and the
# Lxmert Authors.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.... | AdaMix/src/transformers/models/lxmert/modeling_tf_lxmert.py/0 | {
"file_path": "AdaMix/src/transformers/models/lxmert/modeling_tf_lxmert.py",
"repo_id": "AdaMix",
"token_count": 29062
} | 59 |
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | AdaMix/src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py/0 | {
"file_path": "AdaMix/src/transformers/models/mbart/convert_mbart_original_checkpoint_to_pytorch.py",
"repo_id": "AdaMix",
"token_count": 1117
} | 60 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.