repo
stringclasses
254 values
file_path
stringlengths
29
241
code
stringlengths
100
233k
tokens
int64
14
69.4k
imezx/Gradien
imezx-Gradien-eaf9c89/src/experimental/optim/SwarmPSO.luau
--!native --!optimize 2 -- multi-particle PSO Optimizer for QIMHNN local Types = require("../../Types") local Tensor = require("../../Tensor") local Util = require("../../Util") local assert = Util.Assert export type SwarmPSOConfig = { swarmSize: number, inertia: number?, cognitive: number?, social: number?, lr:...
1,494
imezx/Gradien
imezx-Gradien-eaf9c89/src/experimental/rl/Feudal.luau
--!native --!optimize 2 local Types = require("../../Types") local Tensor = require("../../Tensor") local Linear = require("../../nn/Linear") local LSTM = require("../../nn/LSTM") local Softmax = require("../../nn/Softmax") local Util = require("../../Util") local Activations = require("../../nn/Activations") local ass...
2,149
imezx/Gradien
imezx-Gradien-eaf9c89/src/extra/RND.luau
--!native --!optimize 2 local Tape = require("../autograd/Tape") local Tensor = require("../Tensor") local Types = require("../Types") local function step(self, s: Types.Tensor): number self.opt:zeroGrad() local g = self.target:forward(s) :: Types.Tensor -- {E,1} local f = self.predictor:forward(s) :: Types.Tensor ...
359
imezx/Gradien
imezx-Gradien-eaf9c89/src/init.luau
--[[ Gradien (by @Eternity_Devs) A strictly-typed, optimized parallel-first ML & DL library for Roblox. Version 1.4.0-rc5 ]] --!strict local Gradien = {} Gradien.Autograd = require("@self/autograd/Tape") Gradien.Init = require("@self/Initializer") Gradien.Metrics = require("@self/Metrics") Gradien.GradClip = require("...
1,552
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/AutoEncoder.luau
--!native --!optimize 2 local Types = require("../Types") local Utils = require("./Utils") local Util = require("../Util") local Linear = require("../nn/Linear") local LayerNorm = require("../nn/LayerNorm") local Dropout = require("../nn/Dropout") local Sequential = require("../nn/Sequential") local assert = Util.Asse...
941
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/ConvNet.luau
--!native --!optimize 2 local Types = require("../Types") local Utils = require("./Utils") local Util = require("../Util") local Conv2d = require("../nn/Conv2d") local BatchNorm2d = require("../nn/BatchNorm2d") local MaxPool2d = require("../nn/MaxPool2d") local AvgPool2d = require("../nn/AvgPool2d") local Dropout = req...
1,420
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/MLP.luau
--!native --!optimize 2 local Types = require("../Types") local Utils = require("./Utils") local Util = require("../Util") local Linear = require("../nn/Linear") local LayerNorm = require("../nn/LayerNorm") local Dropout = require("../nn/Dropout") local Sequential = require("../nn/Sequential") local assert = Util.Asse...
544
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/ResMLP.luau
--!native --!optimize 2 local Types = require("../Types") local Utils = require("./Utils") local Util = require("../Util") local Linear = require("../nn/Linear") local LayerNorm = require("../nn/LayerNorm") local Dropout = require("../nn/Dropout") local Sequential = require("../nn/Sequential") local assert = Util.Asse...
1,011
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/SequenceClassifier.luau
--!native --!optimize 2 local Types = require("../Types") local Utils = require("./Utils") local Tensor = require("../Tensor") local GRU = require("../nn/GRU") local LSTM = require("../nn/LSTM") local RNN = require("../nn/RNN") local Dropout = require("../nn/Dropout") local LayerNorm = require("../nn/LayerNorm") local ...
1,430
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/TransformerEncoder.luau
--!native --!optimize 2 local Types = require("../Types") local Utils = require("./Utils") local Linear = require("../nn/Linear") local LayerNorm = require("../nn/LayerNorm") local Dropout = require("../nn/Dropout") local Sequential = require("../nn/Sequential") local Attention = require("../nn/Attention") export type...
1,196
imezx/Gradien
imezx-Gradien-eaf9c89/src/models/Utils.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Activations = require("../nn/Activations") local Util = require("../Util") local assert = Util.Assert export type ActivationName = "relu" | "gelu" | "tanh" | "silu" | "sigmoid" ...
899
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Activations.luau
--!native --!optimize 2 local Tape = require("../autograd/Tape") local Tensor = require("../Tensor") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert local A = {} local function prep(x: Types.Tensor): Types.Tensor return Tensor.zeros(x._shape, x._dtype, x._requiresGrad) e...
3,198
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Attention.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Linear = require("./Linear") local SoftmaxOps = require("../ops/Softmax") local Util = require("../Util") local assert = Util.Assert type TensorT = Types.Tensor type Module = Ty...
3,181
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/AvgPool2d.luau
--!native --!optimize 2 local Avg = require("../ops/AvgPool2d") local Types = require("../Types") return function(KH: number, KW: number, stride: number) return { forward = function(_, x: Types.Tensor): Types.Tensor return Avg(x, KH, KW, stride) end, parameters = function(_) return {} end, train = func...
98
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/BatchNorm1d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert -- BatchNorm1d (shape {D,B}) --@param D number --@param eps number? --@param momentum number? --@return table return function...
1,559
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/BatchNorm2d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert --@param C number --@return table return function(C: number): Types.Module<Types.Tensor, Types.Tensor> local gamma = Tensor....
1,470
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Conv2d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local convop = require("../ops/Conv2d") local Types = require("../Types") -- {C_in,H,W,B} -> {C_out,H-KH+1,W-KW+1,B}, stride=1, pad=0 (for now), with bias return function(C_in: number, C_out: number, KH: number, KW: nu...
609
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/ConvTranspose2d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local convT = require("../ops/ConvTranspose2d") local Types = require("../Types") --{C_in,H,W,B} -> {C_out, H+KH-1, W+KW-1, B} return function(C_in: number, C_out: number, KH: number, KW: number): Types.Module<Types.Te...
686
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/DropPath.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(p: number?, training: boolean?, rng: Random?): Types.Module<Types.Tensor, Types.Tensor> local p = p or 0.2 local rng = rng or Random.new() local keep = 1 - p return...
494
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Dropout.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(p: number?, training: boolean?, rng: Random?): Types.Module<Types.Tensor, Types.Tensor> local rng = rng or Random.new() local drop = p or 0.2 local keep = 1 - drop ...
390
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Embedding.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(numEmbeds: number, dim: number): Types.Module<Types.Tensor, Types.Tensor> local W = Tensor.zeros({ dim, numEmbeds }, "f64", true) return { forward = function(_, idx...
366
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Flatten.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") -- Flattens everything but the last dimension (batch) for NCHW+B -> {C*H*W,B} return function(): Types.Module<Types.Tensor, Types.Tensor> return { forward = function(_, x: Types.Ten...
574
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Functional.luau
--!native --!optimize 2 local Tape = require("../autograd/Tape") local Types = require("../Types") local FN = {} function FN.relu(x: Types.Tensor): Types.Tensor return Tape.relu(x) end function FN.linear(w: Types.Tensor, b: Types.Tensor, x: Types.Tensor): Types.Tensor task.desynchronize() local out = Tape.matmul(w...
160
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Fused.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Types = require("../Types") local Linear = require("./Linear") local Activations = require("./Activations") local Dropout = require("./Dropout") local Util = require("../Util") local assert = Util.Assert type Tensor = Types.Tensor type Module = Types.Mo...
1,914
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/GRU.luau
--!native --!optimize 2 local Linear = require("./Linear") local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(inF: number, hidF: number): Types.Module<Types.Tensor, Types.Tensor> local Wx = Linear(inF, 3 * hidF) local Wh = Linear(hidF, 3 * h...
1,217
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/GatedMLP.luau
--!native --!optimize 2 local Linear = require("./Linear") local A = require("./Activations") local Types = require("../Types") --@class GatedMLP --@field fc_in any -- Linear(in, 2H) --@field fc_out any -- Linear(H, out) local function forward(self, x: Types.Tensor): Types.Tensor local twoH = self.fc_in:forward(x) -...
276
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/GroupNorm.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert return function(num_groups: number, num_channels: number, eps: number?): Types.Module<Types.Tensor, Types.Tensor> eps = eps ...
1,614
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/LSTM.luau
--!native --!optimize 2 local Linear = require("./Linear") local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(inF: number, hidF: number): Types.Module<Types.Tensor, Types.Tensor> local Wx = Linear(inF, 4 * hidF) local Wh = Linear(hidF, 4 * h...
2,065
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/LayerNorm.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Reduce = require("../ops/Reduce") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert -- LayerNorm across features for each column (shape {D,B}). return function(D: number...
1,150
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Linear.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local function ensureGrad(t: Types.Tensor): Types.Tensor? if not t._grad then t._grad = Tensor.zeros(t._shape, t._dtype, false) end return t._grad end return function( inF: numb...
1,386
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/MaxPool2d.luau
--!native --!optimize 2 local MP = require("../ops/Pool") local Types = require("../Types") return function(KH: number, KW: number, stride: number) return { forward = function(_, x: Types.Tensor): Types.Tensor return MP(x, KH, KW, stride) end, parameters = function(_) return {} end, train = function(_)...
95
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/NoisyLinear.luau
local Tensor = require("../Tensor") local Base = require("./Linear") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(inF: number, outF: number, sigma: number?, rng: Random?): Types.Module<Types.Tensor, Types.Tensor> sigma = sigma or 0.5 rng = rng or Random.new() local base...
548
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/RMSNorm.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local function forward(self: { weight: Types.Tensor, eps: number }, x: Types.Tensor): Types.Tensor task.desynchronize() local D, B = x._shape[1], x._shape[2] local y = Tensor.zeros(...
747
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/RNN.luau
--!native --!optimize 2 local Linear = require("./Linear") local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(inF: number, hidF: number): Types.Module<Types.Tensor, Types.Tensor> local Wx = Linear(inF, hidF) local Wh = Linear(hidF, hidF) lo...
556
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Residual.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert --@param f Types.Module<Types.Tensor, Types.Tensor> return function(f: Types.Module<Types.Tensor, Types.Tensor>): Types.Modul...
413
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Sequential.luau
--!native --!optimize 2 local Types = require("../Types") return function(mods: { Types.Module<Types.Tensor, Types.Tensor> }): Types.Module<Types.Tensor, Types.Tensor> return { _layers = mods, forward = function(_, x: Types.Tensor): Types.Tensor local v = x for _, m in ipairs(mods) do if type(m) == "tabl...
226
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/Softmax.luau
--!native --!optimize 2 local SoftmaxOps = require("../ops/Softmax") local Types = require("../Types") local S = {} function S.forward(logits: Types.Tensor): Types.Tensor return SoftmaxOps.forward(logits) end return S
56
imezx/Gradien
imezx-Gradien-eaf9c89/src/nn/SwiGLU.luau
--!native --!optimize 2 local Activations = require("./Activations") local Types = require("../Types") return function(): Types.Module<Types.Tensor, Types.Tensor> return { --@param x Types.Tensor @ shape {2H, B} --@return Types.Tensor @ shape {H, B} forward = function(_, x: Types.Tensor): Types.Tensor return...
121
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/AvgPool2d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") return function(X: Types.Tensor, KH: number, KW: number, stride: number): Types.Tensor local C, H, W, B = X._shape[1], X._shape[2], X._shape[3], X._shape[4] local Ho = math.floor((H ...
933
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/BLAS.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert -- Matrix multiply: A{M,K} x B{K,N} -> C{M,N} --@param A Tensor --@param B Tensor --@return Tensor local function matmul(A: Types.Tensor, B: Types.Tensor): Types.Tens...
314
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/Conv2d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local BLAS = require("../ops/BLAS") local Util = require("../Util") local assert = Util.Assert local function copyShape(shape: { number }): { number } local out = table.create(#shape)...
1,659
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/ConvTranspose2d.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local BLAS = require("../ops/BLAS") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert local function copyShape(shape: { number }): { number } local out = table.create(#shape)...
1,608
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/Math.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Types = require("../Types") local Math = {} local exp = math.exp local function checkShapes(A: Types.Tensor, B: Types.Tensor, op: string) if #A._storage ~= #B._storage then error( `Math.{op}: Shape mismatch. A has {#A._storage} elements, B has {#...
764
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/Pool.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") -- MaxPool2d kernel KHxKW stride s; input {C,H,W,B} --@param X Tensor --@param KH number --@param KW number --@param stride number --@return Tensor return function(X: Types.Tensor, KH:...
719
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/Reduce.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Types = require("../Types") local Reduce = {} -- Sum per column (shape {D,B} -> {1,B}) --@param X Tensor --@return Tensor function Reduce.sumCols(X: Types.Tensor): Types.Tensor task.desynchronize() local D, B = X._shape[1], X._shape[2] local out = Te...
359
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/Scan.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Types = require("../Types") local Tape = require("../autograd/Tape") local Scan = {} local function ensureGrad(t: Types.Tensor): Types.Tensor? if not t._grad then t._grad = Tensor.zeros(t._shape, t._dtype, false) end return t._grad end function S...
1,646
imezx/Gradien
imezx-Gradien-eaf9c89/src/ops/Softmax.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local SoftmaxOps = {} local function stableSoftmax(logits: Types.Tensor, out: Types.Tensor) local src = logits._storage local dst = out._storage local C, B = logits._shape[1], logi...
598
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Accumulated.luau
--!native --!optimize 2 local Util = require("../Util") local Types = require("../Types") local assert = Util.Assert local function scaleGrads(params: { Types.Tensor }, factor: number) task.desynchronize() for _, p in params do local g = p._grad if g then local gs = g._storage for i = 1, #gs do gs[i] *...
579
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Adafactor.luau
--!native --!optimize 2 local Types = require("../Types") local function flatShape(shape: { number }): (number, number) local rank = #shape local rows = 1 for i = 1, rank - 1 do rows *= shape[i] end return rows, shape[rank] end local function ensureZeros(buf: { number }?, size: number): { number } if not buf ...
1,795
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Adagrad.luau
--!native --!optimize 2 export type Optimizer = { step: (self: Optimizer) -> (), zeroGrad: (self: Optimizer) -> () } return function(params: { any }, lr: number, eps: number?) eps = eps or 1e-8 local G: { [any]: { [number]: number } } = {} local self: Optimizer = { step = function(_) task.desynchronize() fo...
265
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Adam.luau
--!native --!optimize 2 local Types = require("../Types") return function(params: { Types.Tensor }, lr: number, b1: number?, b2: number?, eps: number?, weight_decay: number?) b1 = b1 or 0.9 b2 = b2 or 0.999 eps = eps or 1e-8 local wd = weight_decay or 0 local m: { [Types.Tensor]: { [number]: number } } = {} loca...
790
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/AdamW.luau
--!native --!optimize 2 local Types = require("../Types") return function( params: { Types.Tensor }, lr: number, wd: number, b1: number?, b2: number?, eps: number? ) b1 = b1 or 0.9 b2 = b2 or 0.999 eps = eps or 1e-8 local m: { [Types.Tensor]: { [number]: number } } = {} local v: { [Types.Tensor]: { [number]...
671
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/EMA.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local function update(self, params) task.desynchronize() local m = self.decay for i = 1, #params do local t = params[i] local s = self.shadow[i] for k = 1, #t._storage do s._storage[k] = m * s._storage[k] + (1 - m) * t._storage[k] end end task...
425
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Lion.luau
--!native --!optimize 2 -- Chen et al., 2023. Update uses sign of momentum (no per-param v). --@param params {any} --@param lr number --@param beta1 number? @default 0.9 --@param beta2 number? @default 0.99 --@param weightDecay number? @default 0.0 (decoupled) return function(params, lr: number, beta1: number?, beta2:...
482
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Lookahead.luau
--!native --!optimize 2 --@param params { Tensor } --@param base any -- base optimizer already constructed on params --@param k integer? -- steps between sync (default 5) --@param alpha number? -- interpolation factor (default 0.5) return function(params, base, k: number?, alpha: number?) k = k or 5; alpha = alpha or...
311
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Muon.luau
--!native --!optimize 2 local Types = require("../Types") local Tensor = require("../Tensor") local BLAS = require("../ops/BLAS") -- X_{k+1} = 1.5 * X_k - 0.5 * X_k @ X_k^T @ X_k local function newtonSchulz(G: Types.Tensor, steps: number): Types.Tensor local X = Tensor.detach(G) local _rows, _cols = X._shape[1], X._...
843
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/RMSProp.luau
--!native --!optimize 2 export type Optimizer = { step: (self: Optimizer) -> (), zeroGrad: (self: Optimizer) -> () } return function(params: { any }, lr: number, decay: number?, eps: number?) decay = decay or 0.99 eps = eps or 1e-8 local avg2: { [any]: { [number]: number } } = {} local self: Optimizer = { step =...
306
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/SGD.luau
--!native --!optimize 2 local Types = require("../Types") return function(params: { Types.Tensor }, lr: number, momentum: number?, nesterov: boolean?) momentum = momentum or 0 local vel: { [Types.Tensor]: { [number]: number } } = {} local step = 0 local self = { step = function(_) step += 1 task.desynchron...
464
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Schedulers.luau
--!native --!optimize 2 local S = {} function S.step(lr: number, gamma: number, stepSize: number) return function(step: number) return lr * (gamma ^ math.floor(step / stepSize)) end end function S.cosine(lr: number, lrMin: number, T: number) return function(step: number) local t = math.min(step / T, 1) retur...
559
imezx/Gradien
imezx-Gradien-eaf9c89/src/optim/Sophia.luau
--!native --!optimize 2 local Types = require("../Types") --@param params {Tensor} --@param lr number --@param betas {number}? @default {0.965, 0.99} --@param rho number? @default 0.04 --@param weightDecay number? @default 0.0 return function(params: { Types.Tensor }, lr: number, betas: { number }?, rho: number?, weig...
735
imezx/Gradien
imezx-Gradien-eaf9c89/src/preprocess/MinMaxScaler.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Util = require("../Util") local assert = Util.Assert return function(a: number, b: number) a = a or 0 b = b or 1 local lo: { number }? = nil local hi: { number }? = nil local self self = { fit = function(_, X) task.desynchronize() local D,...
445
imezx/Gradien
imezx-Gradien-eaf9c89/src/preprocess/OneHot.luau
--!native --!optimize 2 local Tensor = require("../Tensor") return function(C: number) return function(idx: { number }) task.desynchronize() local B = #idx local Y = Tensor.zeros({ C, B }) for j = 1, B do Y._storage[(idx[j] - 1) * B + j] = 1 end task.synchronize() return Y end end
100
imezx/Gradien
imezx-Gradien-eaf9c89/src/preprocess/PCA.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local function matvec(D: number, C: Types.Tensor, vec: {number}): {number} -- C * v local out = table.create(D, 0.0) for r = 1, D do local s = 0.0 for c = 1, D do s += C._storage[(r - 1) * D + c] * vec[c] end out[r] = s end return out end loca...
1,016
imezx/Gradien
imezx-Gradien-eaf9c89/src/preprocess/SinusoidalPE.luau
--!native --!optimize 2 local Tensor = require("../Tensor") --@param x Tensor -- {D, B*T} --@param T number -- sequence length --@return Tensor -- new tensor {D, B*T} with PE added return function(x, T: number) local D, BT = x._shape[1], x._shape[2] local B = math.floor(BT / T) local out = Tensor.zeros(x._shape, x....
337
imezx/Gradien
imezx-Gradien-eaf9c89/src/preprocess/StandardScaler.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Util = require("../Util") local assert = Util.Assert return function() local mu: { number } = nil local sigma: { number } = nil local self self = { fit = function(_, X) task.desynchronize() local D, N = X._shape[1], X._shape[2] mu = table...
396
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/A2C.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Losses = require("../nn/Losses") local Softmax = require("../nn/Softmax") local Util = require("../Util") local assert = Util.Assert return function(cfg) local gamma = cfg.gamma or 0.99 local minBatch = cfg.min...
1,597
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/C51DQN.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert local Stabilizers_ok, Stabilizers = pcall(require, "../Utils/Stabilizers") local function softUpdateMaybe(tp, fp, tau: numbe...
2,534
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/DQL.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Losses = require("../nn/Losses") local Util = require("../Util") local assert = Util.Assert -- argmax over actions per column (batch) local function maxPerColumn(Q) task.desynchronize() local A, B = Q._shape[1]...
1,797
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/DoubleDQN.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Losses = require("../nn/Losses") local Util = require("../Util") local assert = Util.Assert local function maxPerColumn(Q) local A, B = Q._shape[1], Q._shape[2] local maxv = table.create(B, -1 / 0) local arg =...
1,917
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/Dueling.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") return function(V, A) local ADim, B = A._shape[1], A._shape[2] local Q = Tensor.zeros({ ADim, B }, A._dtype, V._requiresGrad or A._requiresGrad) task.desynchronize() for j = 1, B do local meanA = 0 for i = 1, A...
479
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/LambdaBuffer.luau
--!native --!optimize 2 local function push(self, t) local out = {} local b = self.buf table.insert(b, t) if t.done then while #b > 0 do local G, w, k = 0, 1, 0 task.desynchronize() for i = 1, #b do k += 1 local ti = b[i] G += w * self.gamma ^ (i - 1) * ti.reward w *= self.lambda if ...
267
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/MultiAgent.luau
--!native --!optimize 2 local Types = require("../Types") local Util = require("../Util") local assert = Util.Assert local function getAgent(self, i: number): Types.Module assert(i >= 1 and i <= self._numAgents, ("MultiAgent: index %d out of range [1,%d]"):format(i, self._numAgents)) return self._agents[i] end loca...
822
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/NStep.luau
--!native --!optimize 2 local NStep = {} function NStep.returns(rewards: { number }, dones: { boolean }, gamma: number, n: number): { number } task.desynchronize() local out = table.create(#rewards, 0) for t = 1, #rewards do local G, g = 0, 1 for k = 0, n - 1 do local i = t + k if i > #rewards or dones[i] ...
175
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/PPO.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Tape = require("../autograd/Tape") local Losses = require("../nn/Losses") local Softmax = require("../nn/Softmax") local Util = require("../Util") local assert = Util.Assert return function(cfg) assert(cfg.policy, "PPO: cfg.policy (policy network) requ...
2,258
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/PrioritizedReplay.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Buffer = require("../Util/Buffer") local Util = require("../Util") local assert = Util.Assert return function(capacity: number, alpha: number, beta0: number) alpha = alpha or 0.6 beta0 = beta0 or 0.4 local size, pos = 0, 1 local S, A, R, NS, D = {},...
1,039
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/Replay.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Buffer = require("../Util/Buffer") local Util = require("../Util") local assert = Util.Assert return function(capacity: number) local size, pos = 0, 1 local S, A, R, NS, D = {}, {}, {}, {}, {} local stateDim: number? = nil local stateDType: string? ...
553
imezx/Gradien
imezx-Gradien-eaf9c89/src/rl/UniformReplay.luau
--!native --!optimize 2 local Tensor = require("../Tensor") local Buffer = require("../Util/Buffer") local Util = require("../Util") local assert = Util.Assert local function packBatch(self, idxs: { number }) local B = #idxs assert(B > 0, "UniformReplay: empty index list") assert(self.stateDim, "UniformReplay: buff...
935
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/Tokenizer.luau
--!native --!optimize 2 local Types = require("./Types") local NFKC = require("./normalizers/NFKC") local ByteLevel = require("./pre_tokenizers/ByteLevel") local ByteLevelProcessor = require("./processors/ByteLevel") local BPE = require("./models/BPE") local BPEDecoder = require("./decoders/BPEDecoder") local Tokenize...
1,466
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/Types.luau
--!strict export type Encoding = { ids: { number }, tokens: { string }, offsets: { { number } }, attention_mask: { number }, special_tokens_mask: { number }, type_ids: { number }, } export type Model = { tokenize: (self: Model, sequences: { string }) -> { any }, token_to_id: (self: Model, token: string) -> num...
283
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/decoders/BPEDecoder.luau
--!optimize 2 local BPEDecoder = {} BPEDecoder.__index = BPEDecoder local magic = { ["^"] = "%^", ["$"] = "%$", ["("] = "%(", [")"] = "%)", ["%"] = "%%", ["."] = "%.", ["["] = "%[", ["]"] = "%]", ["*"] = "%*", ["+"] = "%+", ["-"] = "%-", ["?"] = "%?", } function BPEDecoder.new(suffix: string?, suffix_repl...
241
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/init.luau
--!native --!optimize 2 local Tokenizer = require("@self/Tokenizer") local Types = require("@self/Types") local models = require("@self/models") local decoders = require("@self/decoders") local pre_tokenizers = require("@self/pre_tokenizers") local processors = require("@self/processors") local normalizers = require("@...
141
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/models/BPE.luau
--!native --!optimize 2 local BPE = {} BPE.__index = BPE function BPE.new(vocab: { [string]: number }?, merges: { { string } }?, unk_token: string?, suffix: string?) local self = setmetatable({}, BPE) self.vocab = vocab or {} self.merges = merges or {} self.unk_token = unk_token or "[UNK]" self.suffix = suffix or...
847
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/normalizers/NFKC.luau
--!optimize 2 local NFKC = {} NFKC.__index = NFKC function NFKC.new() local self = setmetatable({}, NFKC) return self end function NFKC:normalize(text: string) text = string.gsub(text, "\194\160", " ") return text end function NFKC:dump() return { type = "NFKC", } end NFKC.from_dump = NFKC.new return NFKC
108
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/normalizers/Sequence.luau
--!optimize 2 local Sequence = {} Sequence.__index = Sequence local Types = require("../Types") function Sequence.new(normalizers: { Types.Normalizer }) local self = setmetatable({}, Sequence) self.normalizers = normalizers or {} return self end function Sequence:normalize(text: string) for _, n in ipairs(self.n...
99
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/normalizers/init.luau
local Sequence = require("@self/Sequence") local NFKC = require("@self/NFKC") return { Sequence = Sequence, NFKC = NFKC, }
37
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/pre_tokenizers/ByteLevel.luau
--!native --!optimize 2 local ByteLevel = {} ByteLevel.__index = ByteLevel function ByteLevel.new(add_prefix_space: boolean?, use_regex: boolean?) local self = setmetatable({}, ByteLevel) self.add_prefix_space = add_prefix_space == nil and true or add_prefix_space self.use_regex = use_regex == nil and true or use_r...
603
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/pre_tokenizers/Sequence.luau
--!native --!optimize 2 local Sequence = {} Sequence.__index = Sequence local Types = require("../Types") function Sequence.new(pre_tokenizers: { Types.PreTokenizer }) local self = setmetatable({}, Sequence) self.pre_tokenizers = pre_tokenizers or {} return self end function Sequence:pre_tokenize(text: string) l...
287
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/pre_tokenizers/Split.luau
--!native --!optimize 2 local Split = {} Split.__index = Split type Behavior = "Removed" | "Isolated" | "MergedWithPrevious" | "MergedWithNext" function Split.new(pattern: string, behavior: Behavior, invert: boolean?) local self = setmetatable({}, Split) self.pattern = pattern self.behavior = behavior or "Removed"...
563
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/processors/ByteLevel.luau
--!native --!optimize 2 local ByteLevel = {} ByteLevel.__index = ByteLevel local Types = require("../Types") function ByteLevel.new(trim_offsets: boolean?) local self = setmetatable({}, ByteLevel) self.trim_offsets = trim_offsets == nil and true or trim_offsets return self end function ByteLevel:process(encoding:...
307
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/processors/TemplateProcessing.luau
--!native --!optimize 2 local Types = require("../Types") local TemplateProcessing = {} TemplateProcessing.__index = TemplateProcessing export type Template = { string | { id: string, type_id: number } } function TemplateProcessing.new(single: Template, pair: Template, special_tokens: { [string]: number }) local se...
536
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/processors/init.luau
local TemplateProcessing = require("@self/TemplateProcessing") local ByteLevel = require("@self/ByteLevel") return { TemplateProcessing = TemplateProcessing, ByteLevel = ByteLevel, }
39
imezx/Gradien
imezx-Gradien-eaf9c89/src/tokenizers/trainers/BpeTrainer.luau
--!native --!optimize 2 local BpeTrainer = {} BpeTrainer.__index = BpeTrainer local Types = require("../Types") local BPE = require("../models/BPE") type BpeTrainerOptions = { vocab_size: number?, min_frequency: number?, special_tokens: { string }?, show_progress: boolean?, initial_alphabet: { string }?, limit_...
1,654
imezx/Gradien
imezx-Gradien-eaf9c89/src/tools/Gradcheck.luau
--!native --!optimize 2 local Tape = require("../autograd/Tape") local Gradcheck = {} function Gradcheck.check(fn: (...number) -> (number, number), inputs: { any }, eps: number?, tol: number?, max_check: number?) eps = eps or 1e-4 tol = tol or 1e-2 local out, grad = fn(table.unpack(inputs)) Tape.backwardFrom(out,...
365
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/ci/RunTests.luau
local ReplicatedStorage = game:GetService("ReplicatedStorage") local ServerScriptService = game:GetService("ServerScriptService") local Test = require(ServerScriptService.TestRunner.Test) -- Run tests return Test.run({ ReplicatedStorage.Modules })
50
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/modules/buffer-util/Buffer.test.luau
local ServerScriptService = game:GetService("ServerScriptService") local Test = require(ServerScriptService.TestRunner.Test) -- local BufferReader = require(script.Parent.BufferReader) -- local BufferWriter = require(script.Parent.BufferWriter) return function(ctx: Test.TestContext) ctx:Describe("Some test", functi...
191
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/modules/buffer-util/BufferReader.luau
--!native local BufferWriter = require(script.Parent.BufferWriter) local DataTypeBuffer = require(script.Parent.DataTypeBuffer) local Types = require(script.Parent.Types) --[=[ @class BufferReader A BufferReader is an abstraction wrapper for `buffer` objects that provides a convenient way of reading out data fro...
1,322
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/modules/buffer-util/BufferWriter.luau
--!native local MAX_SIZE = 1073741824 local DataTypeBuffer = require(script.Parent.DataTypeBuffer) local Types = require(script.Parent.Types) --[=[ @class BufferWriter A BufferWriter is an abstraction wrapper for `buffer` objects that provides a convenient way of writing data to buffers. The internal buffer i...
1,715
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/modules/buffer-util/DataTypeBuffer.luau
--!native local Types = require(script.Parent.Types) type ReadWritePair = { read: <T>(reader: Types.BufferReader) -> T, write: <T>(writer: Types.BufferWriter, value: T) -> (), } local DataTypeBuffer = {} DataTypeBuffer.DataTypesToString = { [BrickColor] = "BrickColor", [CFrame] = "CFrame", [Color3] = "Color3",...
1,732
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/modules/buffer-util/Types.luau
export type DataTypes = | BrickColor | CFrame | Color3 | DateTime | Ray | Rect | Region3 | Region3int16 | UDim | UDim2 | Vector2 | Vector3 | Vector2int16 | Vector3int16 export type BufferReader = { ReadInt8: (self: BufferReader) -> number, ReadUInt8: (self: BufferReader) -> number, ReadInt16: (self: B...
634
Sleitnick/RbxUtil
Sleitnick-RbxUtil-1616bfd/modules/buffer-util/init.luau
--!native local BufferReader = require(script.BufferReader) local BufferWriter = require(script.BufferWriter) --[=[ @class BufferUtil ]=] --[=[ @within BufferUtil @function reader @param buf buffer | string @return BufferReader Creates a BufferReader. The reader can be constructed from either a string or a `...
177