repo stringclasses 254
values | file_path stringlengths 29 241 | code stringlengths 100 233k | tokens int64 14 69.4k |
|---|---|---|---|
imezx/Gradien | imezx-Gradien-eaf9c89/src/experimental/optim/SwarmPSO.luau | --!native
--!optimize 2
-- multi-particle PSO Optimizer for QIMHNN
local Types = require("../../Types")
local Tensor = require("../../Tensor")
local Util = require("../../Util")
local assert = Util.Assert
export type SwarmPSOConfig = {
swarmSize: number,
inertia: number?,
cognitive: number?,
social: number?,
lr:... | 1,494 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/experimental/rl/Feudal.luau | --!native
--!optimize 2
local Types = require("../../Types")
local Tensor = require("../../Tensor")
local Linear = require("../../nn/Linear")
local LSTM = require("../../nn/LSTM")
local Softmax = require("../../nn/Softmax")
local Util = require("../../Util")
local Activations = require("../../nn/Activations")
local ass... | 2,149 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/extra/RND.luau | --!native
--!optimize 2
local Tape = require("../autograd/Tape")
local Tensor = require("../Tensor")
local Types = require("../Types")
local function step(self, s: Types.Tensor): number
self.opt:zeroGrad()
local g = self.target:forward(s) :: Types.Tensor -- {E,1}
local f = self.predictor:forward(s) :: Types.Tensor ... | 359 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/init.luau | --[[
Gradien (by @Eternity_Devs)
A strictly-typed, optimized parallel-first ML & DL library for Roblox.
Version 1.4.0-rc5
]]
--!strict
local Gradien = {}
Gradien.Autograd = require("@self/autograd/Tape")
Gradien.Init = require("@self/Initializer")
Gradien.Metrics = require("@self/Metrics")
Gradien.GradClip = require("... | 1,552 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/AutoEncoder.luau | --!native
--!optimize 2
local Types = require("../Types")
local Utils = require("./Utils")
local Util = require("../Util")
local Linear = require("../nn/Linear")
local LayerNorm = require("../nn/LayerNorm")
local Dropout = require("../nn/Dropout")
local Sequential = require("../nn/Sequential")
local assert = Util.Asse... | 941 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/ConvNet.luau | --!native
--!optimize 2
local Types = require("../Types")
local Utils = require("./Utils")
local Util = require("../Util")
local Conv2d = require("../nn/Conv2d")
local BatchNorm2d = require("../nn/BatchNorm2d")
local MaxPool2d = require("../nn/MaxPool2d")
local AvgPool2d = require("../nn/AvgPool2d")
local Dropout = req... | 1,420 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/MLP.luau | --!native
--!optimize 2
local Types = require("../Types")
local Utils = require("./Utils")
local Util = require("../Util")
local Linear = require("../nn/Linear")
local LayerNorm = require("../nn/LayerNorm")
local Dropout = require("../nn/Dropout")
local Sequential = require("../nn/Sequential")
local assert = Util.Asse... | 544 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/ResMLP.luau | --!native
--!optimize 2
local Types = require("../Types")
local Utils = require("./Utils")
local Util = require("../Util")
local Linear = require("../nn/Linear")
local LayerNorm = require("../nn/LayerNorm")
local Dropout = require("../nn/Dropout")
local Sequential = require("../nn/Sequential")
local assert = Util.Asse... | 1,011 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/SequenceClassifier.luau | --!native
--!optimize 2
local Types = require("../Types")
local Utils = require("./Utils")
local Tensor = require("../Tensor")
local GRU = require("../nn/GRU")
local LSTM = require("../nn/LSTM")
local RNN = require("../nn/RNN")
local Dropout = require("../nn/Dropout")
local LayerNorm = require("../nn/LayerNorm")
local ... | 1,430 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/TransformerEncoder.luau | --!native
--!optimize 2
local Types = require("../Types")
local Utils = require("./Utils")
local Linear = require("../nn/Linear")
local LayerNorm = require("../nn/LayerNorm")
local Dropout = require("../nn/Dropout")
local Sequential = require("../nn/Sequential")
local Attention = require("../nn/Attention")
export type... | 1,196 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/models/Utils.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Activations = require("../nn/Activations")
local Util = require("../Util")
local assert = Util.Assert
export type ActivationName = "relu" | "gelu" | "tanh" | "silu" | "sigmoid" ... | 899 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Activations.luau | --!native
--!optimize 2
local Tape = require("../autograd/Tape")
local Tensor = require("../Tensor")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
local A = {}
local function prep(x: Types.Tensor): Types.Tensor
return Tensor.zeros(x._shape, x._dtype, x._requiresGrad)
e... | 3,198 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Attention.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Linear = require("./Linear")
local SoftmaxOps = require("../ops/Softmax")
local Util = require("../Util")
local assert = Util.Assert
type TensorT = Types.Tensor
type Module = Ty... | 3,181 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/AvgPool2d.luau | --!native
--!optimize 2
local Avg = require("../ops/AvgPool2d")
local Types = require("../Types")
return function(KH: number, KW: number, stride: number)
return {
forward = function(_, x: Types.Tensor): Types.Tensor
return Avg(x, KH, KW, stride)
end,
parameters = function(_)
return {}
end,
train = func... | 98 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/BatchNorm1d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
-- BatchNorm1d (shape {D,B})
--@param D number
--@param eps number?
--@param momentum number?
--@return table
return function... | 1,559 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/BatchNorm2d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
--@param C number
--@return table
return function(C: number): Types.Module<Types.Tensor, Types.Tensor>
local gamma = Tensor.... | 1,470 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Conv2d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local convop = require("../ops/Conv2d")
local Types = require("../Types")
-- {C_in,H,W,B} -> {C_out,H-KH+1,W-KW+1,B}, stride=1, pad=0 (for now), with bias
return function(C_in: number, C_out: number, KH: number, KW: nu... | 609 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/ConvTranspose2d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local convT = require("../ops/ConvTranspose2d")
local Types = require("../Types")
--{C_in,H,W,B} -> {C_out, H+KH-1, W+KW-1, B}
return function(C_in: number, C_out: number, KH: number, KW: number): Types.Module<Types.Te... | 686 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/DropPath.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(p: number?, training: boolean?, rng: Random?): Types.Module<Types.Tensor, Types.Tensor>
local p = p or 0.2
local rng = rng or Random.new()
local keep = 1 - p
return... | 494 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Dropout.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(p: number?, training: boolean?, rng: Random?): Types.Module<Types.Tensor, Types.Tensor>
local rng = rng or Random.new()
local drop = p or 0.2
local keep = 1 - drop
... | 390 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Embedding.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(numEmbeds: number, dim: number): Types.Module<Types.Tensor, Types.Tensor>
local W = Tensor.zeros({ dim, numEmbeds }, "f64", true)
return {
forward = function(_, idx... | 366 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Flatten.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
-- Flattens everything but the last dimension (batch) for NCHW+B -> {C*H*W,B}
return function(): Types.Module<Types.Tensor, Types.Tensor>
return {
forward = function(_, x: Types.Ten... | 574 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Functional.luau | --!native
--!optimize 2
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local FN = {}
function FN.relu(x: Types.Tensor): Types.Tensor
return Tape.relu(x)
end
function FN.linear(w: Types.Tensor, b: Types.Tensor, x: Types.Tensor): Types.Tensor
task.desynchronize()
local out = Tape.matmul(w... | 160 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Fused.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Types = require("../Types")
local Linear = require("./Linear")
local Activations = require("./Activations")
local Dropout = require("./Dropout")
local Util = require("../Util")
local assert = Util.Assert
type Tensor = Types.Tensor
type Module = Types.Mo... | 1,914 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/GRU.luau | --!native
--!optimize 2
local Linear = require("./Linear")
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(inF: number, hidF: number): Types.Module<Types.Tensor, Types.Tensor>
local Wx = Linear(inF, 3 * hidF)
local Wh = Linear(hidF, 3 * h... | 1,217 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/GatedMLP.luau | --!native
--!optimize 2
local Linear = require("./Linear")
local A = require("./Activations")
local Types = require("../Types")
--@class GatedMLP
--@field fc_in any -- Linear(in, 2H)
--@field fc_out any -- Linear(H, out)
local function forward(self, x: Types.Tensor): Types.Tensor
local twoH = self.fc_in:forward(x) -... | 276 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/GroupNorm.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
return function(num_groups: number, num_channels: number, eps: number?): Types.Module<Types.Tensor, Types.Tensor>
eps = eps ... | 1,614 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/LSTM.luau | --!native
--!optimize 2
local Linear = require("./Linear")
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(inF: number, hidF: number): Types.Module<Types.Tensor, Types.Tensor>
local Wx = Linear(inF, 4 * hidF)
local Wh = Linear(hidF, 4 * h... | 2,065 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/LayerNorm.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Reduce = require("../ops/Reduce")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
-- LayerNorm across features for each column (shape {D,B}).
return function(D: number... | 1,150 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Linear.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local function ensureGrad(t: Types.Tensor): Types.Tensor?
if not t._grad then
t._grad = Tensor.zeros(t._shape, t._dtype, false)
end
return t._grad
end
return function(
inF: numb... | 1,386 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/MaxPool2d.luau | --!native
--!optimize 2
local MP = require("../ops/Pool")
local Types = require("../Types")
return function(KH: number, KW: number, stride: number)
return {
forward = function(_, x: Types.Tensor): Types.Tensor
return MP(x, KH, KW, stride)
end,
parameters = function(_)
return {}
end,
train = function(_)... | 95 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/NoisyLinear.luau | local Tensor = require("../Tensor")
local Base = require("./Linear")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(inF: number, outF: number, sigma: number?, rng: Random?): Types.Module<Types.Tensor, Types.Tensor>
sigma = sigma or 0.5
rng = rng or Random.new()
local base... | 548 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/RMSNorm.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local function forward(self: { weight: Types.Tensor, eps: number }, x: Types.Tensor): Types.Tensor
task.desynchronize()
local D, B = x._shape[1], x._shape[2]
local y = Tensor.zeros(... | 747 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/RNN.luau | --!native
--!optimize 2
local Linear = require("./Linear")
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(inF: number, hidF: number): Types.Module<Types.Tensor, Types.Tensor>
local Wx = Linear(inF, hidF)
local Wh = Linear(hidF, hidF)
lo... | 556 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Residual.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
--@param f Types.Module<Types.Tensor, Types.Tensor>
return function(f: Types.Module<Types.Tensor, Types.Tensor>): Types.Modul... | 413 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Sequential.luau | --!native
--!optimize 2
local Types = require("../Types")
return function(mods: { Types.Module<Types.Tensor, Types.Tensor> }): Types.Module<Types.Tensor, Types.Tensor>
return {
_layers = mods,
forward = function(_, x: Types.Tensor): Types.Tensor
local v = x
for _, m in ipairs(mods) do
if type(m) == "tabl... | 226 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/Softmax.luau | --!native
--!optimize 2
local SoftmaxOps = require("../ops/Softmax")
local Types = require("../Types")
local S = {}
function S.forward(logits: Types.Tensor): Types.Tensor
return SoftmaxOps.forward(logits)
end
return S
| 56 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/nn/SwiGLU.luau | --!native
--!optimize 2
local Activations = require("./Activations")
local Types = require("../Types")
return function(): Types.Module<Types.Tensor, Types.Tensor>
return {
--@param x Types.Tensor @ shape {2H, B}
--@return Types.Tensor @ shape {H, B}
forward = function(_, x: Types.Tensor): Types.Tensor
return... | 121 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/AvgPool2d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
return function(X: Types.Tensor, KH: number, KW: number, stride: number): Types.Tensor
local C, H, W, B = X._shape[1], X._shape[2], X._shape[3], X._shape[4]
local Ho = math.floor((H ... | 933 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/BLAS.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
-- Matrix multiply: A{M,K} x B{K,N} -> C{M,N}
--@param A Tensor
--@param B Tensor
--@return Tensor
local function matmul(A: Types.Tensor, B: Types.Tensor): Types.Tens... | 314 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/Conv2d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local BLAS = require("../ops/BLAS")
local Util = require("../Util")
local assert = Util.Assert
local function copyShape(shape: { number }): { number }
local out = table.create(#shape)... | 1,659 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/ConvTranspose2d.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local BLAS = require("../ops/BLAS")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
local function copyShape(shape: { number }): { number }
local out = table.create(#shape)... | 1,608 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/Math.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Types = require("../Types")
local Math = {}
local exp = math.exp
local function checkShapes(A: Types.Tensor, B: Types.Tensor, op: string)
if #A._storage ~= #B._storage then
error(
`Math.{op}: Shape mismatch. A has {#A._storage} elements, B has {#... | 764 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/Pool.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
-- MaxPool2d kernel KHxKW stride s; input {C,H,W,B}
--@param X Tensor
--@param KH number
--@param KW number
--@param stride number
--@return Tensor
return function(X: Types.Tensor, KH:... | 719 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/Reduce.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Types = require("../Types")
local Reduce = {}
-- Sum per column (shape {D,B} -> {1,B})
--@param X Tensor
--@return Tensor
function Reduce.sumCols(X: Types.Tensor): Types.Tensor
task.desynchronize()
local D, B = X._shape[1], X._shape[2]
local out = Te... | 359 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/Scan.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Types = require("../Types")
local Tape = require("../autograd/Tape")
local Scan = {}
local function ensureGrad(t: Types.Tensor): Types.Tensor?
if not t._grad then
t._grad = Tensor.zeros(t._shape, t._dtype, false)
end
return t._grad
end
function S... | 1,646 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/ops/Softmax.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local SoftmaxOps = {}
local function stableSoftmax(logits: Types.Tensor, out: Types.Tensor)
local src = logits._storage
local dst = out._storage
local C, B = logits._shape[1], logi... | 598 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Accumulated.luau | --!native
--!optimize 2
local Util = require("../Util")
local Types = require("../Types")
local assert = Util.Assert
local function scaleGrads(params: { Types.Tensor }, factor: number)
task.desynchronize()
for _, p in params do
local g = p._grad
if g then
local gs = g._storage
for i = 1, #gs do
gs[i] *... | 579 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Adafactor.luau | --!native
--!optimize 2
local Types = require("../Types")
local function flatShape(shape: { number }): (number, number)
local rank = #shape
local rows = 1
for i = 1, rank - 1 do
rows *= shape[i]
end
return rows, shape[rank]
end
local function ensureZeros(buf: { number }?, size: number): { number }
if not buf ... | 1,795 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Adagrad.luau | --!native
--!optimize 2
export type Optimizer = { step: (self: Optimizer) -> (), zeroGrad: (self: Optimizer) -> () }
return function(params: { any }, lr: number, eps: number?)
eps = eps or 1e-8
local G: { [any]: { [number]: number } } = {}
local self: Optimizer = {
step = function(_)
task.desynchronize()
fo... | 265 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Adam.luau | --!native
--!optimize 2
local Types = require("../Types")
return function(params: { Types.Tensor }, lr: number, b1: number?, b2: number?, eps: number?, weight_decay: number?)
b1 = b1 or 0.9
b2 = b2 or 0.999
eps = eps or 1e-8
local wd = weight_decay or 0
local m: { [Types.Tensor]: { [number]: number } } = {}
loca... | 790 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/AdamW.luau | --!native
--!optimize 2
local Types = require("../Types")
return function(
params: { Types.Tensor },
lr: number,
wd: number,
b1: number?,
b2: number?,
eps: number?
)
b1 = b1 or 0.9
b2 = b2 or 0.999
eps = eps or 1e-8
local m: { [Types.Tensor]: { [number]: number } } = {}
local v: { [Types.Tensor]: { [number]... | 671 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/EMA.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local function update(self, params)
task.desynchronize()
local m = self.decay
for i = 1, #params do
local t = params[i]
local s = self.shadow[i]
for k = 1, #t._storage do
s._storage[k] = m * s._storage[k] + (1 - m) * t._storage[k]
end
end
task... | 425 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Lion.luau | --!native
--!optimize 2
-- Chen et al., 2023. Update uses sign of momentum (no per-param v).
--@param params {any}
--@param lr number
--@param beta1 number? @default 0.9
--@param beta2 number? @default 0.99
--@param weightDecay number? @default 0.0 (decoupled)
return function(params, lr: number, beta1: number?, beta2:... | 482 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Lookahead.luau | --!native
--!optimize 2
--@param params { Tensor }
--@param base any -- base optimizer already constructed on params
--@param k integer? -- steps between sync (default 5)
--@param alpha number? -- interpolation factor (default 0.5)
return function(params, base, k: number?, alpha: number?)
k = k or 5; alpha = alpha or... | 311 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Muon.luau | --!native
--!optimize 2
local Types = require("../Types")
local Tensor = require("../Tensor")
local BLAS = require("../ops/BLAS")
-- X_{k+1} = 1.5 * X_k - 0.5 * X_k @ X_k^T @ X_k
local function newtonSchulz(G: Types.Tensor, steps: number): Types.Tensor
local X = Tensor.detach(G)
local _rows, _cols = X._shape[1], X._... | 843 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/RMSProp.luau | --!native
--!optimize 2
export type Optimizer = { step: (self: Optimizer) -> (), zeroGrad: (self: Optimizer) -> () }
return function(params: { any }, lr: number, decay: number?, eps: number?)
decay = decay or 0.99
eps = eps or 1e-8
local avg2: { [any]: { [number]: number } } = {}
local self: Optimizer = {
step =... | 306 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/SGD.luau | --!native
--!optimize 2
local Types = require("../Types")
return function(params: { Types.Tensor }, lr: number, momentum: number?, nesterov: boolean?)
momentum = momentum or 0
local vel: { [Types.Tensor]: { [number]: number } } = {}
local step = 0
local self = {
step = function(_)
step += 1
task.desynchron... | 464 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Schedulers.luau | --!native
--!optimize 2
local S = {}
function S.step(lr: number, gamma: number, stepSize: number)
return function(step: number)
return lr * (gamma ^ math.floor(step / stepSize))
end
end
function S.cosine(lr: number, lrMin: number, T: number)
return function(step: number)
local t = math.min(step / T, 1)
retur... | 559 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/optim/Sophia.luau | --!native
--!optimize 2
local Types = require("../Types")
--@param params {Tensor}
--@param lr number
--@param betas {number}? @default {0.965, 0.99}
--@param rho number? @default 0.04
--@param weightDecay number? @default 0.0
return function(params: { Types.Tensor }, lr: number, betas: { number }?, rho: number?, weig... | 735 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/preprocess/MinMaxScaler.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Util = require("../Util")
local assert = Util.Assert
return function(a: number, b: number)
a = a or 0
b = b or 1
local lo: { number }? = nil
local hi: { number }? = nil
local self
self = {
fit = function(_, X)
task.desynchronize()
local D,... | 445 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/preprocess/OneHot.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
return function(C: number)
return function(idx: { number })
task.desynchronize()
local B = #idx
local Y = Tensor.zeros({ C, B })
for j = 1, B do
Y._storage[(idx[j] - 1) * B + j] = 1
end
task.synchronize()
return Y
end
end
| 100 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/preprocess/PCA.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local function matvec(D: number, C: Types.Tensor, vec: {number}): {number} -- C * v
local out = table.create(D, 0.0)
for r = 1, D do
local s = 0.0
for c = 1, D do
s += C._storage[(r - 1) * D + c] * vec[c]
end
out[r] = s
end
return out
end
loca... | 1,016 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/preprocess/SinusoidalPE.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
--@param x Tensor -- {D, B*T}
--@param T number -- sequence length
--@return Tensor -- new tensor {D, B*T} with PE added
return function(x, T: number)
local D, BT = x._shape[1], x._shape[2]
local B = math.floor(BT / T)
local out = Tensor.zeros(x._shape, x.... | 337 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/preprocess/StandardScaler.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Util = require("../Util")
local assert = Util.Assert
return function()
local mu: { number } = nil
local sigma: { number } = nil
local self
self = {
fit = function(_, X)
task.desynchronize()
local D, N = X._shape[1], X._shape[2]
mu = table... | 396 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/A2C.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Losses = require("../nn/Losses")
local Softmax = require("../nn/Softmax")
local Util = require("../Util")
local assert = Util.Assert
return function(cfg)
local gamma = cfg.gamma or 0.99
local minBatch = cfg.min... | 1,597 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/C51DQN.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
local Stabilizers_ok, Stabilizers = pcall(require, "../Utils/Stabilizers")
local function softUpdateMaybe(tp, fp, tau: numbe... | 2,534 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/DQL.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Losses = require("../nn/Losses")
local Util = require("../Util")
local assert = Util.Assert
-- argmax over actions per column (batch)
local function maxPerColumn(Q)
task.desynchronize()
local A, B = Q._shape[1]... | 1,797 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/DoubleDQN.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Losses = require("../nn/Losses")
local Util = require("../Util")
local assert = Util.Assert
local function maxPerColumn(Q)
local A, B = Q._shape[1], Q._shape[2]
local maxv = table.create(B, -1 / 0)
local arg =... | 1,917 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/Dueling.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
return function(V, A)
local ADim, B = A._shape[1], A._shape[2]
local Q = Tensor.zeros({ ADim, B }, A._dtype, V._requiresGrad or A._requiresGrad)
task.desynchronize()
for j = 1, B do
local meanA = 0
for i = 1, A... | 479 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/LambdaBuffer.luau | --!native
--!optimize 2
local function push(self, t)
local out = {}
local b = self.buf
table.insert(b, t)
if t.done then
while #b > 0 do
local G, w, k = 0, 1, 0
task.desynchronize()
for i = 1, #b do
k += 1
local ti = b[i]
G += w * self.gamma ^ (i - 1) * ti.reward
w *= self.lambda
if ... | 267 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/MultiAgent.luau | --!native
--!optimize 2
local Types = require("../Types")
local Util = require("../Util")
local assert = Util.Assert
local function getAgent(self, i: number): Types.Module
assert(i >= 1 and i <= self._numAgents, ("MultiAgent: index %d out of range [1,%d]"):format(i, self._numAgents))
return self._agents[i]
end
loca... | 822 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/NStep.luau | --!native
--!optimize 2
local NStep = {}
function NStep.returns(rewards: { number }, dones: { boolean }, gamma: number, n: number): { number }
task.desynchronize()
local out = table.create(#rewards, 0)
for t = 1, #rewards do
local G, g = 0, 1
for k = 0, n - 1 do
local i = t + k
if i > #rewards or dones[i] ... | 175 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/PPO.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Tape = require("../autograd/Tape")
local Losses = require("../nn/Losses")
local Softmax = require("../nn/Softmax")
local Util = require("../Util")
local assert = Util.Assert
return function(cfg)
assert(cfg.policy, "PPO: cfg.policy (policy network) requ... | 2,258 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/PrioritizedReplay.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Buffer = require("../Util/Buffer")
local Util = require("../Util")
local assert = Util.Assert
return function(capacity: number, alpha: number, beta0: number)
alpha = alpha or 0.6
beta0 = beta0 or 0.4
local size, pos = 0, 1
local S, A, R, NS, D = {},... | 1,039 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/Replay.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Buffer = require("../Util/Buffer")
local Util = require("../Util")
local assert = Util.Assert
return function(capacity: number)
local size, pos = 0, 1
local S, A, R, NS, D = {}, {}, {}, {}, {}
local stateDim: number? = nil
local stateDType: string? ... | 553 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/rl/UniformReplay.luau | --!native
--!optimize 2
local Tensor = require("../Tensor")
local Buffer = require("../Util/Buffer")
local Util = require("../Util")
local assert = Util.Assert
local function packBatch(self, idxs: { number })
local B = #idxs
assert(B > 0, "UniformReplay: empty index list")
assert(self.stateDim, "UniformReplay: buff... | 935 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/Tokenizer.luau | --!native
--!optimize 2
local Types = require("./Types")
local NFKC = require("./normalizers/NFKC")
local ByteLevel = require("./pre_tokenizers/ByteLevel")
local ByteLevelProcessor = require("./processors/ByteLevel")
local BPE = require("./models/BPE")
local BPEDecoder = require("./decoders/BPEDecoder")
local Tokenize... | 1,466 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/Types.luau | --!strict
export type Encoding = {
ids: { number },
tokens: { string },
offsets: { { number } },
attention_mask: { number },
special_tokens_mask: { number },
type_ids: { number },
}
export type Model = {
tokenize: (self: Model, sequences: { string }) -> { any },
token_to_id: (self: Model, token: string) -> num... | 283 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/decoders/BPEDecoder.luau | --!optimize 2
local BPEDecoder = {}
BPEDecoder.__index = BPEDecoder
local magic = {
["^"] = "%^",
["$"] = "%$",
["("] = "%(",
[")"] = "%)",
["%"] = "%%",
["."] = "%.",
["["] = "%[",
["]"] = "%]",
["*"] = "%*",
["+"] = "%+",
["-"] = "%-",
["?"] = "%?",
}
function BPEDecoder.new(suffix: string?, suffix_repl... | 241 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/init.luau | --!native
--!optimize 2
local Tokenizer = require("@self/Tokenizer")
local Types = require("@self/Types")
local models = require("@self/models")
local decoders = require("@self/decoders")
local pre_tokenizers = require("@self/pre_tokenizers")
local processors = require("@self/processors")
local normalizers = require("@... | 141 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/models/BPE.luau | --!native
--!optimize 2
local BPE = {}
BPE.__index = BPE
function BPE.new(vocab: { [string]: number }?, merges: { { string } }?, unk_token: string?, suffix: string?)
local self = setmetatable({}, BPE)
self.vocab = vocab or {}
self.merges = merges or {}
self.unk_token = unk_token or "[UNK]"
self.suffix = suffix or... | 847 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/normalizers/NFKC.luau | --!optimize 2
local NFKC = {}
NFKC.__index = NFKC
function NFKC.new()
local self = setmetatable({}, NFKC)
return self
end
function NFKC:normalize(text: string)
text = string.gsub(text, "\194\160", " ")
return text
end
function NFKC:dump()
return {
type = "NFKC",
}
end
NFKC.from_dump = NFKC.new
return NFKC
| 108 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/normalizers/Sequence.luau | --!optimize 2
local Sequence = {}
Sequence.__index = Sequence
local Types = require("../Types")
function Sequence.new(normalizers: { Types.Normalizer })
local self = setmetatable({}, Sequence)
self.normalizers = normalizers or {}
return self
end
function Sequence:normalize(text: string)
for _, n in ipairs(self.n... | 99 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/normalizers/init.luau | local Sequence = require("@self/Sequence")
local NFKC = require("@self/NFKC")
return {
Sequence = Sequence,
NFKC = NFKC,
}
| 37 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/pre_tokenizers/ByteLevel.luau | --!native
--!optimize 2
local ByteLevel = {}
ByteLevel.__index = ByteLevel
function ByteLevel.new(add_prefix_space: boolean?, use_regex: boolean?)
local self = setmetatable({}, ByteLevel)
self.add_prefix_space = add_prefix_space == nil and true or add_prefix_space
self.use_regex = use_regex == nil and true or use_r... | 603 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/pre_tokenizers/Sequence.luau | --!native
--!optimize 2
local Sequence = {}
Sequence.__index = Sequence
local Types = require("../Types")
function Sequence.new(pre_tokenizers: { Types.PreTokenizer })
local self = setmetatable({}, Sequence)
self.pre_tokenizers = pre_tokenizers or {}
return self
end
function Sequence:pre_tokenize(text: string)
l... | 287 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/pre_tokenizers/Split.luau | --!native
--!optimize 2
local Split = {}
Split.__index = Split
type Behavior = "Removed" | "Isolated" | "MergedWithPrevious" | "MergedWithNext"
function Split.new(pattern: string, behavior: Behavior, invert: boolean?)
local self = setmetatable({}, Split)
self.pattern = pattern
self.behavior = behavior or "Removed"... | 563 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/processors/ByteLevel.luau | --!native
--!optimize 2
local ByteLevel = {}
ByteLevel.__index = ByteLevel
local Types = require("../Types")
function ByteLevel.new(trim_offsets: boolean?)
local self = setmetatable({}, ByteLevel)
self.trim_offsets = trim_offsets == nil and true or trim_offsets
return self
end
function ByteLevel:process(encoding:... | 307 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/processors/TemplateProcessing.luau | --!native
--!optimize 2
local Types = require("../Types")
local TemplateProcessing = {}
TemplateProcessing.__index = TemplateProcessing
export type Template = { string | { id: string, type_id: number } }
function TemplateProcessing.new(single: Template, pair: Template, special_tokens: { [string]: number })
local se... | 536 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/processors/init.luau | local TemplateProcessing = require("@self/TemplateProcessing")
local ByteLevel = require("@self/ByteLevel")
return {
TemplateProcessing = TemplateProcessing,
ByteLevel = ByteLevel,
}
| 39 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tokenizers/trainers/BpeTrainer.luau | --!native
--!optimize 2
local BpeTrainer = {}
BpeTrainer.__index = BpeTrainer
local Types = require("../Types")
local BPE = require("../models/BPE")
type BpeTrainerOptions = {
vocab_size: number?,
min_frequency: number?,
special_tokens: { string }?,
show_progress: boolean?,
initial_alphabet: { string }?,
limit_... | 1,654 |
imezx/Gradien | imezx-Gradien-eaf9c89/src/tools/Gradcheck.luau | --!native
--!optimize 2
local Tape = require("../autograd/Tape")
local Gradcheck = {}
function Gradcheck.check(fn: (...number) -> (number, number), inputs: { any }, eps: number?, tol: number?, max_check: number?)
eps = eps or 1e-4
tol = tol or 1e-2
local out, grad = fn(table.unpack(inputs))
Tape.backwardFrom(out,... | 365 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/ci/RunTests.luau | local ReplicatedStorage = game:GetService("ReplicatedStorage")
local ServerScriptService = game:GetService("ServerScriptService")
local Test = require(ServerScriptService.TestRunner.Test)
-- Run tests
return Test.run({ ReplicatedStorage.Modules })
| 50 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/modules/buffer-util/Buffer.test.luau | local ServerScriptService = game:GetService("ServerScriptService")
local Test = require(ServerScriptService.TestRunner.Test)
-- local BufferReader = require(script.Parent.BufferReader)
-- local BufferWriter = require(script.Parent.BufferWriter)
return function(ctx: Test.TestContext)
ctx:Describe("Some test", functi... | 191 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/modules/buffer-util/BufferReader.luau | --!native
local BufferWriter = require(script.Parent.BufferWriter)
local DataTypeBuffer = require(script.Parent.DataTypeBuffer)
local Types = require(script.Parent.Types)
--[=[
@class BufferReader
A BufferReader is an abstraction wrapper for `buffer` objects
that provides a convenient way of reading out data fro... | 1,322 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/modules/buffer-util/BufferWriter.luau | --!native
local MAX_SIZE = 1073741824
local DataTypeBuffer = require(script.Parent.DataTypeBuffer)
local Types = require(script.Parent.Types)
--[=[
@class BufferWriter
A BufferWriter is an abstraction wrapper for `buffer` objects
that provides a convenient way of writing data to buffers.
The internal buffer i... | 1,715 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/modules/buffer-util/DataTypeBuffer.luau | --!native
local Types = require(script.Parent.Types)
type ReadWritePair = {
read: <T>(reader: Types.BufferReader) -> T,
write: <T>(writer: Types.BufferWriter, value: T) -> (),
}
local DataTypeBuffer = {}
DataTypeBuffer.DataTypesToString = {
[BrickColor] = "BrickColor",
[CFrame] = "CFrame",
[Color3] = "Color3",... | 1,732 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/modules/buffer-util/Types.luau | export type DataTypes =
| BrickColor
| CFrame
| Color3
| DateTime
| Ray
| Rect
| Region3
| Region3int16
| UDim
| UDim2
| Vector2
| Vector3
| Vector2int16
| Vector3int16
export type BufferReader = {
ReadInt8: (self: BufferReader) -> number,
ReadUInt8: (self: BufferReader) -> number,
ReadInt16: (self: B... | 634 |
Sleitnick/RbxUtil | Sleitnick-RbxUtil-1616bfd/modules/buffer-util/init.luau | --!native
local BufferReader = require(script.BufferReader)
local BufferWriter = require(script.BufferWriter)
--[=[
@class BufferUtil
]=]
--[=[
@within BufferUtil
@function reader
@param buf buffer | string
@return BufferReader
Creates a BufferReader. The reader can be constructed from either
a string or a `... | 177 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.