TylerHilbert commited on
Commit
d822a33
·
verified ·
1 Parent(s): 58cc89e

Reorder to put similar projects close to each other

Browse files
Files changed (1) hide show
  1. PyTorchConference2025_GithubRepos.json +374 -380
PyTorchConference2025_GithubRepos.json CHANGED
@@ -7,27 +7,6 @@
7
  "closest_github_tag": "machine-learning",
8
  "category": "machine learning framework"
9
  },
10
- {
11
- "github_repo_link": "https://github.com/ggml-org/llama.cpp",
12
- "repo_name": "llama.cpp",
13
- "repo_description": "LLM inference in C/C++",
14
- "closest_github_tag": "ggml",
15
- "category": "inference engine"
16
- },
17
- {
18
- "github_repo_link": "https://github.com/onnx/onnx",
19
- "repo_name": "onnx",
20
- "repo_description": "Open standard for machine learning interoperability",
21
- "homepage_link": "https://onnx.ai/",
22
- "closest_github_tag": "deep-learning"
23
- },
24
- {
25
- "github_repo_link": "https://github.com/ray-project/ray",
26
- "repo_name": "ray",
27
- "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
28
- "homepage_link": "https://ray.io",
29
- "closest_github_tag": "deep-learning"
30
- },
31
  {
32
  "github_repo_link": "https://github.com/vllm-project/vllm",
33
  "repo_name": "vllm",
@@ -53,18 +32,11 @@
53
  "category": "inference engine"
54
  },
55
  {
56
- "github_repo_link": "https://github.com/modular/modular",
57
- "repo_name": "modular",
58
- "repo_description": "The Modular Platform (includes MAX & Mojo)",
59
- "homepage_link": "https://docs.modular.com/",
60
- "closest_github_tag": "mojo"
61
- },
62
- {
63
- "github_repo_link": "https://github.com/pytorch/ao",
64
- "repo_name": "ao",
65
- "repo_description": "PyTorch native quantization and sparsity for training and inference",
66
- "homepage_link": "https://pytorch.org/ao/stable/index.html",
67
- "closest_github_tag": "quantization"
68
  },
69
  {
70
  "github_repo_link": "https://github.com/triton-lang/triton",
@@ -73,19 +45,44 @@
73
  "homepage_link": "https://triton-lang.org/",
74
  "category": "dsl"
75
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
  {
77
  "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
78
  "repo_name": "ThunderKittens",
79
  "repo_description": "Tile primitives for speedy kernels"
80
  },
81
- {
82
- "github_repo_link": "https://github.com/gpu-mode/reference-kernels",
83
- "repo_name": "reference-kernels",
84
- "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
85
- "homepage_link": "https://gpumode.com",
86
- "closest_github_tag": "gpu",
87
- "category": "kernels"
88
- },
89
  {
90
  "github_repo_link": "https://github.com/pytorch/executorch",
91
  "repo_name": "executorch",
@@ -95,28 +92,91 @@
95
  "category": "model compiler"
96
  },
97
  {
98
- "github_repo_link": "https://github.com/guandeh17/Self-Forcing",
99
- "repo_name": "Self-Forcing",
100
- "repo_description": "Official codebase for \"Self Forcing: Bridging Training and Inference in Autoregressive Video Diffusion\" (NeurIPS 2025 Spotlight)",
101
-
 
102
  },
103
  {
104
- "github_repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
105
- "repo_name": "StreamDiffusion",
106
- "repo_description": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation"
 
 
107
  },
108
  {
109
- "github_repo_link": "https://github.com/comfyanonymous/ComfyUI",
110
- "repo_name": "ComfyUI",
111
- "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
112
- "homepage_link": "https://www.comfy.org/",
113
- "closest_github_tag": "stable-diffusion"
114
  },
115
  {
116
- "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
117
- "repo_name": "streamv2v",
118
- "repo_description": "Official Pytorch implementation of StreamV2V. ",
119
- "homepage_link": "https://jeff-liangf.github.io/projects/streamv2v/"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
120
  },
121
  {
122
  "github_repo_link": "https://github.com/letta-ai/letta",
@@ -126,59 +186,108 @@
126
  "closest_github_tag": "ai-agents"
127
  },
128
  {
129
- "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
130
- "repo_name": "jupyterlab",
131
- "repo_description": "JupyterLab computational environment.",
132
- "homepage_link": "https://jupyterlab.readthedocs.io/",
133
- "closest_github_tag": "jupyter",
134
- "category": "ui"
135
  },
136
  {
137
- "github_repo_link": "https://github.com/ROCm/rocm-systems",
138
- "repo_name": "rocm-systems",
139
- "repo_description": "super repo for rocm systems projects"
 
 
140
  },
141
  {
142
- "github_repo_link": "https://github.com/NVIDIA/cutlass",
143
- "repo_name": "cutlass",
144
- "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
145
- "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
146
- "closest_github_tag": "cuda"
 
147
  },
148
  {
149
- "github_repo_link": "https://github.com/pytorch/helion",
150
- "repo_name": "helion",
151
- "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
152
- "category": "dsl"
153
  },
154
  {
155
- "github_repo_link": "https://github.com/jax-ml/jax",
156
- "repo_name": "jax",
157
- "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
158
- "homepage_link": "https://docs.jax.dev",
159
- "closest_github_tag": "jax"
160
  },
161
  {
162
- "github_repo_link": "https://github.com/tensorflow/tensorflow",
163
- "repo_name": "tensorflow",
164
- "repo_description": "An Open Source Machine Learning Framework for Everyone",
165
- "homepage_link": "https://tensorflow.org",
166
- "closest_github_tag": "deep-learning",
167
- "category": "machine learning framework"
168
  },
169
  {
170
- "github_repo_link": "https://github.com/deepspeedai/DeepSpeed",
171
- "repo_name": "DeepSpeed",
172
- "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
173
- "homepage_link": "https://www.deepspeed.ai/",
174
- "closest_github_tag": "gpu"
 
175
  },
176
  {
177
- "github_repo_link": "https://github.com/triton-inference-server/server",
178
- "repo_name": "server",
179
- "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
180
- "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
181
- "closest_github_tag": "inference"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
182
  },
183
  {
184
  "github_repo_link": "https://github.com/ROCm/ROCm",
@@ -188,11 +297,31 @@
188
  "closest_github_tag": "documentation"
189
  },
190
  {
191
- "github_repo_link": "https://github.com/llvm/llvm-project",
192
- "repo_name": "llvm-project",
193
- "repo_description": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
194
- "homepage_link": "http://llvm.org",
195
- "category": "compiler"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
196
  },
197
  {
198
  "github_repo_link": "https://github.com/cwpearson/cupti",
@@ -222,20 +351,6 @@
222
  "homepage_link": "https://triton-distributed.readthedocs.io/en/latest/",
223
  "category": "model compiler"
224
  },
225
- {
226
- "github_repo_link": "https://github.com/linkedin/Liger-Kernel",
227
- "repo_name": "Liger-Kernel",
228
- "repo_description": "Efficient Triton Kernels for LLM Training",
229
- "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
230
- "closest_github_tag": "triton",
231
- "category": "kernels"
232
- },
233
- {
234
- "github_repo_link": "https://github.com/thunlp/TritonBench",
235
- "repo_name": "TritonBench",
236
- "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
237
- "category": "benchmark"
238
- },
239
  {
240
  "github_repo_link": "https://github.com/meta-pytorch/tritonparse",
241
  "repo_name": "tritonparse",
@@ -243,6 +358,89 @@
243
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
244
  "closest_github_tag": "triton"
245
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
246
  {
247
  "github_repo_link": "https://github.com/elastic/elasticsearch",
248
  "repo_name": "elasticsearch",
@@ -264,12 +462,6 @@
264
  "repo_description": "Specification and documentation for the Model Context Protocol",
265
  "homepage_link": "https://modelcontextprotocol.io"
266
  },
267
- {
268
- "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
269
- "repo_name": "mcp-agent",
270
- "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
271
- "closest_github_tag": "ai-agents"
272
- },
273
  {
274
  "github_repo_link": "https://github.com/milvus-io/milvus",
275
  "repo_name": "milvus",
@@ -305,29 +497,6 @@
305
  "homepage_link": "https://dstack.ai",
306
  "closest_github_tag": "orchestration"
307
  },
308
- {
309
- "github_repo_link": "https://github.com/numpy/numpy",
310
- "repo_name": "numpy",
311
- "repo_description": "The fundamental package for scientific computing with Python.",
312
- "homepage_link": "https://numpy.org",
313
- "closest_github_tag": "python",
314
- "category": "python library"
315
- },
316
- {
317
- "github_repo_link": "https://github.com/scipy/scipy",
318
- "repo_name": "scipy",
319
- "repo_description": "SciPy library main repository",
320
- "homepage_link": "https://scipy.org",
321
- "closest_github_tag": "python",
322
- "category": "python library"
323
- },
324
- {
325
- "github_repo_link": "https://github.com/numba/numba",
326
- "repo_name": "numba",
327
- "repo_description": "NumPy aware dynamic Python compiler using LLVM",
328
- "homepage_link": "https://numba.pydata.org/",
329
- "closest_github_tag": "compiler"
330
- },
331
  {
332
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
333
  "repo_name": "torchdendrite",
@@ -335,21 +504,6 @@
335
  "closest_github_tag": "scr-3078",
336
  "category": "machine learning framework"
337
  },
338
- {
339
- "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
340
- "repo_name": "lightning-thunder",
341
- "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own."
342
- },
343
- {
344
- "github_repo_link": "https://github.com/pytorch/torchdynamo",
345
- "repo_name": "torchdynamo",
346
- "repo_description": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster."
347
- },
348
- {
349
- "github_repo_link": "https://github.com/microsoft/TileIR",
350
- "repo_name": "TileIR",
351
- "category": "dsl"
352
- },
353
  {
354
  "github_repo_link": "https://github.com/pytorch/torchtitan",
355
  "repo_name": "torchtitan",
@@ -365,13 +519,6 @@
365
  "repo_name": "ort",
366
  "repo_description": "Accelerate PyTorch models with ONNX Runtime"
367
  },
368
- {
369
- "github_repo_link": "https://github.com/NVIDIA/nccl",
370
- "repo_name": "nccl",
371
- "repo_description": "Optimized primitives for collective multi-GPU communication",
372
- "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
373
- "closest_github_tag": "cuda"
374
- },
375
  {
376
  "github_repo_link": "https://github.com/sgl-project/ome",
377
  "repo_name": "ome",
@@ -379,12 +526,6 @@
379
  "homepage_link": "http://docs.sglang.ai/ome/",
380
  "closest_github_tag": "k8s"
381
  },
382
- {
383
- "github_repo_link": "https://github.com/volcengine/verl",
384
- "repo_name": "verl",
385
- "repo_description": "verl: Volcano Engine Reinforcement Learning for LLMs",
386
- "homepage_link": "https://verl.readthedocs.io/en/latest/index.html"
387
- },
388
  {
389
  "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
390
  "repo_name": "neuronx-distributed-inference",
@@ -396,11 +537,6 @@
396
  "repo_description": "PyTorch Single Controller",
397
  "homepage_link": "https://meta-pytorch.org/monarch"
398
  },
399
- {
400
- "github_repo_link": "https://github.com/ai-dynamo/nixl",
401
- "repo_name": "nixl",
402
- "repo_description": "NVIDIA Inference Xfer Library (NIXL)"
403
- },
404
  {
405
  "github_repo_link": "https://github.com/LMCache/LMCache",
406
  "repo_name": "LMCache",
@@ -413,14 +549,7 @@
413
  "repo_name": "rdma-core",
414
  "repo_description": "RDMA core userspace libraries and daemons",
415
  "homepage_link": null,
416
- "closest_github_tag": "linux-kernel"
417
- },
418
- {
419
- "github_repo_link": "https://github.com/NVIDIA/TensorRT",
420
- "repo_name": "TensorRT",
421
- "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
422
- "homepage_link": "https://developer.nvidia.com/tensorrt",
423
- "closest_github_tag": "inference"
424
  },
425
  {
426
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
@@ -436,26 +565,6 @@
436
  "homepage_link": "https://hhvm.com",
437
  "closest_github_tag": "hack"
438
  },
439
- {
440
- "github_repo_link": "https://github.com/vosen/ZLUDA",
441
- "repo_name": "ZLUDA",
442
- "repo_description": "CUDA on non-NVIDIA GPUs",
443
- "homepage_link": "https://vosen.github.io/ZLUDA/",
444
- "closest_github_tag": "cuda"
445
- },
446
- {
447
- "github_repo_link": "https://github.com/vtsynergy/CU2CL",
448
- "repo_name": "CU2CL",
449
- "repo_description": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
450
- "homepage_link": "http://chrec.cs.vt.edu/cu2cl"
451
- },
452
- {
453
- "github_repo_link": "https://github.com/pocl/pocl",
454
- "repo_name": "pocl",
455
- "repo_description": "pocl - Portable Computing Language",
456
- "homepage_link": "https://portablecl.org",
457
- "closest_github_tag": "opencl"
458
- },
459
  {
460
  "github_repo_link": "https://github.com/apache/spark",
461
  "repo_name": "spark",
@@ -463,40 +572,6 @@
463
  "homepage_link": "https://spark.apache.org/",
464
  "closest_github_tag": "big-data"
465
  },
466
- {
467
- "github_repo_link": "https://github.com/codelion/openevolve",
468
- "repo_name": "openevolve",
469
- "repo_description": "Open-source implementation of AlphaEvolve",
470
- "homepage_link": "",
471
- "closest_github_tag": "genetic-algorithm"
472
- },
473
- {
474
- "github_repo_link": "https://github.com/ROCm/hipBLAS",
475
- "repo_name": "hipBLAS",
476
- "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
477
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
478
- "closest_github_tag": "hip"
479
- },
480
- {
481
- "github_repo_link": "https://github.com/ROCm/roctracer",
482
- "repo_name": "roctracer",
483
- "repo_description": "[DEPRECATED] Moved to ROCm/rocm-systems repo ",
484
- "homepage_link": "https://github.com/ROCm/rocm-systems"
485
- },
486
- {
487
- "github_repo_link": "https://github.com/huggingface/peft",
488
- "repo_name": "peft",
489
- "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
490
- "homepage_link": "https://huggingface.co/docs/peft",
491
- "closest_github_tag": "lora"
492
- },
493
- {
494
- "github_repo_link": "https://github.com/ROCm/hip",
495
- "repo_name": "hip",
496
- "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
497
- "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
498
- "closest_github_tag": "hip"
499
- },
500
  {
501
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
502
  "repo_name": "composable_kernel",
@@ -508,17 +583,6 @@
508
  "repo_name": "aiter",
509
  "repo_description": "AI Tensor Engine for ROCm"
510
  },
511
- {
512
- "github_repo_link": "https://github.com/AMDResearch/intelliperf",
513
- "repo_name": "intelliperf",
514
- "repo_description": "Automated bottleneck detection and solution orchestration",
515
- "closest_github_tag": "performance"
516
- },
517
- {
518
- "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
519
- "repo_name": "GEAK-agent",
520
- "repo_description": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically."
521
- },
522
  {
523
  "github_repo_link": "https://github.com/AMD-AGI/torchtitan",
524
  "repo_name": "torchtitan",
@@ -539,19 +603,6 @@
539
  "repo_name": "Megakernels",
540
  "repo_description": "kernels, of the mega variety"
541
  },
542
- {
543
- "github_repo_link": "https://github.com/huggingface/kernels",
544
- "repo_name": "kernels",
545
- "repo_description": "Load compute kernels from the Hub",
546
- "category": "kernels"
547
- },
548
- {
549
- "github_repo_link": "https://github.com/tile-ai/tilelang",
550
- "repo_name": "tilelang",
551
- "repo_description": " Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
552
- "homepage_link": "https://tilelang.com/",
553
- "category": "dsl"
554
- },
555
  {
556
  "github_repo_link": "https://github.com/opencv/opencv",
557
  "repo_name": "opencv",
@@ -571,19 +622,6 @@
571
  "homepage_link": "https://burn.dev",
572
  "closest_github_tag": "machine-learning"
573
  },
574
- {
575
- "github_repo_link": "https://github.com/huggingface/kernels-community",
576
- "repo_name": "kernels-community",
577
- "repo_description": "Kernel sources for https://huggingface.co/kernels-community",
578
- "category": "kernels"
579
- },
580
- {
581
- "github_repo_link": "https://github.com/flashinfer-ai/flashinfer-bench",
582
- "repo_name": "flashinfer-bench",
583
- "repo_description": "Building the Virtuous Cycle for AI-driven LLM Systems",
584
- "homepage_link": "https://bench.flashinfer.ai",
585
- "category": "benchmark"
586
- },
587
  {
588
  "github_repo_link": "https://github.com/OSC/ondemand",
589
  "repo_name": "ondemand",
@@ -598,20 +636,6 @@
598
  "homepage_link": "https://flashinfer.ai",
599
  "closest_github_tag": "attention"
600
  },
601
- {
602
- "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
603
- "repo_name": "KernelBench",
604
- "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
605
- "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
606
- "closest_github_tag": "benchmark",
607
- "category": "benchmark"
608
- },
609
- {
610
- "github_repo_link": "https://github.com/thunlp/TritonBench",
611
- "repo_name": "TritonBench",
612
- "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
613
- "category": "benchmark"
614
- },
615
  {
616
  "github_repo_link": "https://github.com/AutomataLab/cuJSON",
617
  "repo_name": "cuJSON",
@@ -655,21 +679,6 @@
655
  "homepage_link": "https://truss.baseten.co",
656
  "closest_github_tag": "machine-learning"
657
  },
658
- {
659
- "github_repo_link": "https://github.com/laude-institute/terminal-bench",
660
- "repo_name": "terminal-bench",
661
- "repo_description": "A benchmark for LLMs on complicated tasks in the terminal",
662
- "homepage_link": "https://www.tbench.ai",
663
- "category": "benchmark"
664
- },
665
- {
666
- "github_repo_link": "https://github.com/block/goose",
667
- "repo_name": "goose",
668
- "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
669
- "homepage_link": "https://block.github.io/goose/",
670
- "closest_github_tag": "mcp",
671
- "category": "agent"
672
- },
673
  {
674
  "github_repo_link": "https://github.com/kvcache-ai/Mooncake",
675
  "repo_name": "Mooncake",
@@ -677,20 +686,6 @@
677
  "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
678
  "closest_github_tag": "inference"
679
  },
680
- {
681
- "github_repo_link": "https://github.com/SWE-bench/SWE-bench",
682
- "repo_name": "SWE-bench",
683
- "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
684
- "homepage_link": "https://www.swebench.com",
685
- "closest_github_tag": "benchmark",
686
- "category": "benchmark"
687
- },
688
- {
689
- "github_repo_link": "https://github.com/Dao-AILab/quack",
690
- "repo_name": "quack",
691
- "repo_description": "A Quirky Assortment of CuTe Kernels",
692
- "category": "kernels"
693
- },
694
  {
695
  "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
696
  "repo_name": "SYCL-Docs",
@@ -715,6 +710,66 @@
715
  "repo_description": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
716
  "category": "compiler"
717
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
718
  {
719
  "github_repo_link": "https://github.com/ROCm/rocSOLVER",
720
  "repo_name": "rocSOLVER",
@@ -767,66 +822,5 @@
767
  "repo_name": "MIOpen",
768
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
769
  "homepage_link": "https://github.com/ROCm/rocm-libraries"
770
- },
771
- {
772
- "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
773
- "repo_name": "lapack",
774
- "repo_description": "LAPACK development repository",
775
- "closest_github_tag": "linear-algebra"
776
- },
777
- {
778
- "github_repo_link": "https://github.com/ccache/ccache",
779
- "repo_name": "ccache",
780
- "repo_description": "ccache – a fast compiler cache",
781
- "homepage_link": "https://ccache.dev",
782
- "closest_github_tag": "compiler",
783
- "category": "compiler"
784
- },
785
- {
786
- "github_repo_link": "https://github.com/ROCm/omnitrace",
787
- "repo_name": "omnitrace",
788
- "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
789
- "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
790
- "closest_github_tag": "performance-analysis"
791
- },
792
- {
793
- "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
794
- "repo_name": "OpenCL-SDK",
795
- "repo_description": "OpenCL SDK"
796
- },
797
- {
798
- "github_repo_link": "https://github.com/meta-llama/synthetic-data-kit",
799
- "repo_name": "synthetic-data-kit",
800
- "repo_description": "Tool for generating high quality Synthetic datasets",
801
- "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
802
- "closest_github_tag": "generation"
803
- },
804
- {
805
- "github_repo_link": "https://github.com/unslothai/unsloth",
806
- "repo_name": "unsloth",
807
- "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
808
- "homepage_link": "https://docs.unsloth.ai/",
809
- "closest_github_tag": "unsloth"
810
- },
811
- {
812
- "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
813
- "repo_name": "Vulkan-Docs",
814
- "repo_description": "The Vulkan API Specification and related tools"
815
- },
816
- {
817
- "github_repo_link": "https://github.com/tensorflow/tflite-micro",
818
- "repo_name": "tflite-micro",
819
- "repo_description": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors)."
820
- },
821
- {
822
- "github_repo_link": "https://github.com/Wan-Video/Wan2.2",
823
- "repo_name": "Wan2.2",
824
- "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
825
- "homepage_link": "https://wan.video",
826
- "closest_github_tag": "video-generation"
827
- },
828
- {
829
- "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
830
- "repo_name": "Primus-Turbo"
831
  }
832
  ]
 
7
  "closest_github_tag": "machine-learning",
8
  "category": "machine learning framework"
9
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  {
11
  "github_repo_link": "https://github.com/vllm-project/vllm",
12
  "repo_name": "vllm",
 
32
  "category": "inference engine"
33
  },
34
  {
35
+ "github_repo_link": "https://github.com/ggml-org/llama.cpp",
36
+ "repo_name": "llama.cpp",
37
+ "repo_description": "LLM inference in C/C++",
38
+ "closest_github_tag": "ggml",
39
+ "category": "inference engine"
 
 
 
 
 
 
 
40
  },
41
  {
42
  "github_repo_link": "https://github.com/triton-lang/triton",
 
45
  "homepage_link": "https://triton-lang.org/",
46
  "category": "dsl"
47
  },
48
+ {
49
+ "github_repo_link": "https://github.com/pytorch/helion",
50
+ "repo_name": "helion",
51
+ "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
52
+ "category": "dsl"
53
+ },
54
+ {
55
+ "github_repo_link": "https://github.com/microsoft/TileIR",
56
+ "repo_name": "TileIR",
57
+ "category": "dsl"
58
+ },
59
+ {
60
+ "github_repo_link": "https://github.com/tile-ai/tilelang",
61
+ "repo_name": "tilelang",
62
+ "repo_description": " Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
63
+ "homepage_link": "https://tilelang.com/",
64
+ "category": "dsl"
65
+ },
66
+ {
67
+ "github_repo_link": "https://github.com/NVIDIA/cutlass",
68
+ "repo_name": "cutlass",
69
+ "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
70
+ "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
71
+ "closest_github_tag": "cuda"
72
+ },
73
+ {
74
+ "github_repo_link": "https://github.com/tensorflow/tensorflow",
75
+ "repo_name": "tensorflow",
76
+ "repo_description": "An Open Source Machine Learning Framework for Everyone",
77
+ "homepage_link": "https://tensorflow.org",
78
+ "closest_github_tag": "deep-learning",
79
+ "category": "machine learning framework"
80
+ },
81
  {
82
  "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
83
  "repo_name": "ThunderKittens",
84
  "repo_description": "Tile primitives for speedy kernels"
85
  },
 
 
 
 
 
 
 
 
86
  {
87
  "github_repo_link": "https://github.com/pytorch/executorch",
88
  "repo_name": "executorch",
 
92
  "category": "model compiler"
93
  },
94
  {
95
+ "github_repo_link": "https://github.com/onnx/onnx",
96
+ "repo_name": "onnx",
97
+ "repo_description": "Open standard for machine learning interoperability",
98
+ "homepage_link": "https://onnx.ai/",
99
+ "closest_github_tag": "deep-learning"
100
  },
101
  {
102
+ "github_repo_link": "https://github.com/ray-project/ray",
103
+ "repo_name": "ray",
104
+ "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
105
+ "homepage_link": "https://ray.io",
106
+ "closest_github_tag": "deep-learning"
107
  },
108
  {
109
+ "github_repo_link": "https://github.com/jax-ml/jax",
110
+ "repo_name": "jax",
111
+ "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
112
+ "homepage_link": "https://docs.jax.dev",
113
+ "closest_github_tag": "jax"
114
  },
115
  {
116
+ "github_repo_link": "https://github.com/llvm/llvm-project",
117
+ "repo_name": "llvm-project",
118
+ "repo_description": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
119
+ "homepage_link": "http://llvm.org",
120
+ "category": "compiler"
121
+ },
122
+ {
123
+ "github_repo_link": "https://github.com/NVIDIA/TensorRT",
124
+ "repo_name": "TensorRT",
125
+ "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
126
+ "homepage_link": "https://developer.nvidia.com/tensorrt",
127
+ "closest_github_tag": "inference"
128
+ },
129
+ {
130
+ "github_repo_link": "https://github.com/pytorch/ao",
131
+ "repo_name": "ao",
132
+ "repo_description": "PyTorch native quantization and sparsity for training and inference",
133
+ "homepage_link": "https://pytorch.org/ao/stable/index.html",
134
+ "closest_github_tag": "quantization"
135
+ },
136
+ {
137
+ "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
138
+ "repo_name": "GEAK-agent",
139
+ "repo_description": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically."
140
+ },
141
+ {
142
+ "github_repo_link": "https://github.com/block/goose",
143
+ "repo_name": "goose",
144
+ "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
145
+ "homepage_link": "https://block.github.io/goose/",
146
+ "closest_github_tag": "mcp",
147
+ "category": "agent"
148
+ },
149
+ {
150
+ "github_repo_link": "https://github.com/codelion/openevolve",
151
+ "repo_name": "openevolve",
152
+ "repo_description": "Open-source implementation of AlphaEvolve",
153
+ "homepage_link": "",
154
+ "closest_github_tag": "genetic-algorithm"
155
+ },
156
+ {
157
+ "github_repo_link": "https://github.com/volcengine/verl",
158
+ "repo_name": "verl",
159
+ "repo_description": "verl: Volcano Engine Reinforcement Learning for LLMs",
160
+ "homepage_link": "https://verl.readthedocs.io/en/latest/index.html"
161
+ },
162
+ {
163
+ "github_repo_link": "https://github.com/huggingface/peft",
164
+ "repo_name": "peft",
165
+ "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
166
+ "homepage_link": "https://huggingface.co/docs/peft",
167
+ "closest_github_tag": "lora"
168
+ },
169
+ {
170
+ "github_repo_link": "https://github.com/Dao-AILab/quack",
171
+ "repo_name": "quack",
172
+ "repo_description": "A Quirky Assortment of CuTe Kernels",
173
+ "category": "kernels"
174
+ },
175
+ {
176
+ "github_repo_link": "https://github.com/AMDResearch/intelliperf",
177
+ "repo_name": "intelliperf",
178
+ "repo_description": "Automated bottleneck detection and solution orchestration",
179
+ "closest_github_tag": "performance"
180
  },
181
  {
182
  "github_repo_link": "https://github.com/letta-ai/letta",
 
186
  "closest_github_tag": "ai-agents"
187
  },
188
  {
189
+ "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
190
+ "repo_name": "mcp-agent",
191
+ "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
192
+ "closest_github_tag": "ai-agents"
 
 
193
  },
194
  {
195
+ "github_repo_link": "https://github.com/modular/modular",
196
+ "repo_name": "modular",
197
+ "repo_description": "The Modular Platform (includes MAX & Mojo)",
198
+ "homepage_link": "https://docs.modular.com/",
199
+ "closest_github_tag": "mojo"
200
  },
201
  {
202
+ "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
203
+ "repo_name": "KernelBench",
204
+ "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
205
+ "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
206
+ "closest_github_tag": "benchmark",
207
+ "category": "benchmark"
208
  },
209
  {
210
+ "github_repo_link": "https://github.com/thunlp/TritonBench",
211
+ "repo_name": "TritonBench",
212
+ "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
213
+ "category": "benchmark"
214
  },
215
  {
216
+ "github_repo_link": "https://github.com/flashinfer-ai/flashinfer-bench",
217
+ "repo_name": "flashinfer-bench",
218
+ "repo_description": "Building the Virtuous Cycle for AI-driven LLM Systems",
219
+ "homepage_link": "https://bench.flashinfer.ai",
220
+ "category": "benchmark"
221
  },
222
  {
223
+ "github_repo_link": "https://github.com/laude-institute/terminal-bench",
224
+ "repo_name": "terminal-bench",
225
+ "repo_description": "A benchmark for LLMs on complicated tasks in the terminal",
226
+ "homepage_link": "https://www.tbench.ai",
227
+ "category": "benchmark"
 
228
  },
229
  {
230
+ "github_repo_link": "https://github.com/SWE-bench/SWE-bench",
231
+ "repo_name": "SWE-bench",
232
+ "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
233
+ "homepage_link": "https://www.swebench.com",
234
+ "closest_github_tag": "benchmark",
235
+ "category": "benchmark"
236
  },
237
  {
238
+ "github_repo_link": "https://github.com/gpu-mode/reference-kernels",
239
+ "repo_name": "reference-kernels",
240
+ "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
241
+ "homepage_link": "https://gpumode.com",
242
+ "closest_github_tag": "gpu",
243
+ "category": "kernels"
244
+ },
245
+ {
246
+ "github_repo_link": "https://github.com/linkedin/Liger-Kernel",
247
+ "repo_name": "Liger-Kernel",
248
+ "repo_description": "Efficient Triton Kernels for LLM Training",
249
+ "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
250
+ "closest_github_tag": "triton",
251
+ "category": "kernels"
252
+ },
253
+ {
254
+ "github_repo_link": "https://github.com/huggingface/kernels",
255
+ "repo_name": "kernels",
256
+ "repo_description": "Load compute kernels from the Hub",
257
+ "category": "kernels"
258
+ },
259
+ {
260
+ "github_repo_link": "https://github.com/huggingface/kernels-community",
261
+ "repo_name": "kernels-community",
262
+ "repo_description": "Kernel sources for https://huggingface.co/kernels-community",
263
+ "category": "kernels"
264
+ },
265
+ {
266
+ "github_repo_link": "https://github.com/unslothai/unsloth",
267
+ "repo_name": "unsloth",
268
+ "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
269
+ "homepage_link": "https://docs.unsloth.ai/",
270
+ "closest_github_tag": "unsloth"
271
+ },
272
+ {
273
+ "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
274
+ "repo_name": "jupyterlab",
275
+ "repo_description": "JupyterLab computational environment.",
276
+ "homepage_link": "https://jupyterlab.readthedocs.io/",
277
+ "closest_github_tag": "jupyter",
278
+ "category": "ui"
279
+ },
280
+ {
281
+ "github_repo_link": "https://github.com/ROCm/rocm-systems",
282
+ "repo_name": "rocm-systems",
283
+ "repo_description": "super repo for rocm systems projects"
284
+ },
285
+ {
286
+ "github_repo_link": "https://github.com/ROCm/hip",
287
+ "repo_name": "hip",
288
+ "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
289
+ "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
290
+ "closest_github_tag": "hip"
291
  },
292
  {
293
  "github_repo_link": "https://github.com/ROCm/ROCm",
 
297
  "closest_github_tag": "documentation"
298
  },
299
  {
300
+ "github_repo_link": "https://github.com/ROCm/omnitrace",
301
+ "repo_name": "omnitrace",
302
+ "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
303
+ "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
304
+ "closest_github_tag": "performance-analysis"
305
+ },
306
+ {
307
+ "github_repo_link": "https://github.com/vosen/ZLUDA",
308
+ "repo_name": "ZLUDA",
309
+ "repo_description": "CUDA on non-NVIDIA GPUs",
310
+ "homepage_link": "https://vosen.github.io/ZLUDA/",
311
+ "closest_github_tag": "cuda"
312
+ },
313
+ {
314
+ "github_repo_link": "https://github.com/vtsynergy/CU2CL",
315
+ "repo_name": "CU2CL",
316
+ "repo_description": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
317
+ "homepage_link": "http://chrec.cs.vt.edu/cu2cl"
318
+ },
319
+ {
320
+ "github_repo_link": "https://github.com/pocl/pocl",
321
+ "repo_name": "pocl",
322
+ "repo_description": "pocl - Portable Computing Language",
323
+ "homepage_link": "https://portablecl.org",
324
+ "closest_github_tag": "opencl"
325
  },
326
  {
327
  "github_repo_link": "https://github.com/cwpearson/cupti",
 
351
  "homepage_link": "https://triton-distributed.readthedocs.io/en/latest/",
352
  "category": "model compiler"
353
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
354
  {
355
  "github_repo_link": "https://github.com/meta-pytorch/tritonparse",
356
  "repo_name": "tritonparse",
 
358
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
359
  "closest_github_tag": "triton"
360
  },
361
+ {
362
+ "github_repo_link": "https://github.com/numpy/numpy",
363
+ "repo_name": "numpy",
364
+ "repo_description": "The fundamental package for scientific computing with Python.",
365
+ "homepage_link": "https://numpy.org",
366
+ "closest_github_tag": "python",
367
+ "category": "python library"
368
+ },
369
+ {
370
+ "github_repo_link": "https://github.com/scipy/scipy",
371
+ "repo_name": "scipy",
372
+ "repo_description": "SciPy library main repository",
373
+ "homepage_link": "https://scipy.org",
374
+ "closest_github_tag": "python",
375
+ "category": "python library"
376
+ },
377
+ {
378
+ "github_repo_link": "https://github.com/numba/numba",
379
+ "repo_name": "numba",
380
+ "repo_description": "NumPy aware dynamic Python compiler using LLVM",
381
+ "homepage_link": "https://numba.pydata.org/",
382
+ "closest_github_tag": "compiler"
383
+ },
384
+ {
385
+ "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
386
+ "repo_name": "lightning-thunder",
387
+ "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own."
388
+ },
389
+ {
390
+ "github_repo_link": "https://github.com/pytorch/torchdynamo",
391
+ "repo_name": "torchdynamo",
392
+ "repo_description": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster."
393
+ },
394
+ {
395
+ "github_repo_link": "https://github.com/NVIDIA/nccl",
396
+ "repo_name": "nccl",
397
+ "repo_description": "Optimized primitives for collective multi-GPU communication",
398
+ "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
399
+ "closest_github_tag": "cuda"
400
+ },
401
+ {
402
+ "github_repo_link": "https://github.com/ai-dynamo/nixl",
403
+ "repo_name": "nixl",
404
+ "repo_description": "NVIDIA Inference Xfer Library (NIXL)"
405
+ },
406
+ {
407
+ "github_repo_link": "https://github.com/guandeh17/Self-Forcing",
408
+ "repo_name": "Self-Forcing",
409
+ "repo_description": "Official codebase for \"Self Forcing: Bridging Training and Inference in Autoregressive Video Diffusion\" (NeurIPS 2025 Spotlight)",
410
+
411
+ },
412
+ {
413
+ "github_repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
414
+ "repo_name": "StreamDiffusion",
415
+ "repo_description": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation"
416
+ },
417
+ {
418
+ "github_repo_link": "https://github.com/comfyanonymous/ComfyUI",
419
+ "repo_name": "ComfyUI",
420
+ "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
421
+ "homepage_link": "https://www.comfy.org/",
422
+ "closest_github_tag": "stable-diffusion"
423
+ },
424
+ {
425
+ "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
426
+ "repo_name": "streamv2v",
427
+ "repo_description": "Official Pytorch implementation of StreamV2V. ",
428
+ "homepage_link": "https://jeff-liangf.github.io/projects/streamv2v/"
429
+ },
430
+ {
431
+ "github_repo_link": "https://github.com/deepspeedai/DeepSpeed",
432
+ "repo_name": "DeepSpeed",
433
+ "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
434
+ "homepage_link": "https://www.deepspeed.ai/",
435
+ "closest_github_tag": "gpu"
436
+ },
437
+ {
438
+ "github_repo_link": "https://github.com/triton-inference-server/server",
439
+ "repo_name": "server",
440
+ "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
441
+ "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
442
+ "closest_github_tag": "inference"
443
+ },
444
  {
445
  "github_repo_link": "https://github.com/elastic/elasticsearch",
446
  "repo_name": "elasticsearch",
 
462
  "repo_description": "Specification and documentation for the Model Context Protocol",
463
  "homepage_link": "https://modelcontextprotocol.io"
464
  },
 
 
 
 
 
 
465
  {
466
  "github_repo_link": "https://github.com/milvus-io/milvus",
467
  "repo_name": "milvus",
 
497
  "homepage_link": "https://dstack.ai",
498
  "closest_github_tag": "orchestration"
499
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
500
  {
501
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
502
  "repo_name": "torchdendrite",
 
504
  "closest_github_tag": "scr-3078",
505
  "category": "machine learning framework"
506
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
507
  {
508
  "github_repo_link": "https://github.com/pytorch/torchtitan",
509
  "repo_name": "torchtitan",
 
519
  "repo_name": "ort",
520
  "repo_description": "Accelerate PyTorch models with ONNX Runtime"
521
  },
 
 
 
 
 
 
 
522
  {
523
  "github_repo_link": "https://github.com/sgl-project/ome",
524
  "repo_name": "ome",
 
526
  "homepage_link": "http://docs.sglang.ai/ome/",
527
  "closest_github_tag": "k8s"
528
  },
 
 
 
 
 
 
529
  {
530
  "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
531
  "repo_name": "neuronx-distributed-inference",
 
537
  "repo_description": "PyTorch Single Controller",
538
  "homepage_link": "https://meta-pytorch.org/monarch"
539
  },
 
 
 
 
 
540
  {
541
  "github_repo_link": "https://github.com/LMCache/LMCache",
542
  "repo_name": "LMCache",
 
549
  "repo_name": "rdma-core",
550
  "repo_description": "RDMA core userspace libraries and daemons",
551
  "homepage_link": null,
552
+ "closest_github_tag": "linux-kernel"
 
 
 
 
 
 
 
553
  },
554
  {
555
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
 
565
  "homepage_link": "https://hhvm.com",
566
  "closest_github_tag": "hack"
567
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
568
  {
569
  "github_repo_link": "https://github.com/apache/spark",
570
  "repo_name": "spark",
 
572
  "homepage_link": "https://spark.apache.org/",
573
  "closest_github_tag": "big-data"
574
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
575
  {
576
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
577
  "repo_name": "composable_kernel",
 
583
  "repo_name": "aiter",
584
  "repo_description": "AI Tensor Engine for ROCm"
585
  },
 
 
 
 
 
 
 
 
 
 
 
586
  {
587
  "github_repo_link": "https://github.com/AMD-AGI/torchtitan",
588
  "repo_name": "torchtitan",
 
603
  "repo_name": "Megakernels",
604
  "repo_description": "kernels, of the mega variety"
605
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
606
  {
607
  "github_repo_link": "https://github.com/opencv/opencv",
608
  "repo_name": "opencv",
 
622
  "homepage_link": "https://burn.dev",
623
  "closest_github_tag": "machine-learning"
624
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
625
  {
626
  "github_repo_link": "https://github.com/OSC/ondemand",
627
  "repo_name": "ondemand",
 
636
  "homepage_link": "https://flashinfer.ai",
637
  "closest_github_tag": "attention"
638
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
639
  {
640
  "github_repo_link": "https://github.com/AutomataLab/cuJSON",
641
  "repo_name": "cuJSON",
 
679
  "homepage_link": "https://truss.baseten.co",
680
  "closest_github_tag": "machine-learning"
681
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
682
  {
683
  "github_repo_link": "https://github.com/kvcache-ai/Mooncake",
684
  "repo_name": "Mooncake",
 
686
  "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
687
  "closest_github_tag": "inference"
688
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
689
  {
690
  "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
691
  "repo_name": "SYCL-Docs",
 
710
  "repo_description": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
711
  "category": "compiler"
712
  },
713
+ {
714
+ "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
715
+ "repo_name": "lapack",
716
+ "repo_description": "LAPACK development repository",
717
+ "closest_github_tag": "linear-algebra"
718
+ },
719
+ {
720
+ "github_repo_link": "https://github.com/ccache/ccache",
721
+ "repo_name": "ccache",
722
+ "repo_description": "ccache – a fast compiler cache",
723
+ "homepage_link": "https://ccache.dev",
724
+ "closest_github_tag": "compiler",
725
+ "category": "compiler"
726
+ },
727
+ {
728
+ "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
729
+ "repo_name": "OpenCL-SDK",
730
+ "repo_description": "OpenCL SDK"
731
+ },
732
+ {
733
+ "github_repo_link": "https://github.com/meta-llama/synthetic-data-kit",
734
+ "repo_name": "synthetic-data-kit",
735
+ "repo_description": "Tool for generating high quality Synthetic datasets",
736
+ "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
737
+ "closest_github_tag": "generation"
738
+ },
739
+ {
740
+ "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
741
+ "repo_name": "Vulkan-Docs",
742
+ "repo_description": "The Vulkan API Specification and related tools"
743
+ },
744
+ {
745
+ "github_repo_link": "https://github.com/tensorflow/tflite-micro",
746
+ "repo_name": "tflite-micro",
747
+ "repo_description": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors)."
748
+ },
749
+ {
750
+ "github_repo_link": "https://github.com/Wan-Video/Wan2.2",
751
+ "repo_name": "Wan2.2",
752
+ "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
753
+ "homepage_link": "https://wan.video",
754
+ "closest_github_tag": "video-generation"
755
+ },
756
+ {
757
+ "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
758
+ "repo_name": "Primus-Turbo"
759
+ }
760
+ {
761
+ "github_repo_link": "https://github.com/ROCm/hipBLAS",
762
+ "repo_name": "hipBLAS",
763
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
764
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
765
+ "closest_github_tag": "hip"
766
+ },
767
+ {
768
+ "github_repo_link": "https://github.com/ROCm/roctracer",
769
+ "repo_name": "roctracer",
770
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-systems repo ",
771
+ "homepage_link": "https://github.com/ROCm/rocm-systems"
772
+ },
773
  {
774
  "github_repo_link": "https://github.com/ROCm/rocSOLVER",
775
  "repo_name": "rocSOLVER",
 
822
  "repo_name": "MIOpen",
823
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
824
  "homepage_link": "https://github.com/ROCm/rocm-libraries"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
825
  }
826
  ]