TylerHilbert commited on
Commit
8dcde9e
·
1 Parent(s): 0bc5c22

Sorted by unique contributors in 2025

Browse files
Files changed (1) hide show
  1. PyTorchConference2025_GithubRepos.json +1038 -1038
PyTorchConference2025_GithubRepos.json CHANGED
@@ -1,130 +1,4 @@
1
  [
2
- {
3
- "repo_name": "goose",
4
- "repo_link": "https://github.com/block/goose",
5
- "category": "agent",
6
- "github_about_section": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
7
- "homepage_link": "https://block.github.io/goose",
8
- "github_topic_closest_fit": "ai-agents",
9
- "contributors_all": 332,
10
- "contributors_2025": 319,
11
- "contributors_2024": 32,
12
- "contributors_2023": 0
13
- },
14
- {
15
- "repo_name": "ray",
16
- "repo_link": "https://github.com/ray-project/ray",
17
- "github_about_section": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
18
- "homepage_link": "https://ray.io",
19
- "contributors_all": 1381,
20
- "contributors_2025": 397,
21
- "contributors_2024": 223,
22
- "contributors_2023": 230
23
- },
24
- {
25
- "repo_name": "flashinfer-bench",
26
- "repo_link": "https://github.com/flashinfer-ai/flashinfer-bench",
27
- "category": "benchmark",
28
- "github_about_section": "Building the Virtuous Cycle for AI-driven LLM Systems",
29
- "homepage_link": "https://bench.flashinfer.ai",
30
- "github_topic_closest_fit": "benchmark",
31
- "contributors_all": 12,
32
- "contributors_2025": 11,
33
- "contributors_2024": 0,
34
- "contributors_2023": 0
35
- },
36
- {
37
- "repo_name": "KernelBench",
38
- "repo_link": "https://github.com/ScalingIntelligence/KernelBench",
39
- "category": "benchmark",
40
- "github_about_section": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
41
- "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench",
42
- "github_topic_closest_fit": "benchmark",
43
- "contributors_all": 19,
44
- "contributors_2025": 16,
45
- "contributors_2024": 3,
46
- "contributors_2023": 0
47
- },
48
- {
49
- "repo_name": "SWE-bench",
50
- "repo_link": "https://github.com/SWE-bench/SWE-bench",
51
- "category": "benchmark",
52
- "github_about_section": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
53
- "homepage_link": "https://swebench.com",
54
- "github_topic_closest_fit": "benchmark",
55
- "contributors_all": 66,
56
- "contributors_2025": 33,
57
- "contributors_2024": 37,
58
- "contributors_2023": 9
59
- },
60
- {
61
- "repo_name": "terminal-bench",
62
- "repo_link": "https://github.com/laude-institute/terminal-bench",
63
- "category": "benchmark",
64
- "github_about_section": "A benchmark for LLMs on complicated tasks in the terminal",
65
- "homepage_link": "https://tbench.ai",
66
- "github_topic_closest_fit": "benchmark",
67
- "contributors_all": 96,
68
- "contributors_2025": 96,
69
- "contributors_2024": 0,
70
- "contributors_2023": 0
71
- },
72
- {
73
- "repo_name": "TritonBench",
74
- "repo_link": "https://github.com/thunlp/TritonBench",
75
- "category": "benchmark",
76
- "github_about_section": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
77
- "homepage_link": "https://arxiv.org/abs/2502.14752",
78
- "github_topic_closest_fit": "benchmark",
79
- "contributors_all": 3,
80
- "contributors_2025": 3,
81
- "contributors_2024": 0,
82
- "contributors_2023": 0
83
- },
84
- {
85
- "repo_name": "BitBLAS",
86
- "repo_link": "https://github.com/microsoft/BitBLAS",
87
- "category": "Basic Linear Algebra Subprograms (BLAS)",
88
- "github_about_section": "BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.",
89
- "github_topic_closest_fit": "matrix-multiplication",
90
- "contributors_all": 17,
91
- "contributors_2025": 5,
92
- "contributors_2024": 14,
93
- "contributors_2023": 0
94
- },
95
- {
96
- "repo_name": "hipBLAS",
97
- "repo_link": "https://github.com/ROCm/hipBLAS",
98
- "category": "Basic Linear Algebra Subprograms (BLAS)",
99
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
100
- "github_topic_closest_fit": "matrix-multiplication",
101
- "contributors_all": 72,
102
- "contributors_2025": 21,
103
- "contributors_2024": 24,
104
- "contributors_2023": 14
105
- },
106
- {
107
- "repo_name": "hipBLASLt",
108
- "repo_link": "https://github.com/AMD-AGI/hipBLASLt",
109
- "category": "Basic Linear Algebra Subprograms (BLAS)",
110
- "github_about_section": "hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditional BLAS library",
111
- "homepage_link": "https://rocm.docs.amd.com/projects/hipBLASLt",
112
- "github_topic_closest_fit": "matrix-multiplication",
113
- "contributors_all": 111,
114
- "contributors_2025": 69,
115
- "contributors_2024": 70,
116
- "contributors_2023": 35
117
- },
118
- {
119
- "repo_name": "AdaptiveCpp",
120
- "repo_link": "https://github.com/AdaptiveCpp/AdaptiveCpp",
121
- "github_about_section": "Compiler for multiple programming models (SYCL, C++ standard parallelism, HIP/CUDA) for CPUs and GPUs from all vendors: The independent, community-driven compiler for C++-based heterogeneous programming models. Lets applications adapt themselves to all the hardware in the system - even at runtime!",
122
- "homepage_link": "https://adaptivecpp.github.io",
123
- "contributors_all": 93,
124
- "contributors_2025": 32,
125
- "contributors_2024": 32,
126
- "contributors_2023": 24
127
- },
128
  {
129
  "repo_name": "llvm-project",
130
  "repo_link": "https://github.com/llvm/llvm-project",
@@ -133,164 +7,9 @@
133
  "homepage_link": "http://llvm.org",
134
  "github_topic_closest_fit": "compiler",
135
  "contributors_all": 6680,
136
- "contributors_2025": 2378,
137
- "contributors_2024": 2130,
138
- "contributors_2023": 1920
139
- },
140
- {
141
- "repo_name": "numba",
142
- "repo_link": "https://github.com/numba/numba",
143
- "github_about_section": "NumPy aware dynamic Python compiler using LLVM",
144
- "homepage_link": "https://numba.pydata.org",
145
- "contributors_all": 430,
146
- "contributors_2025": 36,
147
- "contributors_2024": 32,
148
- "contributors_2023": 55
149
- },
150
- {
151
- "repo_name": "nvcc4jupyter",
152
- "repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
153
- "github_about_section": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
154
- "homepage_link": "https://nvcc4jupyter.readthedocs.io",
155
- "contributors_all": 9,
156
- "contributors_2025": 0,
157
- "contributors_2024": 3,
158
- "contributors_2023": 3
159
- },
160
- {
161
- "repo_name": "CU2CL",
162
- "repo_link": "https://github.com/vtsynergy/CU2CL",
163
- "github_about_section": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
164
- "homepage_link": "http://chrec.cs.vt.edu/cu2cl",
165
- "github_topic_closest_fit": "parallel-programming",
166
- "contributors_all": 3,
167
- "contributors_2025": 0,
168
- "contributors_2024": 0,
169
- "contributors_2023": 0
170
- },
171
- {
172
- "repo_name": "cuda-python",
173
- "repo_link": "https://github.com/NVIDIA/cuda-python",
174
- "github_about_section": "CUDA Python: Performance meets Productivity",
175
- "homepage_link": "https://nvidia.github.io/cuda-python",
176
- "github_topic_closest_fit": "parallel-programming",
177
- "contributors_all": 48,
178
- "contributors_2025": 41,
179
- "contributors_2024": 12,
180
- "contributors_2023": 1
181
- },
182
- {
183
- "repo_name": "OpenCL-SDK",
184
- "repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
185
- "github_about_section": "OpenCL SDK",
186
- "homepage_link": "https://khronos.org/opencl",
187
- "github_topic_closest_fit": "parallel-programming",
188
- "contributors_all": 25,
189
- "contributors_2025": 8,
190
- "contributors_2024": 6,
191
- "contributors_2023": 9
192
- },
193
- {
194
- "repo_name": "pocl",
195
- "repo_link": "https://github.com/pocl/pocl",
196
- "github_about_section": "pocl - Portable Computing Language",
197
- "homepage_link": "https://portablecl.org",
198
- "github_topic_closest_fit": "parallel-programming",
199
- "contributors_all": 166,
200
- "contributors_2025": 26,
201
- "contributors_2024": 27,
202
- "contributors_2023": 21
203
- },
204
- {
205
- "repo_name": "SYCL-Docs",
206
- "repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
207
- "github_about_section": "SYCL Open Source Specification",
208
- "homepage_link": "https://khronos.org/sycl",
209
- "github_topic_closest_fit": "parallel-programming",
210
- "contributors_all": 67,
211
- "contributors_2025": 13,
212
- "contributors_2024": 20,
213
- "contributors_2023": 27
214
- },
215
- {
216
- "repo_name": "triSYCL",
217
- "repo_link": "https://github.com/triSYCL/triSYCL",
218
- "github_about_section": "Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
219
- "homepage_link": "https://trisycl.github.io/triSYCL/Doxygen/triSYCL/html/index.html",
220
- "github_topic_closest_fit": "parallel-programming",
221
- "contributors_all": 31,
222
- "contributors_2025": 0,
223
- "contributors_2024": 1,
224
- "contributors_2023": 3
225
- },
226
- {
227
- "repo_name": "ZLUDA",
228
- "repo_link": "https://github.com/vosen/ZLUDA",
229
- "github_about_section": "CUDA on non-NVIDIA GPUs",
230
- "homepage_link": "https://vosen.github.io/ZLUDA",
231
- "github_topic_closest_fit": "parallel-programming",
232
- "contributors_all": 15,
233
- "contributors_2025": 8,
234
- "contributors_2024": 4,
235
- "contributors_2023": 0
236
- },
237
- {
238
- "repo_name": "llama.cpp",
239
- "repo_link": "https://github.com/ggml-org/llama.cpp",
240
- "category": "inference engine",
241
- "github_about_section": "LLM inference in C/C++",
242
- "homepage_link": "https://ggml.ai",
243
- "github_topic_closest_fit": "inference",
244
- "contributors_all": 1374,
245
- "contributors_2025": 535,
246
- "contributors_2024": 575,
247
- "contributors_2023": 461
248
- },
249
- {
250
- "repo_name": "mistral-inference",
251
- "repo_link": "https://github.com/mistralai/mistral-inference",
252
- "category": "inference engine",
253
- "github_about_section": "Official inference library for Mistral models",
254
- "homepage_link": "https://mistral.ai",
255
- "github_topic_closest_fit": "inference",
256
- "contributors_all": 29,
257
- "contributors_2025": 2,
258
- "contributors_2024": 17,
259
- "contributors_2023": 14
260
- },
261
- {
262
- "repo_name": "ollama",
263
- "repo_link": "https://github.com/ollama/ollama",
264
- "category": "inference engine",
265
- "github_about_section": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
266
- "homepage_link": "https://ollama.com",
267
- "github_topic_closest_fit": "inference",
268
- "contributors_all": 574,
269
- "contributors_2025": 202,
270
- "contributors_2024": 314,
271
- "contributors_2023": 97
272
- },
273
- {
274
- "repo_name": "sglang",
275
- "repo_link": "https://github.com/sgl-project/sglang",
276
- "category": "inference engine",
277
- "github_about_section": "SGLang is a fast serving framework for large language models and vision language models.",
278
- "homepage_link": "https://docs.sglang.ai",
279
- "github_topic_closest_fit": "inference",
280
- "contributors_all": 937,
281
- "contributors_2025": 796,
282
- "contributors_2024": 189,
283
- "contributors_2023": 1
284
- },
285
- {
286
- "repo_name": "TensorRT",
287
- "repo_link": "https://github.com/NVIDIA/TensorRT",
288
- "github_about_section": "NVIDIA TensorRT is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
289
- "homepage_link": "https://developer.nvidia.com/tensorrt",
290
- "contributors_all": 104,
291
- "contributors_2025": 10,
292
- "contributors_2024": 18,
293
- "contributors_2023": 19
294
  },
295
  {
296
  "repo_name": "vllm",
@@ -304,60 +23,6 @@
304
  "contributors_2024": 579,
305
  "contributors_2023": 145
306
  },
307
- {
308
- "repo_name": "kernels",
309
- "repo_link": "https://github.com/huggingface/kernels",
310
- "category": "gpu kernels",
311
- "github_about_section": "Load compute kernels from the Hub",
312
- "contributors_all": 15,
313
- "contributors_2025": 14,
314
- "contributors_2024": 2,
315
- "contributors_2023": 0
316
- },
317
- {
318
- "repo_name": "kernels-community",
319
- "repo_link": "https://github.com/huggingface/kernels-community",
320
- "category": "gpu kernels",
321
- "homepage_link": "https://huggingface.co/kernels-community",
322
- "github_about_section": "Kernel sources for https://huggingface.co/kernels-community",
323
- "contributors_all": 9,
324
- "contributors_2025": 9,
325
- "contributors_2024": 0,
326
- "contributors_2023": 0
327
- },
328
- {
329
- "repo_name": "Liger-Kernel",
330
- "repo_link": "https://github.com/linkedin/Liger-Kernel",
331
- "category": "kernel examples",
332
- "github_about_section": "Efficient Triton Kernels for LLM Training",
333
- "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
334
- "github_topic_closest_fit": "triton",
335
- "contributors_all": 120,
336
- "contributors_2025": 78,
337
- "contributors_2024": 61,
338
- "contributors_2023": 0
339
- },
340
- {
341
- "repo_name": "quack",
342
- "repo_link": "https://github.com/Dao-AILab/quack",
343
- "category": "kernel examples",
344
- "github_about_section": "A Quirky Assortment of CuTe Kernels",
345
- "contributors_all": 17,
346
- "contributors_2025": 17,
347
- "contributors_2024": 0,
348
- "contributors_2023": 0
349
- },
350
- {
351
- "repo_name": "reference-kernels",
352
- "repo_link": "https://github.com/gpu-mode/reference-kernels",
353
- "category": "kernel examples",
354
- "github_about_section": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
355
- "homepage_link": "https://gpumode.com",
356
- "contributors_all": 16,
357
- "contributors_2025": 16,
358
- "contributors_2024": 0,
359
- "contributors_2023": 0
360
- },
361
  {
362
  "repo_name": "pytorch",
363
  "repo_link": "https://github.com/pytorch/pytorch",
@@ -371,167 +36,138 @@
371
  "contributors_2023": 1024
372
  },
373
  {
374
- "repo_name": "tensorflow",
375
- "repo_link": "https://github.com/tensorflow/tensorflow",
376
- "category": "machine learning framework",
377
- "github_about_section": "An Open Source Machine Learning Framework for Everyone",
378
- "homepage_link": "https://tensorflow.org",
379
- "github_topic_closest_fit": "machine-learning",
380
- "contributors_all": 4618,
381
- "contributors_2025": 500,
382
- "contributors_2024": 523,
383
- "contributors_2023": 630
384
- },
385
- {
386
- "repo_name": "torchdendrite",
387
- "repo_link": "https://github.com/sandialabs/torchdendrite",
388
- "category": "machine learning framework",
389
- "github_about_section": "Dendrites for PyTorch and SNNTorch neural networks",
390
- "contributors_all": 2,
391
- "contributors_2025": 1,
392
- "contributors_2024": 1,
393
- "contributors_2023": 0
394
- },
395
- {
396
- "repo_name": "onnx",
397
- "repo_link": "https://github.com/onnx/onnx",
398
- "category": "machine learning interoperability",
399
- "github_about_section": "Open standard for machine learning interoperability",
400
- "homepage_link": "https://onnx.ai",
401
- "github_topic_closest_fit": "onnx",
402
- "contributors_all": 370,
403
- "contributors_2025": 56,
404
- "contributors_2024": 45,
405
- "contributors_2023": 61
406
  },
407
  {
408
- "repo_name": "executorch",
409
- "repo_link": "https://github.com/pytorch/executorch",
410
- "category": "model compiler",
411
- "github_about_section": "On-device AI across mobile, embedded and edge for PyTorch",
412
- "homepage_link": "https://executorch.ai",
413
- "contributors_all": 437,
414
- "contributors_2025": 267,
415
- "contributors_2024": 243,
416
- "contributors_2023": 77
 
417
  },
418
  {
419
- "repo_name": "cutlass",
420
- "repo_link": "https://github.com/NVIDIA/cutlass",
421
- "category": "parallel computing",
422
- "github_about_section": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
423
- "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
424
- "github_topic_closest_fit": "parallel-programming",
425
- "contributors_all": 238,
426
- "contributors_2025": 94,
427
- "contributors_2024": 64,
428
- "contributors_2023": 66
429
  },
430
  {
431
- "repo_name": "ThunderKittens",
432
- "repo_link": "https://github.com/HazyResearch/ThunderKittens",
433
- "category": "parallel computing",
434
- "github_about_section": "Tile primitives for speedy kernels",
435
- "homepage_link": "https://hazyresearch.stanford.edu/blog/2024-10-29-tk2",
436
- "github_topic_closest_fit": "parallel-programming",
437
- "contributors_all": 34,
438
- "contributors_2025": 29,
439
- "contributors_2024": 13,
440
- "contributors_2023": 0
441
  },
442
  {
443
- "repo_name": "helion",
444
- "repo_link": "https://github.com/pytorch/helion",
445
- "category": "parallel computing dsl",
446
- "github_about_section": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
447
- "homepage_link": "https://helionlang.com",
448
- "github_topic_closest_fit": "parallel-programming",
449
- "contributors_all": 49,
450
- "contributors_2025": 49,
451
- "contributors_2024": 0,
452
- "contributors_2023": 0
453
  },
454
  {
455
- "repo_name": "TileIR",
456
- "repo_link": "https://github.com/microsoft/TileIR",
457
- "category": "parallel computing dsl",
458
- "github_about_section": "TileIR (tile-ir) is a concise domain-specific IR designed to streamline the development of high-performance GPU/CPU kernels (e.g., GEMM, Dequant GEMM, FlashAttention, LinearAttention). By employing a Pythonic syntax with an underlying compiler infrastructure on top of TVM, TileIR allows developers to focus on productivity without sacrificing the low-level optimizations necessary for state-of-the-art performance.",
459
- "github_topic_closest_fit": "parallel-programming",
460
- "contributors_all": 10,
461
- "contributors_2025": 10,
462
- "contributors_2024": 1,
463
- "contributors_2023": 0
 
464
  },
465
  {
466
- "repo_name": "tilelang",
467
- "repo_link": "https://github.com/tile-ai/tilelang",
468
- "category": "parallel computing dsl",
469
- "github_about_section": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
470
- "homepage_link": "https://tilelang.com",
471
- "github_topic_closest_fit": "parallel-programming",
472
- "contributors_all": 90,
473
- "contributors_2025": 89,
474
- "contributors_2024": 1,
475
  "contributors_2023": 0
476
  },
477
  {
478
- "repo_name": "triton",
479
- "repo_link": "https://github.com/triton-lang/triton",
480
- "category": "parallel computing dsl",
481
- "github_about_section": "Development repository for the Triton language and compiler",
482
- "homepage_link": "https://triton-lang.org",
483
- "github_topic_closest_fit": "parallel-programming",
484
- "contributors_all": 522,
485
- "contributors_2025": 233,
486
- "contributors_2024": 206,
487
- "contributors_2023": 159
488
  },
489
  {
490
- "repo_name": "cupti",
491
- "repo_link": "https://github.com/cwpearson/cupti",
492
- "category": "performance testing",
493
- "github_about_section": "Profile how CUDA applications create and modify data in memory.",
494
- "github_topic_closest_fit": "profiling",
495
- "contributors_all": 1,
496
- "contributors_2025": 0,
497
- "contributors_2024": 0,
498
- "contributors_2023": 0
499
  },
500
  {
501
- "repo_name": "hatchet",
502
- "repo_link": "https://github.com/LLNL/hatchet",
503
- "category": "performance testing",
504
- "github_about_section": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
505
- "homepage_link": "https://llnl-hatchet.readthedocs.io",
506
- "github_topic_closest_fit": "profiling",
507
- "contributors_all": 25,
508
- "contributors_2025": 3,
509
- "contributors_2024": 6,
510
- "contributors_2023": 8
511
  },
512
  {
513
- "repo_name": "intelliperf",
514
- "repo_link": "https://github.com/AMDResearch/intelliperf",
515
- "category": "performance testing",
516
- "github_about_section": "Automated bottleneck detection and solution orchestration",
517
- "homepage_link": "https://arxiv.org/html/2508.20258v1",
518
- "github_topic_closest_fit": "profiling",
519
- "contributors_all": 7,
520
- "contributors_2025": 7,
521
- "contributors_2024": 0,
522
  "contributors_2023": 0
523
  },
524
  {
525
- "repo_name": "omnitrace",
526
- "repo_link": "https://github.com/ROCm/omnitrace",
527
- "category": "performance testing",
528
- "github_about_section": "Omnitrace: Application Profiling, Tracing, and Analysis",
529
- "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace",
530
- "github_topic_closest_fit": "profiling",
531
- "contributors_all": 16,
532
- "contributors_2025": 2,
533
- "contributors_2024": 12,
534
- "contributors_2023": 2
535
  },
536
  {
537
  "repo_name": "jax",
@@ -545,6 +181,29 @@
545
  "contributors_2024": 280,
546
  "contributors_2023": 202
547
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
548
  {
549
  "repo_name": "numpy",
550
  "repo_link": "https://github.com/numpy/numpy",
@@ -557,6 +216,30 @@
557
  "contributors_2024": 233,
558
  "contributors_2023": 252
559
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
560
  {
561
  "repo_name": "scipy",
562
  "repo_link": "https://github.com/scipy/scipy",
@@ -570,50 +253,38 @@
570
  "contributors_2023": 245
571
  },
572
  {
573
- "repo_name": "elasticsearch",
574
- "repo_link": "https://github.com/elastic/elasticsearch",
575
- "category": "search engine",
576
- "github_about_section": "Free and Open Source, Distributed, RESTful Search Engine",
577
- "homepage_link": "https://elastic.co/products/elasticsearch",
578
- "github_topic_closest_fit": "search-engine",
579
- "contributors_all": 2297,
580
- "contributors_2025": 316,
581
- "contributors_2024": 284,
582
- "contributors_2023": 270
583
- },
584
- {
585
- "repo_name": "jupyterlab",
586
- "repo_link": "https://github.com/jupyterlab/jupyterlab",
587
- "category": "user interface",
588
- "github_about_section": "JupyterLab computational environment.",
589
- "homepage_link": "https://jupyterlab.readthedocs.io",
590
- "github_topic_closest_fit": "jupyter",
591
- "contributors_all": 698,
592
- "contributors_2025": 77,
593
- "contributors_2024": 85,
594
- "contributors_2023": 100
595
  },
596
  {
597
- "repo_name": "milvus",
598
- "repo_link": "https://github.com/milvus-io/milvus",
599
- "category": "vector database",
600
- "github_about_section": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
601
- "homepage_link": "https://milvus.io",
602
- "github_topic_closest_fit": "vector-search",
603
- "contributors_all": 387,
604
- "contributors_2025": 95,
605
- "contributors_2024": 84,
606
- "contributors_2023": 72
607
  },
608
  {
609
- "repo_name": "accelerate",
610
- "repo_link": "https://github.com/huggingface/accelerate",
611
- "github_about_section": "A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support.",
612
- "homepage_link": "https://huggingface.co/docs/accelerate",
613
- "contributors_all": 392,
614
- "contributors_2025": 97,
615
- "contributors_2024": 124,
616
- "contributors_2023": 149
 
 
617
  },
618
  {
619
  "repo_name": "aiter",
@@ -625,6 +296,47 @@
625
  "contributors_2024": 10,
626
  "contributors_2023": 0
627
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
628
  {
629
  "repo_name": "ao",
630
  "repo_link": "https://github.com/pytorch/ao",
@@ -637,24 +349,16 @@
637
  "contributors_2023": 5
638
  },
639
  {
640
- "repo_name": "burn",
641
- "repo_link": "https://github.com/tracel-ai/burn",
642
- "github_about_section": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
643
- "homepage_link": "https://burn.dev",
644
- "contributors_all": 237,
645
- "contributors_2025": 99,
646
- "contributors_2024": 104,
647
- "contributors_2023": 62
648
- },
649
- {
650
- "repo_name": "ccache",
651
- "repo_link": "https://github.com/ccache/ccache",
652
- "github_about_section": "ccache - a fast compiler cache",
653
- "homepage_link": "https://ccache.dev",
654
- "contributors_all": 218,
655
- "contributors_2025": 20,
656
- "contributors_2024": 28,
657
- "contributors_2023": 22
658
  },
659
  {
660
  "repo_name": "ComfyUI",
@@ -669,38 +373,47 @@
669
  "contributors_2023": 94
670
  },
671
  {
672
- "repo_name": "composable_kernel",
673
- "repo_link": "https://github.com/ROCm/composable_kernel",
674
- "category": "gpu kernels",
675
- "github_about_section": "Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators",
676
- "homepage_link": "https://rocm.docs.amd.com/projects/composable_kernel",
677
- "contributors_all": 190,
678
- "contributors_2025": 140,
679
- "contributors_2024": 58,
680
- "contributors_2023": 33
 
 
 
 
 
 
 
 
 
 
 
681
  },
682
  {
683
- "repo_name": "cudnn-frontend",
684
- "repo_link": "https://github.com/NVIDIA/cudnn-frontend",
685
- "category": "parallel computing",
686
- "github_about_section": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
687
- "homepage_link": "https://developer.nvidia.com/cudnn",
688
- "github_topic_closest_fit": "parallel-programming",
689
- "contributors_all": 12,
690
- "contributors_2025": 6,
691
- "contributors_2024": 5,
692
- "contributors_2023": 1
693
  },
694
  {
695
- "repo_name": "cuJSON",
696
- "repo_link": "https://github.com/AutomataLab/cuJSON",
697
- "category": "library leveraging parallel compute",
698
- "github_about_section": "cuJSON: A Highly Parallel JSON Parser for GPUs",
699
- "homepage_link": "https://dl.acm.org/doi/10.1145/3760250.3762222",
700
- "github_topic_closest_fit": "json-parser",
701
- "contributors_all": 2,
702
- "contributors_2025": 2,
703
- "contributors_2024": 2,
704
  "contributors_2023": 0
705
  },
706
  {
@@ -714,129 +427,148 @@
714
  "contributors_2023": 165
715
  },
716
  {
717
- "repo_name": "dstack",
718
- "repo_link": "https://github.com/dstackai/dstack",
719
- "category": "gpu provisioning and orchestration",
720
- "github_about_section": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs-across hyperscalers, neoclouds, or on-prem.",
721
- "homepage_link": "https://dstack.ai",
722
- "github_topic_closest_fit": "orchestration",
723
- "contributors_all": 69,
724
- "contributors_2025": 28,
725
- "contributors_2024": 42,
726
- "contributors_2023": 14
727
  },
728
  {
729
- "repo_name": "flashinfer",
730
- "repo_link": "https://github.com/flashinfer-ai/flashinfer",
731
- "category": "gpu kernels",
732
- "github_about_section": "FlashInfer: Kernel Library for LLM Serving",
733
- "homepage_link": "https://flashinfer.ai",
734
- "github_topic_closest_fit": "attention",
735
- "contributors_all": 205,
736
- "contributors_2025": 158,
737
- "contributors_2024": 50,
738
- "contributors_2023": 11
739
  },
740
  {
741
- "repo_name": "FTorch",
742
- "repo_link": "https://github.com/Cambridge-ICCS/FTorch",
743
- "category": "wrapper",
744
- "github_about_section": "A library for directly calling PyTorch ML models from Fortran.",
745
- "homepage_link": "https://cambridge-iccs.github.io/FTorch",
746
- "github_topic_closest_fit": "machine-learning",
747
- "contributors_all": 20,
748
- "contributors_2025": 11,
749
- "contributors_2024": 8,
750
- "contributors_2023": 9
751
  },
752
  {
753
- "repo_name": "GEAK-agent",
754
- "repo_link": "https://github.com/AMD-AGI/GEAK-agent",
755
- "category": "agent",
756
- "github_about_section": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
757
- "github_topic_closest_fit": "ai-agents",
758
- "contributors_all": 9,
759
- "contributors_2025": 9,
760
  "contributors_2024": 0,
761
  "contributors_2023": 0
762
  },
763
  {
764
- "repo_name": "hhvm",
765
- "repo_link": "https://github.com/facebook/hhvm",
766
- "github_about_section": "A virtual machine for executing programs written in Hack.",
767
- "homepage_link": "https://hhvm.com",
768
- "contributors_all": 2624,
769
- "contributors_2025": 692,
770
- "contributors_2024": 648,
771
- "contributors_2023": 604
 
 
772
  },
773
  {
774
- "repo_name": "hip",
775
- "repo_link": "https://github.com/ROCm/hip",
776
- "github_about_section": "HIP: C++ Heterogeneous-Compute Interface for Portability",
777
- "homepage_link": "https://rocmdocs.amd.com/projects/HIP",
778
- "contributors_all": 288,
779
- "contributors_2025": 46,
780
- "contributors_2024": 31,
781
- "contributors_2023": 25
782
  },
783
  {
784
- "repo_name": "hipCUB",
785
- "repo_link": "https://github.com/ROCm/hipCUB",
786
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
787
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
788
- "contributors_all": 54,
789
- "contributors_2025": 10,
790
- "contributors_2024": 19,
791
- "contributors_2023": 13
 
 
792
  },
793
  {
794
- "repo_name": "IMO2025",
795
- "repo_link": "https://github.com/harmonic-ai/IMO2025",
796
- "category": "formal mathematical reasoning",
797
- "github_about_section": "Harmonic's model Aristotle achieved gold medal performance, solving 5 problems. This repository contains the lean statement files and proofs for Problems 1-5.",
798
- "homepage_link": "https://harmonic.fun",
799
- "github_topic_closest_fit": "lean",
800
- "contributors_all": 2,
801
- "contributors_2025": 2,
802
- "contributors_2024": 0,
803
- "contributors_2023": 0
804
  },
805
  {
806
- "repo_name": "kubernetes",
807
- "repo_link": "https://github.com/kubernetes/kubernetes",
808
- "category": "container orchestration",
809
- "github_about_section": "Production-Grade Container Scheduling and Management",
810
- "homepage_link": "https://kubernetes.io",
811
- "github_topic_closest_fit": "kubernetes",
812
- "contributors_all": 5041,
813
- "contributors_2025": 509,
814
- "contributors_2024": 498,
815
- "contributors_2023": 565
816
  },
817
  {
818
- "repo_name": "lapack",
819
- "repo_link": "https://github.com/Reference-LAPACK/lapack",
820
- "category": "linear algebra",
821
- "github_about_section": "LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.",
822
- "homepage_link": "https://netlib.org/lapack",
823
- "github_topic_closest_fit": "linear-algebra",
824
- "contributors_all": 178,
825
- "contributors_2025": 20,
826
- "contributors_2024": 24,
827
- "contributors_2023": 42
828
  },
829
  {
830
- "repo_name": "lean4",
831
- "repo_link": "https://github.com/leanprover/lean4",
832
- "category": "theorem prover",
833
- "github_about_section": "Lean 4 programming language and theorem prover",
834
- "homepage_link": "https://lean-lang.org",
835
- "github_topic_closest_fit": "lean",
836
- "contributors_all": 278,
837
- "contributors_2025": 110,
838
- "contributors_2024": 85,
839
- "contributors_2023": 64
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
840
  },
841
  {
842
  "repo_name": "letta",
@@ -850,6 +582,48 @@
850
  "contributors_2024": 75,
851
  "contributors_2023": 47
852
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
853
  {
854
  "repo_name": "lightning-thunder",
855
  "repo_link": "https://github.com/Lightning-AI/lightning-thunder",
@@ -860,25 +634,61 @@
860
  "contributors_2023": 29
861
  },
862
  {
863
- "repo_name": "LMCache",
864
- "repo_link": "https://github.com/LMCache/LMCache",
865
- "github_about_section": "Supercharge Your LLM with the Fastest KV Cache Layer",
866
- "homepage_link": "https://lmcache.ai",
867
- "contributors_all": 152,
868
- "contributors_2025": 144,
869
- "contributors_2024": 18,
870
- "contributors_2023": 0
 
 
871
  },
872
  {
873
- "repo_name": "mcp-agent",
874
- "repo_link": "https://github.com/lastmile-ai/mcp-agent",
875
- "category": "mcp",
876
- "github_about_section": "Build effective agents using Model Context Protocol and simple workflow patterns",
877
- "github_topic_closest_fit": "mcp",
878
- "contributors_all": 63,
879
- "contributors_2025": 63,
880
- "contributors_2024": 1,
881
- "contributors_2023": 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
882
  },
883
  {
884
  "repo_name": "metaflow",
@@ -891,109 +701,212 @@
891
  "contributors_2023": 28
892
  },
893
  {
894
- "repo_name": "MIOpen",
895
- "repo_link": "https://github.com/ROCm/MIOpen",
896
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
897
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
898
- "contributors_all": 204,
899
- "contributors_2025": 47,
900
- "contributors_2024": 62,
901
- "contributors_2023": 44
902
  },
903
  {
904
- "repo_name": "modelcontextprotocol",
905
- "repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
906
- "category": "mcp",
907
- "github_about_section": "Specification and documentation for the Model Context Protocol",
908
- "homepage_link": "https://modelcontextprotocol.io",
909
- "github_topic_closest_fit": "mcp",
910
- "contributors_all": 327,
911
- "contributors_2025": 298,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
912
  "contributors_2024": 42,
 
 
 
 
 
 
 
 
 
 
 
913
  "contributors_2023": 0
914
  },
915
  {
916
- "repo_name": "modular",
917
- "repo_link": "https://github.com/modular/modular",
918
- "category": "parallel computing",
919
- "github_about_section": "The Modular Platform (includes MAX & Mojo)",
920
- "homepage_link": "https://docs.modular.com",
921
- "github_topic_closest_fit": "parallel-programming",
922
- "contributors_all": 366,
923
- "contributors_2025": 222,
924
- "contributors_2024": 205,
925
- "contributors_2023": 99
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
926
  },
927
  {
928
- "repo_name": "monarch",
929
- "repo_link": "https://github.com/meta-pytorch/monarch",
930
- "github_about_section": "PyTorch Single Controller",
931
- "homepage_link": "https://meta-pytorch.org/monarch",
932
- "contributors_all": 85,
933
- "contributors_2025": 85,
934
- "contributors_2024": 0,
935
- "contributors_2023": 0
936
  },
937
  {
938
- "repo_name": "Mooncake",
939
- "repo_link": "https://github.com/kvcache-ai/Mooncake",
940
- "github_about_section": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
941
- "homepage_link": "https://kvcache-ai.github.io/Mooncake",
942
- "github_topic_closest_fit": "inference",
943
- "contributors_all": 138,
944
- "contributors_2025": 133,
945
- "contributors_2024": 13,
946
- "contributors_2023": 0
 
947
  },
948
  {
949
- "repo_name": "nccl",
950
- "repo_link": "https://github.com/NVIDIA/nccl",
951
- "github_about_section": "Optimized primitives for collective multi-GPU communication",
952
- "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
953
- "contributors_all": 51,
954
- "contributors_2025": 7,
955
- "contributors_2024": 5,
956
- "contributors_2023": 6
957
  },
958
  {
959
- "repo_name": "neuronx-distributed-inference",
960
- "repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
961
- "contributors_all": 11,
962
- "contributors_2025": 9,
963
- "contributors_2024": 3,
964
- "contributors_2023": 0
 
965
  },
966
  {
967
- "repo_name": "nixl",
968
- "repo_link": "https://github.com/ai-dynamo/nixl",
969
- "github_about_section": "NVIDIA Inference Xfer Library (NIXL)",
970
- "contributors_all": 78,
971
- "contributors_2025": 78,
972
- "contributors_2024": 0,
973
- "contributors_2023": 0
 
 
 
974
  },
975
  {
976
- "repo_name": "ome",
977
- "repo_link": "https://github.com/sgl-project/ome",
978
- "github_about_section": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
979
- "homepage_link": "http://docs.sglang.ai/ome",
980
- "github_topic_closest_fit": "k8s",
981
- "contributors_all": 28,
982
- "contributors_2025": 28,
983
  "contributors_2024": 0,
984
  "contributors_2023": 0
985
  },
986
- {
987
- "repo_name": "ondemand",
988
- "repo_link": "https://github.com/OSC/ondemand",
989
- "github_about_section": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
990
- "homepage_link": "https://openondemand.org",
991
- "github_topic_closest_fit": "hpc",
992
- "contributors_all": 117,
993
- "contributors_2025": 43,
994
- "contributors_2024": 23,
995
- "contributors_2023": 21
996
- },
997
  {
998
  "repo_name": "oneDPL",
999
  "repo_link": "https://github.com/uxlfoundation/oneDPL",
@@ -1005,102 +918,110 @@
1005
  "contributors_2023": 28
1006
  },
1007
  {
1008
- "repo_name": "openevolve",
1009
- "repo_link": "https://github.com/codelion/openevolve",
1010
- "github_about_section": "Open-source implementation of AlphaEvolve",
1011
- "github_topic_closest_fit": "genetic-algorithm",
1012
- "contributors_all": 46,
1013
- "contributors_2025": 46,
1014
- "contributors_2024": 0,
 
 
 
 
 
 
 
 
 
 
 
 
1015
  "contributors_2023": 0
1016
  },
1017
  {
1018
- "repo_name": "ort",
1019
- "repo_link": "https://github.com/pytorch/ort",
1020
- "github_about_section": "Accelerate PyTorch models with ONNX Runtime",
1021
- "contributors_all": 47,
1022
- "contributors_2025": 0,
1023
- "contributors_2024": 7,
1024
- "contributors_2023": 9
 
 
1025
  },
1026
  {
1027
- "repo_name": "peft",
1028
- "repo_link": "https://github.com/huggingface/peft",
1029
- "github_about_section": "PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
1030
- "homepage_link": "https://huggingface.co/docs/peft",
1031
- "github_topic_closest_fit": "lora",
1032
- "contributors_all": 272,
1033
- "contributors_2025": 69,
1034
- "contributors_2024": 111,
1035
- "contributors_2023": 115
1036
  },
1037
  {
1038
- "repo_name": "Primus-Turbo",
1039
- "repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
1040
- "contributors_all": 12,
1041
- "contributors_2025": 12,
 
 
 
 
1042
  "contributors_2024": 0,
1043
  "contributors_2023": 0
1044
  },
1045
  {
1046
- "repo_name": "pybind11",
1047
- "repo_link": "https://github.com/pybind/pybind11",
1048
- "github_about_section": "Seamless operability between C++11 and Python",
1049
- "homepage_link": "https://pybind11.readthedocs.io",
1050
- "github_topic_closest_fit": "bindings",
1051
- "contributors_all": 404,
1052
- "contributors_2025": 43,
1053
- "contributors_2024": 45,
1054
- "contributors_2023": 42
1055
  },
1056
  {
1057
- "repo_name": "RaBitQ",
1058
- "repo_link": "https://github.com/gaoj0017/RaBitQ",
1059
- "github_about_section": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
1060
- "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
1061
- "github_topic_closest_fit": "nearest-neighbor-search",
1062
- "contributors_all": 2,
1063
- "contributors_2025": 2,
1064
- "contributors_2024": 1,
1065
  "contributors_2023": 0
1066
  },
1067
  {
1068
- "repo_name": "rdma-core",
1069
- "repo_link": "https://github.com/linux-rdma/rdma-core",
1070
- "github_about_section": "RDMA core userspace libraries and daemons",
1071
- "contributors_all": 437,
1072
- "contributors_2025": 58,
1073
- "contributors_2024": 61,
1074
- "contributors_2023": 66
 
 
 
1075
  },
1076
  {
1077
- "repo_name": "rocFFT",
1078
- "repo_link": "https://github.com/ROCm/rocFFT",
1079
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1080
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
1081
- "contributors_all": 81,
1082
- "contributors_2025": 17,
 
1083
  "contributors_2024": 20,
1084
- "contributors_2023": 19
1085
- },
1086
- {
1087
- "repo_name": "ROCm",
1088
- "repo_link": "https://github.com/ROCm/ROCm",
1089
- "github_about_section": "AMD ROCm Software - GitHub Home",
1090
- "homepage_link": "https://rocm.docs.amd.com",
1091
- "contributors_all": 166,
1092
- "contributors_2025": 67,
1093
- "contributors_2024": 61,
1094
- "contributors_2023": 44
1095
  },
1096
  {
1097
- "repo_name": "rocm-systems",
1098
- "repo_link": "https://github.com/ROCm/rocm-systems",
1099
- "github_about_section": "super repo for rocm systems projects",
1100
- "contributors_all": 1032,
1101
- "contributors_2025": 440,
1102
- "contributors_2024": 323,
1103
- "contributors_2023": 204
1104
  },
1105
  {
1106
  "repo_name": "rocPRIM",
@@ -1112,6 +1033,61 @@
1112
  "contributors_2024": 28,
1113
  "contributors_2023": 15
1114
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1115
  {
1116
  "repo_name": "rocRAND",
1117
  "repo_link": "https://github.com/ROCm/rocRAND",
@@ -1123,24 +1099,56 @@
1123
  "contributors_2023": 26
1124
  },
1125
  {
1126
- "repo_name": "rocSOLVER",
1127
- "repo_link": "https://github.com/ROCm/rocSOLVER",
1128
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1129
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
1130
- "contributors_all": 59,
1131
- "contributors_2025": 20,
1132
- "contributors_2024": 23,
1133
- "contributors_2023": 15
 
1134
  },
1135
  {
1136
- "repo_name": "rocSPARSE",
1137
- "repo_link": "https://github.com/ROCm/rocSPARSE",
1138
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1139
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
1140
- "contributors_all": 65,
1141
- "contributors_2025": 19,
1142
- "contributors_2024": 24,
1143
- "contributors_2023": 18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1144
  },
1145
  {
1146
  "repo_name": "roctracer",
@@ -1152,6 +1160,51 @@
1152
  "contributors_2024": 11,
1153
  "contributors_2023": 6
1154
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1155
  {
1156
  "repo_name": "Self-Forcing",
1157
  "repo_link": "https://github.com/guandeh17/Self-Forcing",
@@ -1165,38 +1218,28 @@
1165
  "contributors_2023": 0
1166
  },
1167
  {
1168
- "repo_name": "server",
1169
- "repo_link": "https://github.com/triton-inference-server/server",
1170
- "github_about_section": "The Triton Inference Server provides an optimized cloud and edge inferencing solution.",
1171
- "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
1172
- "github_topic_closest_fit": "inference",
1173
- "contributors_all": 147,
1174
- "contributors_2025": 24,
1175
- "contributors_2024": 36,
1176
- "contributors_2023": 34
1177
- },
1178
- {
1179
- "repo_name": "spark",
1180
- "repo_link": "https://github.com/apache/spark",
1181
- "github_about_section": "Apache Spark - A unified analytics engine for large-scale data processing",
1182
- "homepage_link": "https://spark.apache.org",
1183
- "github_topic_closest_fit": "big-data",
1184
- "contributors_all": 3083,
1185
- "contributors_2025": 322,
1186
- "contributors_2024": 300,
1187
- "contributors_2023": 336
1188
  },
1189
  {
1190
- "repo_name": "StreamDiffusion",
1191
- "repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
1192
- "category": "image generation",
1193
- "github_about_section": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation",
1194
- "homepage_link": "https://arxiv.org/abs/2312.12491",
1195
- "github_topic_closest_fit": "diffusion-models",
1196
- "contributors_all": 29,
1197
- "contributors_2025": 0,
1198
- "contributors_2024": 9,
1199
- "contributors_2023": 25
1200
  },
1201
  {
1202
  "repo_name": "streamv2v",
@@ -1207,76 +1250,76 @@
1207
  "github_topic_closest_fit": "diffusion-models",
1208
  "contributors_all": 7,
1209
  "contributors_2025": 3,
1210
- "contributors_2024": 6,
1211
- "contributors_2023": 0
1212
- },
1213
- {
1214
- "repo_name": "synthetic-data-kit",
1215
- "repo_link": "https://github.com/meta-llama/synthetic-data-kit",
1216
- "category": "synthetic data generation",
1217
- "github_about_section": "Tool for generating high quality Synthetic datasets",
1218
- "homepage_link": "https://pypi.org/project/synthetic-data-kit",
1219
- "github_topic_closest_fit": "synthetic-dataset-generation",
1220
- "contributors_all": 15,
1221
- "contributors_2025": 15,
1222
- "contributors_2024": 0,
1223
  "contributors_2023": 0
1224
  },
1225
  {
1226
- "repo_name": "Tensile",
1227
- "repo_link": "https://github.com/ROCm/Tensile",
1228
- "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1229
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
1230
- "contributors_all": 137,
1231
- "contributors_2025": 16,
1232
- "contributors_2024": 25,
1233
- "contributors_2023": 22
 
 
1234
  },
1235
  {
1236
- "repo_name": "tflite-micro",
1237
- "repo_link": "https://github.com/tensorflow/tflite-micro",
1238
- "github_about_section": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors).",
1239
- "contributors_all": 111,
1240
- "contributors_2025": 19,
1241
- "contributors_2024": 25,
1242
- "contributors_2023": 31
 
 
 
1243
  },
1244
  {
1245
- "repo_name": "torchdynamo",
1246
- "repo_link": "https://github.com/pytorch/torchdynamo",
1247
- "github_about_section": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster.",
1248
- "contributors_all": 63,
1249
- "contributors_2025": 0,
1250
- "contributors_2024": 1,
1251
- "contributors_2023": 4
 
 
 
1252
  },
1253
  {
1254
- "repo_name": "torchtitan",
1255
- "repo_link": "https://github.com/pytorch/torchtitan",
1256
- "github_about_section": "A PyTorch native platform for training generative AI models",
1257
- "contributors_all": 145,
1258
- "contributors_2025": 119,
1259
- "contributors_2024": 43,
1260
- "contributors_2023": 1
 
 
 
1261
  },
1262
  {
1263
- "repo_name": "transformers",
1264
- "repo_link": "https://github.com/huggingface/transformers",
1265
- "github_about_section": "Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.",
1266
- "homepage_link": "https://huggingface.co/transformers",
1267
- "contributors_all": 3582,
1268
- "contributors_2025": 860,
1269
- "contributors_2024": 769,
1270
- "contributors_2023": 758
 
1271
  },
1272
  {
1273
- "repo_name": "Triton-distributed",
1274
- "repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
1275
- "github_about_section": "Distributed Compiler based on Triton for Parallel Systems",
1276
- "homepage_link": "https://triton-distributed.readthedocs.io",
1277
- "contributors_all": 30,
1278
- "contributors_2025": 30,
1279
- "contributors_2024": 0,
1280
  "contributors_2023": 0
1281
  },
1282
  {
@@ -1290,119 +1333,76 @@
1290
  "contributors_2023": 0
1291
  },
1292
  {
1293
- "repo_name": "tritonparse",
1294
- "repo_link": "https://github.com/meta-pytorch/tritonparse",
1295
- "github_about_section": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
1296
- "homepage_link": "https://meta-pytorch.org/tritonparse",
1297
- "contributors_all": 15,
1298
- "contributors_2025": 15,
1299
- "contributors_2024": 0,
1300
- "contributors_2023": 0
1301
- },
1302
- {
1303
- "repo_name": "trl",
1304
- "repo_link": "https://github.com/huggingface/trl",
1305
- "github_about_section": "Train transformer language models with reinforcement learning.",
1306
- "homepage_link": "http://hf.co/docs/trl",
1307
- "contributors_all": 433,
1308
- "contributors_2025": 189,
1309
- "contributors_2024": 154,
1310
- "contributors_2023": 122
1311
- },
1312
- {
1313
- "repo_name": "truss",
1314
- "repo_link": "https://github.com/basetenlabs/truss",
1315
- "category": "inference engine",
1316
- "github_about_section": "The simplest way to serve AI/ML models in production",
1317
- "homepage_link": "https://truss.baseten.co",
1318
- "github_topic_closest_fit": "inference",
1319
- "contributors_all": 72,
1320
- "contributors_2025": 44,
1321
- "contributors_2024": 30,
1322
- "contributors_2023": 21
1323
- },
1324
- {
1325
- "repo_name": "unsloth",
1326
- "repo_link": "https://github.com/unslothai/unsloth",
1327
- "category": "fine tuning",
1328
- "github_about_section": "Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
1329
- "homepage_link": "https://docs.unsloth.ai",
1330
- "github_topic_closest_fit": "fine-tuning",
1331
- "contributors_all": 127,
1332
- "contributors_2025": 102,
1333
- "contributors_2024": 27,
1334
  "contributors_2023": 3
1335
  },
1336
  {
1337
- "repo_name": "verl",
1338
- "repo_link": "https://github.com/volcengine/verl",
1339
- "category": "reinforcement learning",
1340
- "github_about_section": "verl: Volcano Engine Reinforcement Learning for LLMs",
1341
- "homepage_link": "https://verl.readthedocs.io",
1342
- "github_topic_closest_fit": "deep-reinforcement-learning",
1343
- "contributors_all": 462,
1344
- "contributors_2025": 454,
1345
- "contributors_2024": 10,
1346
  "contributors_2023": 0
1347
  },
1348
  {
1349
- "repo_name": "Vulkan-Hpp",
1350
- "repo_link": "https://github.com/KhronosGroup/Vulkan-Hpp",
1351
- "category": "graphics api",
1352
- "github_about_section": "Open-Source Vulkan C++ API",
1353
- "homepage_link": "https://vulkan.org",
1354
- "github_topic_closest_fit": "vulkan",
1355
- "contributors_all": 102,
1356
- "contributors_2025": 21,
1357
- "contributors_2024": 15,
1358
- "contributors_2023": 15
1359
  },
1360
  {
1361
- "repo_name": "Vulkan-Tools",
1362
- "repo_link": "https://github.com/KhronosGroup/Vulkan-Tools",
1363
- "category": "graphics api",
1364
- "github_about_section": "Vulkan Development Tools",
1365
- "homepage_link": "https://vulkan.org",
1366
- "github_topic_closest_fit": "vulkan",
1367
- "contributors_all": 248,
1368
- "contributors_2025": 20,
1369
- "contributors_2024": 24,
1370
- "contributors_2023": 24
1371
  },
1372
  {
1373
- "repo_name": "Vulkan-Docs",
1374
- "repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
1375
- "category": "graphics api",
1376
- "github_about_section": "The Vulkan API Specification and related tools",
1377
- "homepage_link": "https://vulkan.org",
1378
- "github_topic_closest_fit": "vulkan",
1379
- "contributors_all": 141,
1380
- "contributors_2025": 18,
1381
- "contributors_2024": 21,
1382
- "contributors_2023": 34
1383
  },
1384
  {
1385
- "repo_name": "Wan2.2",
1386
- "repo_link": "https://github.com/Wan-Video/Wan2.2",
1387
- "category": "video generation",
1388
- "github_about_section": "Wan: Open and Advanced Large-Scale Video Generative Models",
1389
- "homepage_link": "https://wan.video",
1390
  "github_topic_closest_fit": "diffusion-models",
1391
- "contributors_all": 14,
1392
- "contributors_2025": 14,
1393
- "contributors_2024": 0,
1394
- "contributors_2023": 0
1395
  },
1396
  {
1397
- "repo_name": "warp",
1398
- "repo_link": "https://github.com/NVIDIA/warp",
1399
- "category": "spatial computing",
1400
- "github_about_section": "A Python framework for accelerated simulation, data generation and spatial computing.",
1401
- "homepage_link": "https://nvidia.github.io/warp",
1402
- "github_topic_closest_fit": "physics-simulation",
1403
- "contributors_all": 79,
1404
- "contributors_2025": 40,
1405
- "contributors_2024": 29,
1406
- "contributors_2023": 17
1407
  }
1408
  ]
 
1
  [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  {
3
  "repo_name": "llvm-project",
4
  "repo_link": "https://github.com/llvm/llvm-project",
 
7
  "homepage_link": "http://llvm.org",
8
  "github_topic_closest_fit": "compiler",
9
  "contributors_all": 6680,
10
+ "contributors_2025": 2378,
11
+ "contributors_2024": 2130,
12
+ "contributors_2023": 1920
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  },
14
  {
15
  "repo_name": "vllm",
 
23
  "contributors_2024": 579,
24
  "contributors_2023": 145
25
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  {
27
  "repo_name": "pytorch",
28
  "repo_link": "https://github.com/pytorch/pytorch",
 
36
  "contributors_2023": 1024
37
  },
38
  {
39
+ "repo_name": "transformers",
40
+ "repo_link": "https://github.com/huggingface/transformers",
41
+ "github_about_section": "Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.",
42
+ "homepage_link": "https://huggingface.co/transformers",
43
+ "contributors_all": 3582,
44
+ "contributors_2025": 860,
45
+ "contributors_2024": 769,
46
+ "contributors_2023": 758
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  },
48
  {
49
+ "repo_name": "sglang",
50
+ "repo_link": "https://github.com/sgl-project/sglang",
51
+ "category": "inference engine",
52
+ "github_about_section": "SGLang is a fast serving framework for large language models and vision language models.",
53
+ "homepage_link": "https://docs.sglang.ai",
54
+ "github_topic_closest_fit": "inference",
55
+ "contributors_all": 937,
56
+ "contributors_2025": 796,
57
+ "contributors_2024": 189,
58
+ "contributors_2023": 1
59
  },
60
  {
61
+ "repo_name": "hhvm",
62
+ "repo_link": "https://github.com/facebook/hhvm",
63
+ "github_about_section": "A virtual machine for executing programs written in Hack.",
64
+ "homepage_link": "https://hhvm.com",
65
+ "contributors_all": 2624,
66
+ "contributors_2025": 692,
67
+ "contributors_2024": 648,
68
+ "contributors_2023": 604
 
 
69
  },
70
  {
71
+ "repo_name": "llama.cpp",
72
+ "repo_link": "https://github.com/ggml-org/llama.cpp",
73
+ "category": "inference engine",
74
+ "github_about_section": "LLM inference in C/C++",
75
+ "homepage_link": "https://ggml.ai",
76
+ "github_topic_closest_fit": "inference",
77
+ "contributors_all": 1374,
78
+ "contributors_2025": 535,
79
+ "contributors_2024": 575,
80
+ "contributors_2023": 461
81
  },
82
  {
83
+ "repo_name": "kubernetes",
84
+ "repo_link": "https://github.com/kubernetes/kubernetes",
85
+ "category": "container orchestration",
86
+ "github_about_section": "Production-Grade Container Scheduling and Management",
87
+ "homepage_link": "https://kubernetes.io",
88
+ "github_topic_closest_fit": "kubernetes",
89
+ "contributors_all": 5041,
90
+ "contributors_2025": 509,
91
+ "contributors_2024": 498,
92
+ "contributors_2023": 565
93
  },
94
  {
95
+ "repo_name": "tensorflow",
96
+ "repo_link": "https://github.com/tensorflow/tensorflow",
97
+ "category": "machine learning framework",
98
+ "github_about_section": "An Open Source Machine Learning Framework for Everyone",
99
+ "homepage_link": "https://tensorflow.org",
100
+ "github_topic_closest_fit": "machine-learning",
101
+ "contributors_all": 4618,
102
+ "contributors_2025": 500,
103
+ "contributors_2024": 523,
104
+ "contributors_2023": 630
105
  },
106
  {
107
+ "repo_name": "verl",
108
+ "repo_link": "https://github.com/volcengine/verl",
109
+ "category": "reinforcement learning",
110
+ "github_about_section": "verl: Volcano Engine Reinforcement Learning for LLMs",
111
+ "homepage_link": "https://verl.readthedocs.io",
112
+ "github_topic_closest_fit": "deep-reinforcement-learning",
113
+ "contributors_all": 462,
114
+ "contributors_2025": 454,
115
+ "contributors_2024": 10,
116
  "contributors_2023": 0
117
  },
118
  {
119
+ "repo_name": "rocm-systems",
120
+ "repo_link": "https://github.com/ROCm/rocm-systems",
121
+ "github_about_section": "super repo for rocm systems projects",
122
+ "contributors_all": 1032,
123
+ "contributors_2025": 440,
124
+ "contributors_2024": 323,
125
+ "contributors_2023": 204
 
 
 
126
  },
127
  {
128
+ "repo_name": "ray",
129
+ "repo_link": "https://github.com/ray-project/ray",
130
+ "github_about_section": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
131
+ "homepage_link": "https://ray.io",
132
+ "contributors_all": 1381,
133
+ "contributors_2025": 397,
134
+ "contributors_2024": 223,
135
+ "contributors_2023": 230
 
136
  },
137
  {
138
+ "repo_name": "spark",
139
+ "repo_link": "https://github.com/apache/spark",
140
+ "github_about_section": "Apache Spark - A unified analytics engine for large-scale data processing",
141
+ "homepage_link": "https://spark.apache.org",
142
+ "github_topic_closest_fit": "big-data",
143
+ "contributors_all": 3083,
144
+ "contributors_2025": 322,
145
+ "contributors_2024": 300,
146
+ "contributors_2023": 336
 
147
  },
148
  {
149
+ "repo_name": "goose",
150
+ "repo_link": "https://github.com/block/goose",
151
+ "category": "agent",
152
+ "github_about_section": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
153
+ "homepage_link": "https://block.github.io/goose",
154
+ "github_topic_closest_fit": "ai-agents",
155
+ "contributors_all": 332,
156
+ "contributors_2025": 319,
157
+ "contributors_2024": 32,
158
  "contributors_2023": 0
159
  },
160
  {
161
+ "repo_name": "elasticsearch",
162
+ "repo_link": "https://github.com/elastic/elasticsearch",
163
+ "category": "search engine",
164
+ "github_about_section": "Free and Open Source, Distributed, RESTful Search Engine",
165
+ "homepage_link": "https://elastic.co/products/elasticsearch",
166
+ "github_topic_closest_fit": "search-engine",
167
+ "contributors_all": 2297,
168
+ "contributors_2025": 316,
169
+ "contributors_2024": 284,
170
+ "contributors_2023": 270
171
  },
172
  {
173
  "repo_name": "jax",
 
181
  "contributors_2024": 280,
182
  "contributors_2023": 202
183
  },
184
+ {
185
+ "repo_name": "modelcontextprotocol",
186
+ "repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
187
+ "category": "mcp",
188
+ "github_about_section": "Specification and documentation for the Model Context Protocol",
189
+ "homepage_link": "https://modelcontextprotocol.io",
190
+ "github_topic_closest_fit": "mcp",
191
+ "contributors_all": 327,
192
+ "contributors_2025": 298,
193
+ "contributors_2024": 42,
194
+ "contributors_2023": 0
195
+ },
196
+ {
197
+ "repo_name": "executorch",
198
+ "repo_link": "https://github.com/pytorch/executorch",
199
+ "category": "model compiler",
200
+ "github_about_section": "On-device AI across mobile, embedded and edge for PyTorch",
201
+ "homepage_link": "https://executorch.ai",
202
+ "contributors_all": 437,
203
+ "contributors_2025": 267,
204
+ "contributors_2024": 243,
205
+ "contributors_2023": 77
206
+ },
207
  {
208
  "repo_name": "numpy",
209
  "repo_link": "https://github.com/numpy/numpy",
 
216
  "contributors_2024": 233,
217
  "contributors_2023": 252
218
  },
219
+ {
220
+ "repo_name": "triton",
221
+ "repo_link": "https://github.com/triton-lang/triton",
222
+ "category": "parallel computing dsl",
223
+ "github_about_section": "Development repository for the Triton language and compiler",
224
+ "homepage_link": "https://triton-lang.org",
225
+ "github_topic_closest_fit": "parallel-programming",
226
+ "contributors_all": 522,
227
+ "contributors_2025": 233,
228
+ "contributors_2024": 206,
229
+ "contributors_2023": 159
230
+ },
231
+ {
232
+ "repo_name": "modular",
233
+ "repo_link": "https://github.com/modular/modular",
234
+ "category": "parallel computing",
235
+ "github_about_section": "The Modular Platform (includes MAX & Mojo)",
236
+ "homepage_link": "https://docs.modular.com",
237
+ "github_topic_closest_fit": "parallel-programming",
238
+ "contributors_all": 366,
239
+ "contributors_2025": 222,
240
+ "contributors_2024": 205,
241
+ "contributors_2023": 99
242
+ },
243
  {
244
  "repo_name": "scipy",
245
  "repo_link": "https://github.com/scipy/scipy",
 
253
  "contributors_2023": 245
254
  },
255
  {
256
+ "repo_name": "ollama",
257
+ "repo_link": "https://github.com/ollama/ollama",
258
+ "category": "inference engine",
259
+ "github_about_section": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
260
+ "homepage_link": "https://ollama.com",
261
+ "github_topic_closest_fit": "inference",
262
+ "contributors_all": 574,
263
+ "contributors_2025": 202,
264
+ "contributors_2024": 314,
265
+ "contributors_2023": 97
 
 
 
 
 
 
 
 
 
 
 
 
266
  },
267
  {
268
+ "repo_name": "trl",
269
+ "repo_link": "https://github.com/huggingface/trl",
270
+ "github_about_section": "Train transformer language models with reinforcement learning.",
271
+ "homepage_link": "http://hf.co/docs/trl",
272
+ "contributors_all": 433,
273
+ "contributors_2025": 189,
274
+ "contributors_2024": 154,
275
+ "contributors_2023": 122
 
 
276
  },
277
  {
278
+ "repo_name": "flashinfer",
279
+ "repo_link": "https://github.com/flashinfer-ai/flashinfer",
280
+ "category": "gpu kernels",
281
+ "github_about_section": "FlashInfer: Kernel Library for LLM Serving",
282
+ "homepage_link": "https://flashinfer.ai",
283
+ "github_topic_closest_fit": "attention",
284
+ "contributors_all": 205,
285
+ "contributors_2025": 158,
286
+ "contributors_2024": 50,
287
+ "contributors_2023": 11
288
  },
289
  {
290
  "repo_name": "aiter",
 
296
  "contributors_2024": 10,
297
  "contributors_2023": 0
298
  },
299
+ {
300
+ "repo_name": "LMCache",
301
+ "repo_link": "https://github.com/LMCache/LMCache",
302
+ "github_about_section": "Supercharge Your LLM with the Fastest KV Cache Layer",
303
+ "homepage_link": "https://lmcache.ai",
304
+ "contributors_all": 152,
305
+ "contributors_2025": 144,
306
+ "contributors_2024": 18,
307
+ "contributors_2023": 0
308
+ },
309
+ {
310
+ "repo_name": "composable_kernel",
311
+ "repo_link": "https://github.com/ROCm/composable_kernel",
312
+ "category": "gpu kernels",
313
+ "github_about_section": "Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators",
314
+ "homepage_link": "https://rocm.docs.amd.com/projects/composable_kernel",
315
+ "contributors_all": 190,
316
+ "contributors_2025": 140,
317
+ "contributors_2024": 58,
318
+ "contributors_2023": 33
319
+ },
320
+ {
321
+ "repo_name": "Mooncake",
322
+ "repo_link": "https://github.com/kvcache-ai/Mooncake",
323
+ "github_about_section": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
324
+ "homepage_link": "https://kvcache-ai.github.io/Mooncake",
325
+ "github_topic_closest_fit": "inference",
326
+ "contributors_all": 138,
327
+ "contributors_2025": 133,
328
+ "contributors_2024": 13,
329
+ "contributors_2023": 0
330
+ },
331
+ {
332
+ "repo_name": "torchtitan",
333
+ "repo_link": "https://github.com/pytorch/torchtitan",
334
+ "github_about_section": "A PyTorch native platform for training generative AI models",
335
+ "contributors_all": 145,
336
+ "contributors_2025": 119,
337
+ "contributors_2024": 43,
338
+ "contributors_2023": 1
339
+ },
340
  {
341
  "repo_name": "ao",
342
  "repo_link": "https://github.com/pytorch/ao",
 
349
  "contributors_2023": 5
350
  },
351
  {
352
+ "repo_name": "lean4",
353
+ "repo_link": "https://github.com/leanprover/lean4",
354
+ "category": "theorem prover",
355
+ "github_about_section": "Lean 4 programming language and theorem prover",
356
+ "homepage_link": "https://lean-lang.org",
357
+ "github_topic_closest_fit": "lean",
358
+ "contributors_all": 278,
359
+ "contributors_2025": 110,
360
+ "contributors_2024": 85,
361
+ "contributors_2023": 64
 
 
 
 
 
 
 
 
362
  },
363
  {
364
  "repo_name": "ComfyUI",
 
373
  "contributors_2023": 94
374
  },
375
  {
376
+ "repo_name": "unsloth",
377
+ "repo_link": "https://github.com/unslothai/unsloth",
378
+ "category": "fine tuning",
379
+ "github_about_section": "Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
380
+ "homepage_link": "https://docs.unsloth.ai",
381
+ "github_topic_closest_fit": "fine-tuning",
382
+ "contributors_all": 127,
383
+ "contributors_2025": 102,
384
+ "contributors_2024": 27,
385
+ "contributors_2023": 3
386
+ },
387
+ {
388
+ "repo_name": "burn",
389
+ "repo_link": "https://github.com/tracel-ai/burn",
390
+ "github_about_section": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
391
+ "homepage_link": "https://burn.dev",
392
+ "contributors_all": 237,
393
+ "contributors_2025": 99,
394
+ "contributors_2024": 104,
395
+ "contributors_2023": 62
396
  },
397
  {
398
+ "repo_name": "accelerate",
399
+ "repo_link": "https://github.com/huggingface/accelerate",
400
+ "github_about_section": "A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support.",
401
+ "homepage_link": "https://huggingface.co/docs/accelerate",
402
+ "contributors_all": 392,
403
+ "contributors_2025": 97,
404
+ "contributors_2024": 124,
405
+ "contributors_2023": 149
 
 
406
  },
407
  {
408
+ "repo_name": "terminal-bench",
409
+ "repo_link": "https://github.com/laude-institute/terminal-bench",
410
+ "category": "benchmark",
411
+ "github_about_section": "A benchmark for LLMs on complicated tasks in the terminal",
412
+ "homepage_link": "https://tbench.ai",
413
+ "github_topic_closest_fit": "benchmark",
414
+ "contributors_all": 96,
415
+ "contributors_2025": 96,
416
+ "contributors_2024": 0,
417
  "contributors_2023": 0
418
  },
419
  {
 
427
  "contributors_2023": 165
428
  },
429
  {
430
+ "repo_name": "milvus",
431
+ "repo_link": "https://github.com/milvus-io/milvus",
432
+ "category": "vector database",
433
+ "github_about_section": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
434
+ "homepage_link": "https://milvus.io",
435
+ "github_topic_closest_fit": "vector-search",
436
+ "contributors_all": 387,
437
+ "contributors_2025": 95,
438
+ "contributors_2024": 84,
439
+ "contributors_2023": 72
440
  },
441
  {
442
+ "repo_name": "cutlass",
443
+ "repo_link": "https://github.com/NVIDIA/cutlass",
444
+ "category": "parallel computing",
445
+ "github_about_section": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
446
+ "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
447
+ "github_topic_closest_fit": "parallel-programming",
448
+ "contributors_all": 238,
449
+ "contributors_2025": 94,
450
+ "contributors_2024": 64,
451
+ "contributors_2023": 66
452
  },
453
  {
454
+ "repo_name": "tilelang",
455
+ "repo_link": "https://github.com/tile-ai/tilelang",
456
+ "category": "parallel computing dsl",
457
+ "github_about_section": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
458
+ "homepage_link": "https://tilelang.com",
459
+ "github_topic_closest_fit": "parallel-programming",
460
+ "contributors_all": 90,
461
+ "contributors_2025": 89,
462
+ "contributors_2024": 1,
463
+ "contributors_2023": 0
464
  },
465
  {
466
+ "repo_name": "monarch",
467
+ "repo_link": "https://github.com/meta-pytorch/monarch",
468
+ "github_about_section": "PyTorch Single Controller",
469
+ "homepage_link": "https://meta-pytorch.org/monarch",
470
+ "contributors_all": 85,
471
+ "contributors_2025": 85,
 
472
  "contributors_2024": 0,
473
  "contributors_2023": 0
474
  },
475
  {
476
+ "repo_name": "Liger-Kernel",
477
+ "repo_link": "https://github.com/linkedin/Liger-Kernel",
478
+ "category": "kernel examples",
479
+ "github_about_section": "Efficient Triton Kernels for LLM Training",
480
+ "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
481
+ "github_topic_closest_fit": "triton",
482
+ "contributors_all": 120,
483
+ "contributors_2025": 78,
484
+ "contributors_2024": 61,
485
+ "contributors_2023": 0
486
  },
487
  {
488
+ "repo_name": "nixl",
489
+ "repo_link": "https://github.com/ai-dynamo/nixl",
490
+ "github_about_section": "NVIDIA Inference Xfer Library (NIXL)",
491
+ "contributors_all": 78,
492
+ "contributors_2025": 78,
493
+ "contributors_2024": 0,
494
+ "contributors_2023": 0
 
495
  },
496
  {
497
+ "repo_name": "jupyterlab",
498
+ "repo_link": "https://github.com/jupyterlab/jupyterlab",
499
+ "category": "user interface",
500
+ "github_about_section": "JupyterLab computational environment.",
501
+ "homepage_link": "https://jupyterlab.readthedocs.io",
502
+ "github_topic_closest_fit": "jupyter",
503
+ "contributors_all": 698,
504
+ "contributors_2025": 77,
505
+ "contributors_2024": 85,
506
+ "contributors_2023": 100
507
  },
508
  {
509
+ "repo_name": "hipBLASLt",
510
+ "repo_link": "https://github.com/AMD-AGI/hipBLASLt",
511
+ "category": "Basic Linear Algebra Subprograms (BLAS)",
512
+ "github_about_section": "hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditional BLAS library",
513
+ "homepage_link": "https://rocm.docs.amd.com/projects/hipBLASLt",
514
+ "github_topic_closest_fit": "matrix-multiplication",
515
+ "contributors_all": 111,
516
+ "contributors_2025": 69,
517
+ "contributors_2024": 70,
518
+ "contributors_2023": 35
519
  },
520
  {
521
+ "repo_name": "peft",
522
+ "repo_link": "https://github.com/huggingface/peft",
523
+ "github_about_section": "PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
524
+ "homepage_link": "https://huggingface.co/docs/peft",
525
+ "github_topic_closest_fit": "lora",
526
+ "contributors_all": 272,
527
+ "contributors_2025": 69,
528
+ "contributors_2024": 111,
529
+ "contributors_2023": 115
 
530
  },
531
  {
532
+ "repo_name": "ROCm",
533
+ "repo_link": "https://github.com/ROCm/ROCm",
534
+ "github_about_section": "AMD ROCm Software - GitHub Home",
535
+ "homepage_link": "https://rocm.docs.amd.com",
536
+ "contributors_all": 166,
537
+ "contributors_2025": 67,
538
+ "contributors_2024": 61,
539
+ "contributors_2023": 44
 
 
540
  },
541
  {
542
+ "repo_name": "mcp-agent",
543
+ "repo_link": "https://github.com/lastmile-ai/mcp-agent",
544
+ "category": "mcp",
545
+ "github_about_section": "Build effective agents using Model Context Protocol and simple workflow patterns",
546
+ "github_topic_closest_fit": "mcp",
547
+ "contributors_all": 63,
548
+ "contributors_2025": 63,
549
+ "contributors_2024": 1,
550
+ "contributors_2023": 0
551
+ },
552
+ {
553
+ "repo_name": "rdma-core",
554
+ "repo_link": "https://github.com/linux-rdma/rdma-core",
555
+ "github_about_section": "RDMA core userspace libraries and daemons",
556
+ "contributors_all": 437,
557
+ "contributors_2025": 58,
558
+ "contributors_2024": 61,
559
+ "contributors_2023": 66
560
+ },
561
+ {
562
+ "repo_name": "onnx",
563
+ "repo_link": "https://github.com/onnx/onnx",
564
+ "category": "machine learning interoperability",
565
+ "github_about_section": "Open standard for machine learning interoperability",
566
+ "homepage_link": "https://onnx.ai",
567
+ "github_topic_closest_fit": "onnx",
568
+ "contributors_all": 370,
569
+ "contributors_2025": 56,
570
+ "contributors_2024": 45,
571
+ "contributors_2023": 61
572
  },
573
  {
574
  "repo_name": "letta",
 
582
  "contributors_2024": 75,
583
  "contributors_2023": 47
584
  },
585
+ {
586
+ "repo_name": "helion",
587
+ "repo_link": "https://github.com/pytorch/helion",
588
+ "category": "parallel computing dsl",
589
+ "github_about_section": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
590
+ "homepage_link": "https://helionlang.com",
591
+ "github_topic_closest_fit": "parallel-programming",
592
+ "contributors_all": 49,
593
+ "contributors_2025": 49,
594
+ "contributors_2024": 0,
595
+ "contributors_2023": 0
596
+ },
597
+ {
598
+ "repo_name": "MIOpen",
599
+ "repo_link": "https://github.com/ROCm/MIOpen",
600
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
601
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
602
+ "contributors_all": 204,
603
+ "contributors_2025": 47,
604
+ "contributors_2024": 62,
605
+ "contributors_2023": 44
606
+ },
607
+ {
608
+ "repo_name": "hip",
609
+ "repo_link": "https://github.com/ROCm/hip",
610
+ "github_about_section": "HIP: C++ Heterogeneous-Compute Interface for Portability",
611
+ "homepage_link": "https://rocmdocs.amd.com/projects/HIP",
612
+ "contributors_all": 288,
613
+ "contributors_2025": 46,
614
+ "contributors_2024": 31,
615
+ "contributors_2023": 25
616
+ },
617
+ {
618
+ "repo_name": "openevolve",
619
+ "repo_link": "https://github.com/codelion/openevolve",
620
+ "github_about_section": "Open-source implementation of AlphaEvolve",
621
+ "github_topic_closest_fit": "genetic-algorithm",
622
+ "contributors_all": 46,
623
+ "contributors_2025": 46,
624
+ "contributors_2024": 0,
625
+ "contributors_2023": 0
626
+ },
627
  {
628
  "repo_name": "lightning-thunder",
629
  "repo_link": "https://github.com/Lightning-AI/lightning-thunder",
 
634
  "contributors_2023": 29
635
  },
636
  {
637
+ "repo_name": "truss",
638
+ "repo_link": "https://github.com/basetenlabs/truss",
639
+ "category": "inference engine",
640
+ "github_about_section": "The simplest way to serve AI/ML models in production",
641
+ "homepage_link": "https://truss.baseten.co",
642
+ "github_topic_closest_fit": "inference",
643
+ "contributors_all": 72,
644
+ "contributors_2025": 44,
645
+ "contributors_2024": 30,
646
+ "contributors_2023": 21
647
  },
648
  {
649
+ "repo_name": "ondemand",
650
+ "repo_link": "https://github.com/OSC/ondemand",
651
+ "github_about_section": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
652
+ "homepage_link": "https://openondemand.org",
653
+ "github_topic_closest_fit": "hpc",
654
+ "contributors_all": 117,
655
+ "contributors_2025": 43,
656
+ "contributors_2024": 23,
657
+ "contributors_2023": 21
658
+ },
659
+ {
660
+ "repo_name": "pybind11",
661
+ "repo_link": "https://github.com/pybind/pybind11",
662
+ "github_about_section": "Seamless operability between C++11 and Python",
663
+ "homepage_link": "https://pybind11.readthedocs.io",
664
+ "github_topic_closest_fit": "bindings",
665
+ "contributors_all": 404,
666
+ "contributors_2025": 43,
667
+ "contributors_2024": 45,
668
+ "contributors_2023": 42
669
+ },
670
+ {
671
+ "repo_name": "cuda-python",
672
+ "repo_link": "https://github.com/NVIDIA/cuda-python",
673
+ "github_about_section": "CUDA Python: Performance meets Productivity",
674
+ "homepage_link": "https://nvidia.github.io/cuda-python",
675
+ "github_topic_closest_fit": "parallel-programming",
676
+ "contributors_all": 48,
677
+ "contributors_2025": 41,
678
+ "contributors_2024": 12,
679
+ "contributors_2023": 1
680
+ },
681
+ {
682
+ "repo_name": "warp",
683
+ "repo_link": "https://github.com/NVIDIA/warp",
684
+ "category": "spatial computing",
685
+ "github_about_section": "A Python framework for accelerated simulation, data generation and spatial computing.",
686
+ "homepage_link": "https://nvidia.github.io/warp",
687
+ "github_topic_closest_fit": "physics-simulation",
688
+ "contributors_all": 79,
689
+ "contributors_2025": 40,
690
+ "contributors_2024": 29,
691
+ "contributors_2023": 17
692
  },
693
  {
694
  "repo_name": "metaflow",
 
701
  "contributors_2023": 28
702
  },
703
  {
704
+ "repo_name": "numba",
705
+ "repo_link": "https://github.com/numba/numba",
706
+ "github_about_section": "NumPy aware dynamic Python compiler using LLVM",
707
+ "homepage_link": "https://numba.pydata.org",
708
+ "contributors_all": 430,
709
+ "contributors_2025": 36,
710
+ "contributors_2024": 32,
711
+ "contributors_2023": 55
712
  },
713
  {
714
+ "repo_name": "SWE-bench",
715
+ "repo_link": "https://github.com/SWE-bench/SWE-bench",
716
+ "category": "benchmark",
717
+ "github_about_section": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
718
+ "homepage_link": "https://swebench.com",
719
+ "github_topic_closest_fit": "benchmark",
720
+ "contributors_all": 66,
721
+ "contributors_2025": 33,
722
+ "contributors_2024": 37,
723
+ "contributors_2023": 9
724
+ },
725
+ {
726
+ "repo_name": "AdaptiveCpp",
727
+ "repo_link": "https://github.com/AdaptiveCpp/AdaptiveCpp",
728
+ "github_about_section": "Compiler for multiple programming models (SYCL, C++ standard parallelism, HIP/CUDA) for CPUs and GPUs from all vendors: The independent, community-driven compiler for C++-based heterogeneous programming models. Lets applications adapt themselves to all the hardware in the system - even at runtime!",
729
+ "homepage_link": "https://adaptivecpp.github.io",
730
+ "contributors_all": 93,
731
+ "contributors_2025": 32,
732
+ "contributors_2024": 32,
733
+ "contributors_2023": 24
734
+ },
735
+ {
736
+ "repo_name": "Triton-distributed",
737
+ "repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
738
+ "github_about_section": "Distributed Compiler based on Triton for Parallel Systems",
739
+ "homepage_link": "https://triton-distributed.readthedocs.io",
740
+ "contributors_all": 30,
741
+ "contributors_2025": 30,
742
+ "contributors_2024": 0,
743
+ "contributors_2023": 0
744
+ },
745
+ {
746
+ "repo_name": "ThunderKittens",
747
+ "repo_link": "https://github.com/HazyResearch/ThunderKittens",
748
+ "category": "parallel computing",
749
+ "github_about_section": "Tile primitives for speedy kernels",
750
+ "homepage_link": "https://hazyresearch.stanford.edu/blog/2024-10-29-tk2",
751
+ "github_topic_closest_fit": "parallel-programming",
752
+ "contributors_all": 34,
753
+ "contributors_2025": 29,
754
+ "contributors_2024": 13,
755
+ "contributors_2023": 0
756
+ },
757
+ {
758
+ "repo_name": "dstack",
759
+ "repo_link": "https://github.com/dstackai/dstack",
760
+ "category": "gpu provisioning and orchestration",
761
+ "github_about_section": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs-across hyperscalers, neoclouds, or on-prem.",
762
+ "homepage_link": "https://dstack.ai",
763
+ "github_topic_closest_fit": "orchestration",
764
+ "contributors_all": 69,
765
+ "contributors_2025": 28,
766
  "contributors_2024": 42,
767
+ "contributors_2023": 14
768
+ },
769
+ {
770
+ "repo_name": "ome",
771
+ "repo_link": "https://github.com/sgl-project/ome",
772
+ "github_about_section": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
773
+ "homepage_link": "http://docs.sglang.ai/ome",
774
+ "github_topic_closest_fit": "k8s",
775
+ "contributors_all": 28,
776
+ "contributors_2025": 28,
777
+ "contributors_2024": 0,
778
  "contributors_2023": 0
779
  },
780
  {
781
+ "repo_name": "pocl",
782
+ "repo_link": "https://github.com/pocl/pocl",
783
+ "github_about_section": "pocl - Portable Computing Language",
784
+ "homepage_link": "https://portablecl.org",
785
+ "github_topic_closest_fit": "parallel-programming",
786
+ "contributors_all": 166,
787
+ "contributors_2025": 26,
788
+ "contributors_2024": 27,
789
+ "contributors_2023": 21
790
+ },
791
+ {
792
+ "repo_name": "server",
793
+ "repo_link": "https://github.com/triton-inference-server/server",
794
+ "github_about_section": "The Triton Inference Server provides an optimized cloud and edge inferencing solution.",
795
+ "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
796
+ "github_topic_closest_fit": "inference",
797
+ "contributors_all": 147,
798
+ "contributors_2025": 24,
799
+ "contributors_2024": 36,
800
+ "contributors_2023": 34
801
+ },
802
+ {
803
+ "repo_name": "hipBLAS",
804
+ "repo_link": "https://github.com/ROCm/hipBLAS",
805
+ "category": "Basic Linear Algebra Subprograms (BLAS)",
806
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
807
+ "github_topic_closest_fit": "matrix-multiplication",
808
+ "contributors_all": 72,
809
+ "contributors_2025": 21,
810
+ "contributors_2024": 24,
811
+ "contributors_2023": 14
812
+ },
813
+ {
814
+ "repo_name": "Vulkan-Hpp",
815
+ "repo_link": "https://github.com/KhronosGroup/Vulkan-Hpp",
816
+ "category": "graphics api",
817
+ "github_about_section": "Open-Source Vulkan C++ API",
818
+ "homepage_link": "https://vulkan.org",
819
+ "github_topic_closest_fit": "vulkan",
820
+ "contributors_all": 102,
821
+ "contributors_2025": 21,
822
+ "contributors_2024": 15,
823
+ "contributors_2023": 15
824
+ },
825
+ {
826
+ "repo_name": "ccache",
827
+ "repo_link": "https://github.com/ccache/ccache",
828
+ "github_about_section": "ccache - a fast compiler cache",
829
+ "homepage_link": "https://ccache.dev",
830
+ "contributors_all": 218,
831
+ "contributors_2025": 20,
832
+ "contributors_2024": 28,
833
+ "contributors_2023": 22
834
+ },
835
+ {
836
+ "repo_name": "lapack",
837
+ "repo_link": "https://github.com/Reference-LAPACK/lapack",
838
+ "category": "linear algebra",
839
+ "github_about_section": "LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.",
840
+ "homepage_link": "https://netlib.org/lapack",
841
+ "github_topic_closest_fit": "linear-algebra",
842
+ "contributors_all": 178,
843
+ "contributors_2025": 20,
844
+ "contributors_2024": 24,
845
+ "contributors_2023": 42
846
  },
847
  {
848
+ "repo_name": "rocSOLVER",
849
+ "repo_link": "https://github.com/ROCm/rocSOLVER",
850
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
851
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
852
+ "contributors_all": 59,
853
+ "contributors_2025": 20,
854
+ "contributors_2024": 23,
855
+ "contributors_2023": 15
856
  },
857
  {
858
+ "repo_name": "Vulkan-Tools",
859
+ "repo_link": "https://github.com/KhronosGroup/Vulkan-Tools",
860
+ "category": "graphics api",
861
+ "github_about_section": "Vulkan Development Tools",
862
+ "homepage_link": "https://vulkan.org",
863
+ "github_topic_closest_fit": "vulkan",
864
+ "contributors_all": 248,
865
+ "contributors_2025": 20,
866
+ "contributors_2024": 24,
867
+ "contributors_2023": 24
868
  },
869
  {
870
+ "repo_name": "rocSPARSE",
871
+ "repo_link": "https://github.com/ROCm/rocSPARSE",
872
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
873
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
874
+ "contributors_all": 65,
875
+ "contributors_2025": 19,
876
+ "contributors_2024": 24,
877
+ "contributors_2023": 18
878
  },
879
  {
880
+ "repo_name": "tflite-micro",
881
+ "repo_link": "https://github.com/tensorflow/tflite-micro",
882
+ "github_about_section": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors).",
883
+ "contributors_all": 111,
884
+ "contributors_2025": 19,
885
+ "contributors_2024": 25,
886
+ "contributors_2023": 31
887
  },
888
  {
889
+ "repo_name": "Vulkan-Docs",
890
+ "repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
891
+ "category": "graphics api",
892
+ "github_about_section": "The Vulkan API Specification and related tools",
893
+ "homepage_link": "https://vulkan.org",
894
+ "github_topic_closest_fit": "vulkan",
895
+ "contributors_all": 141,
896
+ "contributors_2025": 18,
897
+ "contributors_2024": 21,
898
+ "contributors_2023": 34
899
  },
900
  {
901
+ "repo_name": "quack",
902
+ "repo_link": "https://github.com/Dao-AILab/quack",
903
+ "category": "kernel examples",
904
+ "github_about_section": "A Quirky Assortment of CuTe Kernels",
905
+ "contributors_all": 17,
906
+ "contributors_2025": 17,
 
907
  "contributors_2024": 0,
908
  "contributors_2023": 0
909
  },
 
 
 
 
 
 
 
 
 
 
 
910
  {
911
  "repo_name": "oneDPL",
912
  "repo_link": "https://github.com/uxlfoundation/oneDPL",
 
918
  "contributors_2023": 28
919
  },
920
  {
921
+ "repo_name": "rocFFT",
922
+ "repo_link": "https://github.com/ROCm/rocFFT",
923
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
924
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
925
+ "contributors_all": 81,
926
+ "contributors_2025": 17,
927
+ "contributors_2024": 20,
928
+ "contributors_2023": 19
929
+ },
930
+ {
931
+ "repo_name": "KernelBench",
932
+ "repo_link": "https://github.com/ScalingIntelligence/KernelBench",
933
+ "category": "benchmark",
934
+ "github_about_section": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
935
+ "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench",
936
+ "github_topic_closest_fit": "benchmark",
937
+ "contributors_all": 19,
938
+ "contributors_2025": 16,
939
+ "contributors_2024": 3,
940
  "contributors_2023": 0
941
  },
942
  {
943
+ "repo_name": "reference-kernels",
944
+ "repo_link": "https://github.com/gpu-mode/reference-kernels",
945
+ "category": "kernel examples",
946
+ "github_about_section": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
947
+ "homepage_link": "https://gpumode.com",
948
+ "contributors_all": 16,
949
+ "contributors_2025": 16,
950
+ "contributors_2024": 0,
951
+ "contributors_2023": 0
952
  },
953
  {
954
+ "repo_name": "Tensile",
955
+ "repo_link": "https://github.com/ROCm/Tensile",
956
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
957
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
958
+ "contributors_all": 137,
959
+ "contributors_2025": 16,
960
+ "contributors_2024": 25,
961
+ "contributors_2023": 22
 
962
  },
963
  {
964
+ "repo_name": "synthetic-data-kit",
965
+ "repo_link": "https://github.com/meta-llama/synthetic-data-kit",
966
+ "category": "synthetic data generation",
967
+ "github_about_section": "Tool for generating high quality Synthetic datasets",
968
+ "homepage_link": "https://pypi.org/project/synthetic-data-kit",
969
+ "github_topic_closest_fit": "synthetic-dataset-generation",
970
+ "contributors_all": 15,
971
+ "contributors_2025": 15,
972
  "contributors_2024": 0,
973
  "contributors_2023": 0
974
  },
975
  {
976
+ "repo_name": "tritonparse",
977
+ "repo_link": "https://github.com/meta-pytorch/tritonparse",
978
+ "github_about_section": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
979
+ "homepage_link": "https://meta-pytorch.org/tritonparse",
980
+ "contributors_all": 15,
981
+ "contributors_2025": 15,
982
+ "contributors_2024": 0,
983
+ "contributors_2023": 0
 
984
  },
985
  {
986
+ "repo_name": "kernels",
987
+ "repo_link": "https://github.com/huggingface/kernels",
988
+ "category": "gpu kernels",
989
+ "github_about_section": "Load compute kernels from the Hub",
990
+ "contributors_all": 15,
991
+ "contributors_2025": 14,
992
+ "contributors_2024": 2,
 
993
  "contributors_2023": 0
994
  },
995
  {
996
+ "repo_name": "Wan2.2",
997
+ "repo_link": "https://github.com/Wan-Video/Wan2.2",
998
+ "category": "video generation",
999
+ "github_about_section": "Wan: Open and Advanced Large-Scale Video Generative Models",
1000
+ "homepage_link": "https://wan.video",
1001
+ "github_topic_closest_fit": "diffusion-models",
1002
+ "contributors_all": 14,
1003
+ "contributors_2025": 14,
1004
+ "contributors_2024": 0,
1005
+ "contributors_2023": 0
1006
  },
1007
  {
1008
+ "repo_name": "SYCL-Docs",
1009
+ "repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
1010
+ "github_about_section": "SYCL Open Source Specification",
1011
+ "homepage_link": "https://khronos.org/sycl",
1012
+ "github_topic_closest_fit": "parallel-programming",
1013
+ "contributors_all": 67,
1014
+ "contributors_2025": 13,
1015
  "contributors_2024": 20,
1016
+ "contributors_2023": 27
 
 
 
 
 
 
 
 
 
 
1017
  },
1018
  {
1019
+ "repo_name": "Primus-Turbo",
1020
+ "repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
1021
+ "contributors_all": 12,
1022
+ "contributors_2025": 12,
1023
+ "contributors_2024": 0,
1024
+ "contributors_2023": 0
 
1025
  },
1026
  {
1027
  "repo_name": "rocPRIM",
 
1033
  "contributors_2024": 28,
1034
  "contributors_2023": 15
1035
  },
1036
+ {
1037
+ "repo_name": "flashinfer-bench",
1038
+ "repo_link": "https://github.com/flashinfer-ai/flashinfer-bench",
1039
+ "category": "benchmark",
1040
+ "github_about_section": "Building the Virtuous Cycle for AI-driven LLM Systems",
1041
+ "homepage_link": "https://bench.flashinfer.ai",
1042
+ "github_topic_closest_fit": "benchmark",
1043
+ "contributors_all": 12,
1044
+ "contributors_2025": 11,
1045
+ "contributors_2024": 0,
1046
+ "contributors_2023": 0
1047
+ },
1048
+ {
1049
+ "repo_name": "FTorch",
1050
+ "repo_link": "https://github.com/Cambridge-ICCS/FTorch",
1051
+ "category": "wrapper",
1052
+ "github_about_section": "A library for directly calling PyTorch ML models from Fortran.",
1053
+ "homepage_link": "https://cambridge-iccs.github.io/FTorch",
1054
+ "github_topic_closest_fit": "machine-learning",
1055
+ "contributors_all": 20,
1056
+ "contributors_2025": 11,
1057
+ "contributors_2024": 8,
1058
+ "contributors_2023": 9
1059
+ },
1060
+ {
1061
+ "repo_name": "TensorRT",
1062
+ "repo_link": "https://github.com/NVIDIA/TensorRT",
1063
+ "github_about_section": "NVIDIA TensorRT is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
1064
+ "homepage_link": "https://developer.nvidia.com/tensorrt",
1065
+ "contributors_all": 104,
1066
+ "contributors_2025": 10,
1067
+ "contributors_2024": 18,
1068
+ "contributors_2023": 19
1069
+ },
1070
+ {
1071
+ "repo_name": "TileIR",
1072
+ "repo_link": "https://github.com/microsoft/TileIR",
1073
+ "category": "parallel computing dsl",
1074
+ "github_about_section": "TileIR (tile-ir) is a concise domain-specific IR designed to streamline the development of high-performance GPU/CPU kernels (e.g., GEMM, Dequant GEMM, FlashAttention, LinearAttention). By employing a Pythonic syntax with an underlying compiler infrastructure on top of TVM, TileIR allows developers to focus on productivity without sacrificing the low-level optimizations necessary for state-of-the-art performance.",
1075
+ "github_topic_closest_fit": "parallel-programming",
1076
+ "contributors_all": 10,
1077
+ "contributors_2025": 10,
1078
+ "contributors_2024": 1,
1079
+ "contributors_2023": 0
1080
+ },
1081
+ {
1082
+ "repo_name": "hipCUB",
1083
+ "repo_link": "https://github.com/ROCm/hipCUB",
1084
+ "github_about_section": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1085
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1086
+ "contributors_all": 54,
1087
+ "contributors_2025": 10,
1088
+ "contributors_2024": 19,
1089
+ "contributors_2023": 13
1090
+ },
1091
  {
1092
  "repo_name": "rocRAND",
1093
  "repo_link": "https://github.com/ROCm/rocRAND",
 
1099
  "contributors_2023": 26
1100
  },
1101
  {
1102
+ "repo_name": "kernels-community",
1103
+ "repo_link": "https://github.com/huggingface/kernels-community",
1104
+ "category": "gpu kernels",
1105
+ "homepage_link": "https://huggingface.co/kernels-community",
1106
+ "github_about_section": "Kernel sources for https://huggingface.co/kernels-community",
1107
+ "contributors_all": 9,
1108
+ "contributors_2025": 9,
1109
+ "contributors_2024": 0,
1110
+ "contributors_2023": 0
1111
  },
1112
  {
1113
+ "repo_name": "GEAK-agent",
1114
+ "repo_link": "https://github.com/AMD-AGI/GEAK-agent",
1115
+ "category": "agent",
1116
+ "github_about_section": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
1117
+ "github_topic_closest_fit": "ai-agents",
1118
+ "contributors_all": 9,
1119
+ "contributors_2025": 9,
1120
+ "contributors_2024": 0,
1121
+ "contributors_2023": 0
1122
+ },
1123
+ {
1124
+ "repo_name": "neuronx-distributed-inference",
1125
+ "repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
1126
+ "contributors_all": 11,
1127
+ "contributors_2025": 9,
1128
+ "contributors_2024": 3,
1129
+ "contributors_2023": 0
1130
+ },
1131
+ {
1132
+ "repo_name": "OpenCL-SDK",
1133
+ "repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
1134
+ "github_about_section": "OpenCL SDK",
1135
+ "homepage_link": "https://khronos.org/opencl",
1136
+ "github_topic_closest_fit": "parallel-programming",
1137
+ "contributors_all": 25,
1138
+ "contributors_2025": 8,
1139
+ "contributors_2024": 6,
1140
+ "contributors_2023": 9
1141
+ },
1142
+ {
1143
+ "repo_name": "ZLUDA",
1144
+ "repo_link": "https://github.com/vosen/ZLUDA",
1145
+ "github_about_section": "CUDA on non-NVIDIA GPUs",
1146
+ "homepage_link": "https://vosen.github.io/ZLUDA",
1147
+ "github_topic_closest_fit": "parallel-programming",
1148
+ "contributors_all": 15,
1149
+ "contributors_2025": 8,
1150
+ "contributors_2024": 4,
1151
+ "contributors_2023": 0
1152
  },
1153
  {
1154
  "repo_name": "roctracer",
 
1160
  "contributors_2024": 11,
1161
  "contributors_2023": 6
1162
  },
1163
+ {
1164
+ "repo_name": "intelliperf",
1165
+ "repo_link": "https://github.com/AMDResearch/intelliperf",
1166
+ "category": "performance testing",
1167
+ "github_about_section": "Automated bottleneck detection and solution orchestration",
1168
+ "homepage_link": "https://arxiv.org/html/2508.20258v1",
1169
+ "github_topic_closest_fit": "profiling",
1170
+ "contributors_all": 7,
1171
+ "contributors_2025": 7,
1172
+ "contributors_2024": 0,
1173
+ "contributors_2023": 0
1174
+ },
1175
+ {
1176
+ "repo_name": "nccl",
1177
+ "repo_link": "https://github.com/NVIDIA/nccl",
1178
+ "github_about_section": "Optimized primitives for collective multi-GPU communication",
1179
+ "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
1180
+ "contributors_all": 51,
1181
+ "contributors_2025": 7,
1182
+ "contributors_2024": 5,
1183
+ "contributors_2023": 6
1184
+ },
1185
+ {
1186
+ "repo_name": "cudnn-frontend",
1187
+ "repo_link": "https://github.com/NVIDIA/cudnn-frontend",
1188
+ "category": "parallel computing",
1189
+ "github_about_section": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
1190
+ "homepage_link": "https://developer.nvidia.com/cudnn",
1191
+ "github_topic_closest_fit": "parallel-programming",
1192
+ "contributors_all": 12,
1193
+ "contributors_2025": 6,
1194
+ "contributors_2024": 5,
1195
+ "contributors_2023": 1
1196
+ },
1197
+ {
1198
+ "repo_name": "BitBLAS",
1199
+ "repo_link": "https://github.com/microsoft/BitBLAS",
1200
+ "category": "Basic Linear Algebra Subprograms (BLAS)",
1201
+ "github_about_section": "BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.",
1202
+ "github_topic_closest_fit": "matrix-multiplication",
1203
+ "contributors_all": 17,
1204
+ "contributors_2025": 5,
1205
+ "contributors_2024": 14,
1206
+ "contributors_2023": 0
1207
+ },
1208
  {
1209
  "repo_name": "Self-Forcing",
1210
  "repo_link": "https://github.com/guandeh17/Self-Forcing",
 
1218
  "contributors_2023": 0
1219
  },
1220
  {
1221
+ "repo_name": "TritonBench",
1222
+ "repo_link": "https://github.com/thunlp/TritonBench",
1223
+ "category": "benchmark",
1224
+ "github_about_section": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
1225
+ "homepage_link": "https://arxiv.org/abs/2502.14752",
1226
+ "github_topic_closest_fit": "benchmark",
1227
+ "contributors_all": 3,
1228
+ "contributors_2025": 3,
1229
+ "contributors_2024": 0,
1230
+ "contributors_2023": 0
 
 
 
 
 
 
 
 
 
 
1231
  },
1232
  {
1233
+ "repo_name": "hatchet",
1234
+ "repo_link": "https://github.com/LLNL/hatchet",
1235
+ "category": "performance testing",
1236
+ "github_about_section": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
1237
+ "homepage_link": "https://llnl-hatchet.readthedocs.io",
1238
+ "github_topic_closest_fit": "profiling",
1239
+ "contributors_all": 25,
1240
+ "contributors_2025": 3,
1241
+ "contributors_2024": 6,
1242
+ "contributors_2023": 8
1243
  },
1244
  {
1245
  "repo_name": "streamv2v",
 
1250
  "github_topic_closest_fit": "diffusion-models",
1251
  "contributors_all": 7,
1252
  "contributors_2025": 3,
1253
+ "contributors_2024": 6,
 
 
 
 
 
 
 
 
 
 
 
 
1254
  "contributors_2023": 0
1255
  },
1256
  {
1257
+ "repo_name": "mistral-inference",
1258
+ "repo_link": "https://github.com/mistralai/mistral-inference",
1259
+ "category": "inference engine",
1260
+ "github_about_section": "Official inference library for Mistral models",
1261
+ "homepage_link": "https://mistral.ai",
1262
+ "github_topic_closest_fit": "inference",
1263
+ "contributors_all": 29,
1264
+ "contributors_2025": 2,
1265
+ "contributors_2024": 17,
1266
+ "contributors_2023": 14
1267
  },
1268
  {
1269
+ "repo_name": "omnitrace",
1270
+ "repo_link": "https://github.com/ROCm/omnitrace",
1271
+ "category": "performance testing",
1272
+ "github_about_section": "Omnitrace: Application Profiling, Tracing, and Analysis",
1273
+ "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace",
1274
+ "github_topic_closest_fit": "profiling",
1275
+ "contributors_all": 16,
1276
+ "contributors_2025": 2,
1277
+ "contributors_2024": 12,
1278
+ "contributors_2023": 2
1279
  },
1280
  {
1281
+ "repo_name": "cuJSON",
1282
+ "repo_link": "https://github.com/AutomataLab/cuJSON",
1283
+ "category": "library leveraging parallel compute",
1284
+ "github_about_section": "cuJSON: A Highly Parallel JSON Parser for GPUs",
1285
+ "homepage_link": "https://dl.acm.org/doi/10.1145/3760250.3762222",
1286
+ "github_topic_closest_fit": "json-parser",
1287
+ "contributors_all": 2,
1288
+ "contributors_2025": 2,
1289
+ "contributors_2024": 2,
1290
+ "contributors_2023": 0
1291
  },
1292
  {
1293
+ "repo_name": "IMO2025",
1294
+ "repo_link": "https://github.com/harmonic-ai/IMO2025",
1295
+ "category": "formal mathematical reasoning",
1296
+ "github_about_section": "Harmonic's model Aristotle achieved gold medal performance, solving 5 problems. This repository contains the lean statement files and proofs for Problems 1-5.",
1297
+ "homepage_link": "https://harmonic.fun",
1298
+ "github_topic_closest_fit": "lean",
1299
+ "contributors_all": 2,
1300
+ "contributors_2025": 2,
1301
+ "contributors_2024": 0,
1302
+ "contributors_2023": 0
1303
  },
1304
  {
1305
+ "repo_name": "RaBitQ",
1306
+ "repo_link": "https://github.com/gaoj0017/RaBitQ",
1307
+ "github_about_section": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
1308
+ "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
1309
+ "github_topic_closest_fit": "nearest-neighbor-search",
1310
+ "contributors_all": 2,
1311
+ "contributors_2025": 2,
1312
+ "contributors_2024": 1,
1313
+ "contributors_2023": 0
1314
  },
1315
  {
1316
+ "repo_name": "torchdendrite",
1317
+ "repo_link": "https://github.com/sandialabs/torchdendrite",
1318
+ "category": "machine learning framework",
1319
+ "github_about_section": "Dendrites for PyTorch and SNNTorch neural networks",
1320
+ "contributors_all": 2,
1321
+ "contributors_2025": 1,
1322
+ "contributors_2024": 1,
1323
  "contributors_2023": 0
1324
  },
1325
  {
 
1333
  "contributors_2023": 0
1334
  },
1335
  {
1336
+ "repo_name": "nvcc4jupyter",
1337
+ "repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
1338
+ "github_about_section": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
1339
+ "homepage_link": "https://nvcc4jupyter.readthedocs.io",
1340
+ "contributors_all": 9,
1341
+ "contributors_2025": 0,
1342
+ "contributors_2024": 3,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1343
  "contributors_2023": 3
1344
  },
1345
  {
1346
+ "repo_name": "CU2CL",
1347
+ "repo_link": "https://github.com/vtsynergy/CU2CL",
1348
+ "github_about_section": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
1349
+ "homepage_link": "http://chrec.cs.vt.edu/cu2cl",
1350
+ "github_topic_closest_fit": "parallel-programming",
1351
+ "contributors_all": 3,
1352
+ "contributors_2025": 0,
1353
+ "contributors_2024": 0,
 
1354
  "contributors_2023": 0
1355
  },
1356
  {
1357
+ "repo_name": "triSYCL",
1358
+ "repo_link": "https://github.com/triSYCL/triSYCL",
1359
+ "github_about_section": "Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
1360
+ "homepage_link": "https://trisycl.github.io/triSYCL/Doxygen/triSYCL/html/index.html",
1361
+ "github_topic_closest_fit": "parallel-programming",
1362
+ "contributors_all": 31,
1363
+ "contributors_2025": 0,
1364
+ "contributors_2024": 1,
1365
+ "contributors_2023": 3
 
1366
  },
1367
  {
1368
+ "repo_name": "cupti",
1369
+ "repo_link": "https://github.com/cwpearson/cupti",
1370
+ "category": "performance testing",
1371
+ "github_about_section": "Profile how CUDA applications create and modify data in memory.",
1372
+ "github_topic_closest_fit": "profiling",
1373
+ "contributors_all": 1,
1374
+ "contributors_2025": 0,
1375
+ "contributors_2024": 0,
1376
+ "contributors_2023": 0
 
1377
  },
1378
  {
1379
+ "repo_name": "ort",
1380
+ "repo_link": "https://github.com/pytorch/ort",
1381
+ "github_about_section": "Accelerate PyTorch models with ONNX Runtime",
1382
+ "contributors_all": 47,
1383
+ "contributors_2025": 0,
1384
+ "contributors_2024": 7,
1385
+ "contributors_2023": 9
 
 
 
1386
  },
1387
  {
1388
+ "repo_name": "StreamDiffusion",
1389
+ "repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
1390
+ "category": "image generation",
1391
+ "github_about_section": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation",
1392
+ "homepage_link": "https://arxiv.org/abs/2312.12491",
1393
  "github_topic_closest_fit": "diffusion-models",
1394
+ "contributors_all": 29,
1395
+ "contributors_2025": 0,
1396
+ "contributors_2024": 9,
1397
+ "contributors_2023": 25
1398
  },
1399
  {
1400
+ "repo_name": "torchdynamo",
1401
+ "repo_link": "https://github.com/pytorch/torchdynamo",
1402
+ "github_about_section": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster.",
1403
+ "contributors_all": 63,
1404
+ "contributors_2025": 0,
1405
+ "contributors_2024": 1,
1406
+ "contributors_2023": 4
 
 
 
1407
  }
1408
  ]