TylerHilbert commited on
Commit
707ecf1
·
verified ·
1 Parent(s): 665c668

Upload PyTorchConference2025_GithubRepos.json

Browse files
Files changed (1) hide show
  1. PyTorchConference2025_GithubRepos.json +104 -662
PyTorchConference2025_GithubRepos.json CHANGED
@@ -4,42 +4,14 @@
4
  "repo_name": "pytorch",
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
- "repo_tags": [
8
- "autograd",
9
- "deep-learning",
10
- "gpu",
11
- "machine-learning",
12
- "neural-network",
13
- "numpy",
14
- "python",
15
- "tensor"
16
- ],
17
  "category": "machine learning framework"
18
  },
19
- {
20
- "github_repo_link": "https://github.com/pytorch/executorch",
21
- "repo_name": "executorch",
22
- "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
23
- "homepage_link": "https://executorch.ai",
24
- "repo_tags": [
25
- "deep-learning",
26
- "embedded",
27
- "gpu",
28
- "machine-learning",
29
- "mobile",
30
- "neural-network",
31
- "tensor"
32
- ],
33
- "category": "model compiler"
34
- },
35
  {
36
  "github_repo_link": "https://github.com/ggml-org/llama.cpp",
37
  "repo_name": "llama.cpp",
38
  "repo_description": "LLM inference in C/C++",
39
- "homepage_link": "",
40
- "repo_tags": [
41
- "ggml"
42
- ],
43
  "category": "inference engine"
44
  },
45
  {
@@ -47,75 +19,21 @@
47
  "repo_name": "onnx",
48
  "repo_description": "Open standard for machine learning interoperability",
49
  "homepage_link": "https://onnx.ai/",
50
- "repo_tags": [
51
- "deep-learning",
52
- "deep-neural-networks",
53
- "dnn",
54
- "keras",
55
- "machine-learning",
56
- "ml",
57
- "neural-network",
58
- "onnx",
59
- "pytorch",
60
- "scikit-learn",
61
- "tensorflow"
62
- ]
63
  },
64
  {
65
  "github_repo_link": "https://github.com/ray-project/ray",
66
  "repo_name": "ray",
67
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
68
  "homepage_link": "https://ray.io",
69
- "repo_tags": [
70
- "data-science",
71
- "deep-learning",
72
- "deployment",
73
- "distributed",
74
- "hyperparameter-optimization",
75
- "hyperparameter-search",
76
- "large-language-models",
77
- "llm",
78
- "llm-inference",
79
- "llm-serving",
80
- "machine-learning",
81
- "optimization",
82
- "parallel",
83
- "python",
84
- "pytorch",
85
- "ray",
86
- "reinforcement-learning",
87
- "rllib",
88
- "serving",
89
- "tensorflow"
90
- ]
91
  },
92
  {
93
  "github_repo_link": "https://github.com/vllm-project/vllm",
94
  "repo_name": "vllm",
95
  "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
96
  "homepage_link": "https://docs.vllm.ai",
97
- "repo_tags": [
98
- "amd",
99
- "blackwell",
100
- "cuda",
101
- "deepseek",
102
- "deepseek-v3",
103
- "gpt",
104
- "gpt-oss",
105
- "inference",
106
- "kimi",
107
- "llama",
108
- "llm",
109
- "llm-serving",
110
- "model-serving",
111
- "moe",
112
- "openai",
113
- "pytorch",
114
- "qwen",
115
- "qwen3",
116
- "tpu",
117
- "transformer"
118
- ],
119
  "category": "inference engine"
120
  },
121
  {
@@ -123,25 +41,7 @@
123
  "repo_name": "ollama",
124
  "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
125
  "homepage_link": "https://ollama.com",
126
- "repo_tags": [
127
- "deepseek",
128
- "gemma",
129
- "gemma3",
130
- "gemma3n",
131
- "go",
132
- "golang",
133
- "gpt-oss",
134
- "llama",
135
- "llama2",
136
- "llama3",
137
- "llava",
138
- "llm",
139
- "llms",
140
- "mistral",
141
- "ollama",
142
- "phi4",
143
- "qwen"
144
- ],
145
  "category": "inference engine"
146
  },
147
  {
@@ -149,173 +49,106 @@
149
  "repo_name": "sglang",
150
  "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
151
  "homepage_link": "https://docs.sglang.ai/",
152
- "repo_tags": [
153
- "blackwell",
154
- "cuda",
155
- "deepseek",
156
- "deepseek-r1",
157
- "deepseek-v3",
158
- "deepseek-v3-2",
159
- "gpt-oss",
160
- "inference",
161
- "kimi",
162
- "llama",
163
- "llama3",
164
- "llava",
165
- "llm",
166
- "llm-serving",
167
- "moe",
168
- "openai",
169
- "pytorch",
170
- "qwen3",
171
- "transformer",
172
- "vlm"
173
- ],
174
  "category": "inference engine"
175
  },
 
 
 
 
 
 
 
176
  {
177
  "github_repo_link": "https://github.com/pytorch/ao",
178
  "repo_name": "ao",
179
  "repo_description": "PyTorch native quantization and sparsity for training and inference",
180
  "homepage_link": "https://pytorch.org/ao/stable/index.html",
181
- "repo_tags": [
182
- "brrr",
183
- "cuda",
184
- "dtypes",
185
- "float8",
186
- "inference",
187
- "llama",
188
- "mx",
189
- "offloading",
190
- "optimizer",
191
- "pytorch",
192
- "quantization",
193
- "sparsity",
194
- "training",
195
- "transformer"
196
- ]
197
  },
198
  {
199
  "github_repo_link": "https://github.com/triton-lang/triton",
200
  "repo_name": "triton",
201
  "repo_description": "Development repository for the Triton language and compiler",
202
  "homepage_link": "https://triton-lang.org/",
203
- "repo_tags": [],
204
  "category": "dsl"
205
  },
206
  {
207
  "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
208
  "repo_name": "ThunderKittens",
209
- "repo_description": "Tile primitives for speedy kernels",
210
- "homepage_link": null,
211
- "repo_tags": []
212
  },
213
  {
214
  "github_repo_link": "https://github.com/gpu-mode/reference-kernels",
215
  "repo_name": "reference-kernels",
216
  "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
217
  "homepage_link": "https://gpumode.com",
218
- "repo_tags": [
219
- "cuda",
220
- "gpu",
221
- "leaderboard",
222
- "triton"
223
- ],
224
  "category": "kernels"
225
  },
 
 
 
 
 
 
 
 
226
  {
227
  "github_repo_link": "https://github.com/guandeh17/Self-Forcing",
228
  "repo_name": "Self-Forcing",
229
  "repo_description": "Official codebase for \"Self Forcing: Bridging Training and Inference in Autoregressive Video Diffusion\" (NeurIPS 2025 Spotlight)",
230
- "homepage_link": "",
231
- "repo_tags": []
232
- },
233
- {
234
- "github_repo_link": "https://github.com/chenfengxu714/StreamDiffusionV2",
235
- "repo_name": "StreamDiffusionV2",
236
- "repo_description": "StreamDiffusion, Live Stream APP",
237
- "homepage_link": "",
238
- "repo_tags": []
239
  },
240
  {
241
  "github_repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
242
  "repo_name": "StreamDiffusion",
243
- "repo_description": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation",
244
- "homepage_link": "",
245
- "repo_tags": []
246
  },
247
  {
248
  "github_repo_link": "https://github.com/comfyanonymous/ComfyUI",
249
  "repo_name": "ComfyUI",
250
  "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
251
  "homepage_link": "https://www.comfy.org/",
252
- "repo_tags": [
253
- "ai",
254
- "comfy",
255
- "comfyui",
256
- "python",
257
- "pytorch",
258
- "stable-diffusion"
259
- ]
260
  },
261
  {
262
  "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
263
  "repo_name": "streamv2v",
264
  "repo_description": "Official Pytorch implementation of StreamV2V. ",
265
- "homepage_link": "https://jeff-liangf.github.io/projects/streamv2v/",
266
- "repo_tags": []
267
  },
268
  {
269
  "github_repo_link": "https://github.com/letta-ai/letta",
270
  "repo_name": "letta",
271
  "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
272
  "homepage_link": "https://docs.letta.com/",
273
- "repo_tags": [
274
- "ai",
275
- "ai-agents",
276
- "llm",
277
- "llm-agent"
278
- ]
279
  },
280
  {
281
  "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
282
  "repo_name": "jupyterlab",
283
  "repo_description": "JupyterLab computational environment.",
284
  "homepage_link": "https://jupyterlab.readthedocs.io/",
285
- "repo_tags": [
286
- "jupyter",
287
- "jupyterlab"
288
- ],
289
  "category": "ui"
290
  },
291
  {
292
  "github_repo_link": "https://github.com/ROCm/rocm-systems",
293
  "repo_name": "rocm-systems",
294
- "repo_description": "super repo for rocm systems projects",
295
- "homepage_link": "",
296
- "repo_tags": []
297
  },
298
  {
299
  "github_repo_link": "https://github.com/NVIDIA/cutlass",
300
  "repo_name": "cutlass",
301
  "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
302
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
303
- "repo_tags": [
304
- "cpp",
305
- "cuda",
306
- "deep-learning",
307
- "deep-learning-library",
308
- "gpu",
309
- "nvidia",
310
- "python"
311
- ]
312
  },
313
  {
314
  "github_repo_link": "https://github.com/pytorch/helion",
315
  "repo_name": "helion",
316
  "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
317
- "homepage_link": null,
318
- "repo_tags": [],
319
  "category": "dsl"
320
  },
321
  {
@@ -323,25 +156,14 @@
323
  "repo_name": "jax",
324
  "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
325
  "homepage_link": "https://docs.jax.dev",
326
- "repo_tags": [
327
- "jax"
328
- ]
329
  },
330
  {
331
  "github_repo_link": "https://github.com/tensorflow/tensorflow",
332
  "repo_name": "tensorflow",
333
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
334
  "homepage_link": "https://tensorflow.org",
335
- "repo_tags": [
336
- "deep-learning",
337
- "deep-neural-networks",
338
- "distributed",
339
- "machine-learning",
340
- "ml",
341
- "neural-network",
342
- "python",
343
- "tensorflow"
344
- ],
345
  "category": "machine learning framework"
346
  },
347
  {
@@ -349,60 +171,33 @@
349
  "repo_name": "DeepSpeed",
350
  "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
351
  "homepage_link": "https://www.deepspeed.ai/",
352
- "repo_tags": [
353
- "billion-parameters",
354
- "compression",
355
- "data-parallelism",
356
- "deep-learning",
357
- "gpu",
358
- "inference",
359
- "machine-learning",
360
- "mixture-of-experts",
361
- "model-parallelism",
362
- "pipeline-parallelism",
363
- "pytorch",
364
- "trillion-parameters",
365
- "zero"
366
- ]
367
  },
368
  {
369
  "github_repo_link": "https://github.com/triton-inference-server/server",
370
  "repo_name": "server",
371
  "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
372
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
373
- "repo_tags": [
374
- "cloud",
375
- "datacenter",
376
- "deep-learning",
377
- "edge",
378
- "gpu",
379
- "inference",
380
- "machine-learning"
381
- ]
382
  },
383
  {
384
  "github_repo_link": "https://github.com/ROCm/ROCm",
385
  "repo_name": "ROCm",
386
  "repo_description": "AMD ROCm™ Software - GitHub Home",
387
  "homepage_link": "https://rocm.docs.amd.com",
388
- "repo_tags": [
389
- "documentation"
390
- ]
391
  },
392
  {
393
  "github_repo_link": "https://github.com/llvm/llvm-project",
394
  "repo_name": "llvm-project",
395
  "repo_description": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
396
  "homepage_link": "http://llvm.org",
397
- "repo_tags": [],
398
  "category": "compiler"
399
  },
400
  {
401
  "github_repo_link": "https://github.com/cwpearson/cupti",
402
  "repo_name": "cupti",
403
  "repo_description": "Profile how CUDA applications create and modify data in memory.",
404
- "homepage_link": "",
405
- "repo_tags": [],
406
  "category": "profiler"
407
  },
408
  {
@@ -410,18 +205,7 @@
410
  "repo_name": "hatchet",
411
  "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
412
  "homepage_link": "https://llnl-hatchet.readthedocs.io",
413
- "repo_tags": [
414
- "comparative-analysis",
415
- "data-analytics",
416
- "graphs",
417
- "hierarchical-data",
418
- "hpc",
419
- "performance",
420
- "performance-analysis",
421
- "python",
422
- "radiuss",
423
- "trees"
424
- ],
425
  "category": "profiler"
426
  },
427
  {
@@ -429,19 +213,13 @@
429
  "repo_name": "triton-runner",
430
  "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
431
  "homepage_link": "https://triton-runner.org",
432
- "repo_tags": [
433
- "ai-infra",
434
- "cuda",
435
- "tools",
436
- "triton"
437
- ]
438
  },
439
  {
440
  "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
441
  "repo_name": "Triton-distributed",
442
  "repo_description": "Distributed Compiler based on Triton for Parallel Systems",
443
  "homepage_link": "https://triton-distributed.readthedocs.io/en/latest/",
444
- "repo_tags": [],
445
  "category": "model compiler"
446
  },
447
  {
@@ -449,27 +227,13 @@
449
  "repo_name": "Liger-Kernel",
450
  "repo_description": "Efficient Triton Kernels for LLM Training",
451
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
452
- "repo_tags": [
453
- "finetuning",
454
- "gemma2",
455
- "hacktoberfest",
456
- "llama",
457
- "llama3",
458
- "llm-training",
459
- "llms",
460
- "mistral",
461
- "phi3",
462
- "triton",
463
- "triton-kernels"
464
- ],
465
  "category": "kernels"
466
  },
467
  {
468
  "github_repo_link": "https://github.com/thunlp/TritonBench",
469
  "repo_name": "TritonBench",
470
  "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
471
- "homepage_link": "",
472
- "repo_tags": [],
473
  "category": "benchmark"
474
  },
475
  {
@@ -477,28 +241,14 @@
477
  "repo_name": "tritonparse",
478
  "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
479
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
480
- "repo_tags": [
481
- "compiler",
482
- "debugging",
483
- "gpu-kernel",
484
- "interactive-visualization",
485
- "ir-analysis",
486
- "ir-visualization",
487
- "pytorch",
488
- "structured-logging",
489
- "triton"
490
- ]
491
  },
492
  {
493
  "github_repo_link": "https://github.com/elastic/elasticsearch",
494
  "repo_name": "elasticsearch",
495
  "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
496
  "homepage_link": "https://www.elastic.co/products/elasticsearch",
497
- "repo_tags": [
498
- "elasticsearch",
499
- "java",
500
- "search-engine"
501
- ],
502
  "category": "search engine"
503
  },
504
  {
@@ -506,61 +256,26 @@
506
  "repo_name": "kubernetes",
507
  "repo_description": "Production-Grade Container Scheduling and Management",
508
  "homepage_link": "https://kubernetes.io",
509
- "repo_tags": [
510
- "cncf",
511
- "containers",
512
- "go",
513
- "kubernetes"
514
- ]
515
  },
516
  {
517
  "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
518
  "repo_name": "modelcontextprotocol",
519
  "repo_description": "Specification and documentation for the Model Context Protocol",
520
- "homepage_link": "https://modelcontextprotocol.io",
521
- "repo_tags": []
522
  },
523
  {
524
  "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
525
  "repo_name": "mcp-agent",
526
  "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
527
- "homepage_link": "",
528
- "repo_tags": [
529
- "agents",
530
- "ai",
531
- "ai-agents",
532
- "llm",
533
- "llms",
534
- "mcp",
535
- "model-context-protocol",
536
- "python"
537
- ]
538
  },
539
  {
540
  "github_repo_link": "https://github.com/milvus-io/milvus",
541
  "repo_name": "milvus",
542
  "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
543
  "homepage_link": "https://milvus.io",
544
- "repo_tags": [
545
- "anns",
546
- "cloud-native",
547
- "diskann",
548
- "distributed",
549
- "embedding-database",
550
- "embedding-similarity",
551
- "embedding-store",
552
- "faiss",
553
- "golang",
554
- "hnsw",
555
- "image-search",
556
- "llm",
557
- "nearest-neighbor-search",
558
- "rag",
559
- "vector-database",
560
- "vector-search",
561
- "vector-similarity",
562
- "vector-store"
563
- ],
564
  "category": "vector databse"
565
  },
566
  {
@@ -568,29 +283,19 @@
568
  "repo_name": "RaBitQ",
569
  "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
570
  "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
571
- "repo_tags": [
572
- "high-dimensional-vectors",
573
- "nearest-neighbor-search",
574
- "quantization"
575
- ]
576
  },
577
  {
578
  "github_repo_link": "https://github.com/Airtable/airtable.js",
579
  "repo_name": "airtable.js",
580
- "repo_description": "Airtable javascript client",
581
- "homepage_link": null,
582
- "repo_tags": []
583
  },
584
  {
585
  "github_repo_link": "https://github.com/mistralai/mistral-inference",
586
  "repo_name": "mistral-inference",
587
  "repo_description": "Official inference library for Mistral models",
588
  "homepage_link": "https://mistral.ai/",
589
- "repo_tags": [
590
- "llm",
591
- "llm-inference",
592
- "mistralai"
593
- ],
594
  "category": "inference engine"
595
  },
596
  {
@@ -622,10 +327,7 @@
622
  "repo_name": "numpy",
623
  "repo_description": "The fundamental package for scientific computing with Python.",
624
  "homepage_link": "https://numpy.org",
625
- "repo_tags": [
626
- "numpy",
627
- "python"
628
- ],
629
  "category": "python library"
630
  },
631
  {
@@ -633,13 +335,7 @@
633
  "repo_name": "scipy",
634
  "repo_description": "SciPy library main repository",
635
  "homepage_link": "https://scipy.org",
636
- "repo_tags": [
637
- "algorithms",
638
- "closember",
639
- "python",
640
- "scientific-computing",
641
- "scipy"
642
- ],
643
  "category": "python library"
644
  },
645
  {
@@ -647,108 +343,51 @@
647
  "repo_name": "numba",
648
  "repo_description": "NumPy aware dynamic Python compiler using LLVM",
649
  "homepage_link": "https://numba.pydata.org/",
650
- "repo_tags": [
651
- "compiler",
652
- "cuda",
653
- "llvm",
654
- "numba",
655
- "numpy",
656
- "parallel",
657
- "python"
658
- ]
659
  },
660
  {
661
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
662
  "repo_name": "torchdendrite",
663
  "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
664
- "homepage_link": "",
665
- "repo_tags": [
666
- "scr-3078"
667
- ],
668
  "category": "machine learning framework"
669
  },
670
  {
671
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
672
  "repo_name": "lightning-thunder",
673
- "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.",
674
- "homepage_link": "",
675
- "repo_tags": []
676
- },
677
- {
678
- "github_repo_link": "https://github.com/Ascend/pytorch",
679
- "repo_name": "pytorch",
680
- "repo_description": "Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch",
681
- "homepage_link": "https://ascend.github.io/docs/",
682
- "repo_tags": [
683
- "ascend",
684
- "deep-learning",
685
- "pytorch"
686
- ]
687
  },
688
  {
689
  "github_repo_link": "https://github.com/pytorch/torchdynamo",
690
  "repo_name": "torchdynamo",
691
- "repo_description": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster.",
692
- "homepage_link": "",
693
- "repo_tags": []
694
- },
695
- {
696
- "github_repo_link": "https://github.com/modular/modular",
697
- "repo_name": "modular",
698
- "repo_description": "The Modular Platform (includes MAX & Mojo)",
699
- "homepage_link": "https://docs.modular.com/",
700
- "repo_tags": [
701
- "ai",
702
- "language",
703
- "machine-learning",
704
- "max",
705
- "modular",
706
- "mojo",
707
- "programming-language"
708
- ]
709
  },
710
  {
711
  "github_repo_link": "https://github.com/microsoft/TileIR",
712
  "repo_name": "TileIR",
713
- "repo_description": null,
714
- "homepage_link": null,
715
- "repo_tags": [],
716
  "category": "dsl"
717
  },
718
  {
719
  "github_repo_link": "https://github.com/pytorch/torchtitan",
720
  "repo_name": "torchtitan",
721
- "repo_description": "A PyTorch native platform for training generative AI models",
722
- "homepage_link": "",
723
- "repo_tags": []
724
  },
725
  {
726
  "github_repo_link": "https://github.com/NVIDIA/cudnn-frontend",
727
  "repo_name": "cudnn-frontend",
728
- "repo_description": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
729
- "homepage_link": null,
730
- "repo_tags": []
731
  },
732
  {
733
  "github_repo_link": "https://github.com/pytorch/ort",
734
  "repo_name": "ort",
735
- "repo_description": "Accelerate PyTorch models with ONNX Runtime",
736
- "homepage_link": "",
737
- "repo_tags": []
738
  },
739
  {
740
  "github_repo_link": "https://github.com/NVIDIA/nccl",
741
  "repo_name": "nccl",
742
  "repo_description": "Optimized primitives for collective multi-GPU communication",
743
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
744
- "repo_tags": [
745
- "communications",
746
- "cpp",
747
- "cuda",
748
- "deep-learning",
749
- "gpu",
750
- "nvidia"
751
- ]
752
  },
753
  {
754
  "github_repo_link": "https://github.com/sgl-project/ome",
@@ -774,48 +413,30 @@
774
  "github_repo_link": "https://github.com/volcengine/verl",
775
  "repo_name": "verl",
776
  "repo_description": "verl: Volcano Engine Reinforcement Learning for LLMs",
777
- "homepage_link": "https://verl.readthedocs.io/en/latest/index.html",
778
- "repo_tags": []
779
  },
780
  {
781
  "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
782
  "repo_name": "neuronx-distributed-inference",
783
- "repo_description": null,
784
- "homepage_link": null,
785
- "repo_tags": [],
786
  "category": "inference engine"
787
  },
788
  {
789
  "github_repo_link": "https://github.com/meta-pytorch/monarch",
790
  "repo_name": "monarch",
791
  "repo_description": "PyTorch Single Controller",
792
- "homepage_link": "https://meta-pytorch.org/monarch",
793
- "repo_tags": []
794
  },
795
  {
796
  "github_repo_link": "https://github.com/ai-dynamo/nixl",
797
  "repo_name": "nixl",
798
- "repo_description": "NVIDIA Inference Xfer Library (NIXL)",
799
- "homepage_link": "",
800
- "repo_tags": []
801
  },
802
  {
803
  "github_repo_link": "https://github.com/LMCache/LMCache",
804
  "repo_name": "LMCache",
805
  "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
806
  "homepage_link": "https://lmcache.ai/",
807
- "repo_tags": [
808
- "amd",
809
- "cuda",
810
- "fast",
811
- "inference",
812
- "kv-cache",
813
- "llm",
814
- "pytorch",
815
- "rocm",
816
- "speed",
817
- "vllm"
818
- ]
819
  },
820
  {
821
  "github_repo_link": "https://github.com/linux-rdma/rdma-core",
@@ -837,85 +458,48 @@
837
  "repo_name": "TensorRT",
838
  "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
839
  "homepage_link": "https://developer.nvidia.com/tensorrt",
840
- "repo_tags": [
841
- "deep-learning",
842
- "gpu-acceleration",
843
- "inference",
844
- "nvidia",
845
- "tensorrt"
846
- ]
847
  },
848
  {
849
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
850
  "repo_name": "FTorch",
851
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
852
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
853
- "repo_tags": [
854
- "deep-learning",
855
- "fortran",
856
- "hacktoberfest",
857
- "interoperability",
858
- "libtorch",
859
- "machine-learning",
860
- "python",
861
- "pytorch",
862
- "torch"
863
- ]
864
  },
865
  {
866
  "github_repo_link": "https://github.com/facebook/hhvm",
867
  "repo_name": "hhvm",
868
  "repo_description": "A virtual machine for executing programs written in Hack.",
869
  "homepage_link": "https://hhvm.com",
870
- "repo_tags": [
871
- "hack",
872
- "hacklang",
873
- "hhvm",
874
- "php"
875
- ]
876
  },
877
  {
878
  "github_repo_link": "https://github.com/vosen/ZLUDA",
879
  "repo_name": "ZLUDA",
880
  "repo_description": "CUDA on non-NVIDIA GPUs",
881
  "homepage_link": "https://vosen.github.io/ZLUDA/",
882
- "repo_tags": [
883
- "cuda",
884
- "rust"
885
- ]
886
  },
887
  {
888
  "github_repo_link": "https://github.com/vtsynergy/CU2CL",
889
  "repo_name": "CU2CL",
890
  "repo_description": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
891
- "homepage_link": "http://chrec.cs.vt.edu/cu2cl",
892
- "repo_tags": []
893
  },
894
  {
895
  "github_repo_link": "https://github.com/pocl/pocl",
896
  "repo_name": "pocl",
897
  "repo_description": "pocl - Portable Computing Language",
898
  "homepage_link": "https://portablecl.org",
899
- "repo_tags": [
900
- "heterogeneous-parallel-programming",
901
- "opencl"
902
- ]
903
  },
904
  {
905
  "github_repo_link": "https://github.com/apache/spark",
906
  "repo_name": "spark",
907
  "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
908
  "homepage_link": "https://spark.apache.org/",
909
- "repo_tags": [
910
- "big-data",
911
- "java",
912
- "jdbc",
913
- "python",
914
- "r",
915
- "scala",
916
- "spark",
917
- "sql"
918
- ]
919
  },
920
  {
921
  "github_repo_link": "https://github.com/codelion/openevolve",
@@ -949,19 +533,13 @@
949
  "repo_name": "hipBLAS",
950
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
951
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
952
- "repo_tags": [
953
- "blas",
954
- "cuda",
955
- "hip",
956
- "rocm"
957
- ]
958
  },
959
  {
960
  "github_repo_link": "https://github.com/ROCm/roctracer",
961
  "repo_name": "roctracer",
962
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-systems repo ",
963
- "homepage_link": "https://github.com/ROCm/rocm-systems",
964
- "repo_tags": []
965
  },
966
  {
967
  "github_repo_link": "https://github.com/huggingface/peft",
@@ -986,34 +564,23 @@
986
  "repo_name": "hip",
987
  "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
988
  "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
989
- "repo_tags": [
990
- "cuda",
991
- "hip",
992
- "hip-kernel-language",
993
- "hip-portability",
994
- "hip-runtime",
995
- "hipify"
996
- ]
997
  },
998
  {
999
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
1000
  "repo_name": "composable_kernel",
1001
  "repo_description": "Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators",
1002
- "homepage_link": "https://rocm.docs.amd.com/projects/composable_kernel/en/latest/",
1003
- "repo_tags": []
1004
  },
1005
  {
1006
  "github_repo_link": "https://github.com/ROCm/aiter",
1007
  "repo_name": "aiter",
1008
- "repo_description": "AI Tensor Engine for ROCm",
1009
- "homepage_link": null,
1010
- "repo_tags": []
1011
  },
1012
  {
1013
  "github_repo_link": "https://github.com/AMDResearch/intelliperf",
1014
  "repo_name": "intelliperf",
1015
  "repo_description": "Automated bottleneck detection and solution orchestration",
1016
- "homepage_link": "",
1017
  "repo_tags": [
1018
  "amd",
1019
  "genai",
@@ -1028,44 +595,32 @@
1028
  {
1029
  "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
1030
  "repo_name": "GEAK-agent",
1031
- "repo_description": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
1032
- "homepage_link": null,
1033
- "repo_tags": []
1034
  },
1035
  {
1036
  "github_repo_link": "https://github.com/AMD-AGI/torchtitan",
1037
  "repo_name": "torchtitan",
1038
- "repo_description": "A PyTorch native platform for training generative AI models",
1039
- "homepage_link": "",
1040
- "repo_tags": []
1041
  },
1042
  {
1043
  "github_repo_link": "https://github.com/AMD-AGI/hipBLASLt",
1044
  "repo_name": "hipBLASLt",
1045
  "repo_description": "hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditional BLAS library",
1046
- "homepage_link": "https://rocm.docs.amd.com/projects/hipBLASLt/en/latest/index.html",
1047
- "repo_tags": []
1048
  },
1049
  {
1050
  "github_repo_link": "https://github.com/AMD-AGI/rocm-torchtitan",
1051
- "repo_name": "rocm-torchtitan",
1052
- "repo_description": null,
1053
- "homepage_link": null,
1054
- "repo_tags": []
1055
  },
1056
  {
1057
  "github_repo_link": "https://github.com/HazyResearch/Megakernels",
1058
  "repo_name": "Megakernels",
1059
- "repo_description": "kernels, of the mega variety",
1060
- "homepage_link": null,
1061
- "repo_tags": []
1062
  },
1063
  {
1064
  "github_repo_link": "https://github.com/huggingface/kernels",
1065
  "repo_name": "kernels",
1066
  "repo_description": "Load compute kernels from the Hub",
1067
- "homepage_link": "",
1068
- "repo_tags": [],
1069
  "category": "kernels"
1070
  },
1071
  {
@@ -1073,7 +628,6 @@
1073
  "repo_name": "tilelang",
1074
  "repo_description": " Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
1075
  "homepage_link": "https://tilelang.com/",
1076
- "repo_tags": [],
1077
  "category": "dsl"
1078
  },
1079
  {
@@ -1081,20 +635,12 @@
1081
  "repo_name": "opencv",
1082
  "repo_description": "Open Source Computer Vision Library",
1083
  "homepage_link": "https://opencv.org",
1084
- "repo_tags": [
1085
- "c-plus-plus",
1086
- "computer-vision",
1087
- "deep-learning",
1088
- "image-processing",
1089
- "opencv"
1090
- ]
1091
  },
1092
  {
1093
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
1094
  "repo_name": "lightning-thunder",
1095
- "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.",
1096
- "homepage_link": "",
1097
- "repo_tags": []
1098
  },
1099
  {
1100
  "github_repo_link": "https://github.com/tracel-ai/burn",
@@ -1126,8 +672,6 @@
1126
  "github_repo_link": "https://github.com/huggingface/kernels-community",
1127
  "repo_name": "kernels-community",
1128
  "repo_description": "Kernel sources for https://huggingface.co/kernels-community",
1129
- "homepage_link": null,
1130
- "repo_tags": [],
1131
  "category": "kernels"
1132
  },
1133
  {
@@ -1135,7 +679,6 @@
1135
  "repo_name": "flashinfer-bench",
1136
  "repo_description": "Building the Virtuous Cycle for AI-driven LLM Systems",
1137
  "homepage_link": "https://bench.flashinfer.ai",
1138
- "repo_tags": [],
1139
  "category": "benchmark"
1140
  },
1141
  {
@@ -1173,28 +716,19 @@
1173
  "repo_name": "KernelBench",
1174
  "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
1175
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
1176
- "repo_tags": [
1177
- "benchmark",
1178
- "codegen",
1179
- "evaluation",
1180
- "gpu"
1181
- ],
1182
  "category": "benchmark"
1183
  },
1184
  {
1185
  "github_repo_link": "https://github.com/thunlp/TritonBench",
1186
  "repo_name": "TritonBench",
1187
  "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
1188
- "homepage_link": "",
1189
- "repo_tags": [],
1190
  "category": "benchmark"
1191
  },
1192
  {
1193
  "github_repo_link": "https://github.com/AutomataLab/cuJSON",
1194
  "repo_name": "cuJSON",
1195
- "repo_description": "cuJSON: A Highly Parallel JSON Parser for GPUs",
1196
- "homepage_link": "",
1197
- "repo_tags": []
1198
  },
1199
  {
1200
  "github_repo_link": "https://github.com/Netflix/metaflow",
@@ -1226,20 +760,14 @@
1226
  },
1227
  {
1228
  "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
1229
- "repo_name": "IMO2025",
1230
- "repo_description": null,
1231
- "homepage_link": null,
1232
- "repo_tags": []
1233
  },
1234
  {
1235
  "github_repo_link": "https://github.com/leanprover/lean4",
1236
  "repo_name": "lean4",
1237
  "repo_description": "Lean 4 programming language and theorem prover",
1238
  "homepage_link": "https://lean-lang.org",
1239
- "repo_tags": [
1240
- "lean",
1241
- "lean4"
1242
- ]
1243
  },
1244
  {
1245
  "github_repo_link": "https://github.com/NVIDIA/warp",
@@ -1260,8 +788,7 @@
1260
  "github_repo_link": "https://github.com/NVIDIA/cuda-python",
1261
  "repo_name": "cuda-python",
1262
  "repo_description": "CUDA Python: Performance meets Productivity",
1263
- "homepage_link": "https://nvidia.github.io/cuda-python/",
1264
- "repo_tags": []
1265
  },
1266
  {
1267
  "github_repo_link": "https://github.com/basetenlabs/truss",
@@ -1288,7 +815,6 @@
1288
  "repo_name": "terminal-bench",
1289
  "repo_description": "A benchmark for LLMs on complicated tasks in the terminal",
1290
  "homepage_link": "https://www.tbench.ai",
1291
- "repo_tags": [],
1292
  "category": "benchmark"
1293
  },
1294
  {
@@ -1296,10 +822,7 @@
1296
  "repo_name": "goose",
1297
  "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
1298
  "homepage_link": "https://block.github.io/goose/",
1299
- "repo_tags": [
1300
- "hacktoberfest",
1301
- "mcp"
1302
- ],
1303
  "category": "agent"
1304
  },
1305
  {
@@ -1333,22 +856,17 @@
1333
  "github_repo_link": "https://github.com/Dao-AILab/quack",
1334
  "repo_name": "quack",
1335
  "repo_description": "A Quirky Assortment of CuTe Kernels",
1336
- "homepage_link": "",
1337
- "repo_tags": [],
1338
  "category": "kernels"
1339
  },
1340
  {
1341
  "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
1342
  "repo_name": "SYCL-Docs",
1343
- "repo_description": "SYCL Open Source Specification",
1344
- "homepage_link": null,
1345
- "repo_tags": []
1346
  },
1347
  {
1348
  "github_repo_link": "https://github.com/triSYCL/triSYCL",
1349
  "repo_name": "triSYCL",
1350
  "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
1351
- "homepage_link": "",
1352
  "repo_tags": [
1353
  "cpp",
1354
  "cpp20",
@@ -1366,17 +884,12 @@
1366
  "repo_name": "pybind11",
1367
  "repo_description": "Seamless operability between C++11 and Python",
1368
  "homepage_link": "https://pybind11.readthedocs.io/",
1369
- "repo_tags": [
1370
- "bindings",
1371
- "python"
1372
- ]
1373
  },
1374
  {
1375
  "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
1376
  "repo_name": "nvcc4jupyter",
1377
  "repo_description": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
1378
- "homepage_link": null,
1379
- "repo_tags": [],
1380
  "category": "compiler"
1381
  },
1382
  {
@@ -1384,11 +897,7 @@
1384
  "repo_name": "rocSOLVER",
1385
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1386
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
1387
- "repo_tags": [
1388
- "lapack",
1389
- "linear-algebra",
1390
- "rocm"
1391
- ]
1392
  },
1393
  {
1394
  "github_repo_link": "https://github.com/ROCm/Tensile",
@@ -1421,72 +930,44 @@
1421
  "repo_name": "rocPRIM",
1422
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
1423
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
1424
- "repo_tags": [
1425
- "amd",
1426
- "cuda",
1427
- "gpu",
1428
- "hip",
1429
- "parallel",
1430
- "primitive",
1431
- "rocm"
1432
- ]
1433
  },
1434
  {
1435
  "github_repo_link": "https://github.com/ROCm/hipCUB",
1436
  "repo_name": "hipCUB",
1437
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
1438
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
1439
- "repo_tags": []
1440
  },
1441
  {
1442
  "github_repo_link": "https://github.com/ROCm/rocFFT",
1443
  "repo_name": "rocFFT",
1444
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1445
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
1446
- "repo_tags": [
1447
- "amd",
1448
- "fast",
1449
- "fft",
1450
- "fourier",
1451
- "gpu",
1452
- "hip",
1453
- "rocm",
1454
- "transform"
1455
- ]
1456
  },
1457
  {
1458
  "github_repo_link": "https://github.com/ROCm/rocSPARSE",
1459
  "repo_name": "rocSPARSE",
1460
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1461
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
1462
- "repo_tags": []
1463
  },
1464
  {
1465
  "github_repo_link": "https://github.com/ROCm/rocRAND",
1466
  "repo_name": "rocRAND",
1467
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
1468
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
1469
- "repo_tags": [
1470
- "cuda",
1471
- "gpu",
1472
- "hip",
1473
- "random",
1474
- "rng",
1475
- "rocm"
1476
- ]
1477
  },
1478
  {
1479
  "github_repo_link": "https://github.com/ROCm/MIOpen",
1480
  "repo_name": "MIOpen",
1481
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1482
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
1483
- "repo_tags": []
1484
  },
1485
  {
1486
  "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
1487
  "repo_name": "lapack",
1488
  "repo_description": "LAPACK development repository",
1489
- "homepage_link": "",
1490
  "repo_tags": [
1491
  "blas",
1492
  "eigenvalues",
@@ -1547,80 +1028,41 @@
1547
  {
1548
  "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
1549
  "repo_name": "OpenCL-SDK",
1550
- "repo_description": "OpenCL SDK",
1551
- "homepage_link": "",
1552
- "repo_tags": []
1553
  },
1554
  {
1555
  "github_repo_link": "https://github.com/meta-llama/synthetic-data-kit",
1556
  "repo_name": "synthetic-data-kit",
1557
  "repo_description": "Tool for generating high quality Synthetic datasets",
1558
  "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
1559
- "repo_tags": [
1560
- "data",
1561
- "generation",
1562
- "llm",
1563
- "python",
1564
- "synthetic"
1565
- ]
1566
  },
1567
  {
1568
  "github_repo_link": "https://github.com/unslothai/unsloth",
1569
  "repo_name": "unsloth",
1570
  "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
1571
  "homepage_link": "https://docs.unsloth.ai/",
1572
- "repo_tags": [
1573
- "agent",
1574
- "deepseek",
1575
- "deepseek-r1",
1576
- "fine-tuning",
1577
- "gemma",
1578
- "gemma3",
1579
- "gpt-oss",
1580
- "llama",
1581
- "llama3",
1582
- "llm",
1583
- "llms",
1584
- "mistral",
1585
- "openai",
1586
- "qwen",
1587
- "qwen3",
1588
- "reinforcement-learning",
1589
- "text-to-speech",
1590
- "tts",
1591
- "unsloth",
1592
- "voice-cloning"
1593
- ]
1594
  },
1595
  {
1596
  "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
1597
  "repo_name": "Vulkan-Docs",
1598
- "repo_description": "The Vulkan API Specification and related tools",
1599
- "homepage_link": null,
1600
- "repo_tags": []
1601
  },
1602
  {
1603
  "github_repo_link": "https://github.com/tensorflow/tflite-micro",
1604
  "repo_name": "tflite-micro",
1605
- "repo_description": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors).",
1606
- "homepage_link": "",
1607
- "repo_tags": []
1608
  },
1609
  {
1610
  "github_repo_link": "https://github.com/Wan-Video/Wan2.2",
1611
  "repo_name": "Wan2.2",
1612
  "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
1613
  "homepage_link": "https://wan.video",
1614
- "repo_tags": [
1615
- "aigc",
1616
- "video-generation"
1617
- ]
1618
  },
1619
  {
1620
  "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
1621
- "repo_name": "Primus-Turbo",
1622
- "repo_description": null,
1623
- "homepage_link": null,
1624
- "repo_tags": []
1625
  }
1626
  ]
 
4
  "repo_name": "pytorch",
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
+ "repo_tag": "machine-learning",
 
 
 
 
 
 
 
 
 
8
  "category": "machine learning framework"
9
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  {
11
  "github_repo_link": "https://github.com/ggml-org/llama.cpp",
12
  "repo_name": "llama.cpp",
13
  "repo_description": "LLM inference in C/C++",
14
+ "repo_tag": "ggml",
 
 
 
15
  "category": "inference engine"
16
  },
17
  {
 
19
  "repo_name": "onnx",
20
  "repo_description": "Open standard for machine learning interoperability",
21
  "homepage_link": "https://onnx.ai/",
22
+ "repo_tags": "deep-learning"
 
 
 
 
 
 
 
 
 
 
 
 
23
  },
24
  {
25
  "github_repo_link": "https://github.com/ray-project/ray",
26
  "repo_name": "ray",
27
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
28
  "homepage_link": "https://ray.io",
29
+ "repo_tag": "deep-learning"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  },
31
  {
32
  "github_repo_link": "https://github.com/vllm-project/vllm",
33
  "repo_name": "vllm",
34
  "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
35
  "homepage_link": "https://docs.vllm.ai",
36
+ "repo_tag": "inference",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  "category": "inference engine"
38
  },
39
  {
 
41
  "repo_name": "ollama",
42
  "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
43
  "homepage_link": "https://ollama.com",
44
+ "repo_tag": "llms",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  "category": "inference engine"
46
  },
47
  {
 
49
  "repo_name": "sglang",
50
  "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
51
  "homepage_link": "https://docs.sglang.ai/",
52
+ "repo_tag": "inference",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  "category": "inference engine"
54
  },
55
+ {
56
+ "github_repo_link": "https://github.com/modular/modular",
57
+ "repo_name": "modular",
58
+ "repo_description": "The Modular Platform (includes MAX & Mojo)",
59
+ "homepage_link": "https://docs.modular.com/",
60
+ "repo_tag": "mojo"
61
+ },
62
  {
63
  "github_repo_link": "https://github.com/pytorch/ao",
64
  "repo_name": "ao",
65
  "repo_description": "PyTorch native quantization and sparsity for training and inference",
66
  "homepage_link": "https://pytorch.org/ao/stable/index.html",
67
+ "repo_tag": "quantization"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
  },
69
  {
70
  "github_repo_link": "https://github.com/triton-lang/triton",
71
  "repo_name": "triton",
72
  "repo_description": "Development repository for the Triton language and compiler",
73
  "homepage_link": "https://triton-lang.org/",
 
74
  "category": "dsl"
75
  },
76
  {
77
  "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
78
  "repo_name": "ThunderKittens",
79
+ "repo_description": "Tile primitives for speedy kernels"
 
 
80
  },
81
  {
82
  "github_repo_link": "https://github.com/gpu-mode/reference-kernels",
83
  "repo_name": "reference-kernels",
84
  "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
85
  "homepage_link": "https://gpumode.com",
86
+ "repo_tag": "gpu",
 
 
 
 
 
87
  "category": "kernels"
88
  },
89
+ {
90
+ "github_repo_link": "https://github.com/pytorch/executorch",
91
+ "repo_name": "executorch",
92
+ "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
93
+ "homepage_link": "https://executorch.ai",
94
+ "repo_tag": "mobile",
95
+ "category": "model compiler"
96
+ },
97
  {
98
  "github_repo_link": "https://github.com/guandeh17/Self-Forcing",
99
  "repo_name": "Self-Forcing",
100
  "repo_description": "Official codebase for \"Self Forcing: Bridging Training and Inference in Autoregressive Video Diffusion\" (NeurIPS 2025 Spotlight)",
101
+
 
 
 
 
 
 
 
 
102
  },
103
  {
104
  "github_repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
105
  "repo_name": "StreamDiffusion",
106
+ "repo_description": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation"
 
 
107
  },
108
  {
109
  "github_repo_link": "https://github.com/comfyanonymous/ComfyUI",
110
  "repo_name": "ComfyUI",
111
  "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
112
  "homepage_link": "https://www.comfy.org/",
113
+ "repo_tags": "stable-diffusion"
 
 
 
 
 
 
 
114
  },
115
  {
116
  "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
117
  "repo_name": "streamv2v",
118
  "repo_description": "Official Pytorch implementation of StreamV2V. ",
119
+ "homepage_link": "https://jeff-liangf.github.io/projects/streamv2v/"
 
120
  },
121
  {
122
  "github_repo_link": "https://github.com/letta-ai/letta",
123
  "repo_name": "letta",
124
  "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
125
  "homepage_link": "https://docs.letta.com/",
126
+ "repo_tag": "ai-agents"
 
 
 
 
 
127
  },
128
  {
129
  "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
130
  "repo_name": "jupyterlab",
131
  "repo_description": "JupyterLab computational environment.",
132
  "homepage_link": "https://jupyterlab.readthedocs.io/",
133
+ "repo_tag": "jupyter",
 
 
 
134
  "category": "ui"
135
  },
136
  {
137
  "github_repo_link": "https://github.com/ROCm/rocm-systems",
138
  "repo_name": "rocm-systems",
139
+ "repo_description": "super repo for rocm systems projects"
 
 
140
  },
141
  {
142
  "github_repo_link": "https://github.com/NVIDIA/cutlass",
143
  "repo_name": "cutlass",
144
  "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
145
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
146
+ "repo_tag": "cuda"
 
 
 
 
 
 
 
 
147
  },
148
  {
149
  "github_repo_link": "https://github.com/pytorch/helion",
150
  "repo_name": "helion",
151
  "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
 
 
152
  "category": "dsl"
153
  },
154
  {
 
156
  "repo_name": "jax",
157
  "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
158
  "homepage_link": "https://docs.jax.dev",
159
+ "repo_tag": "jax"
 
 
160
  },
161
  {
162
  "github_repo_link": "https://github.com/tensorflow/tensorflow",
163
  "repo_name": "tensorflow",
164
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
165
  "homepage_link": "https://tensorflow.org",
166
+ "repo_tag": "deep-learning",
 
 
 
 
 
 
 
 
 
167
  "category": "machine learning framework"
168
  },
169
  {
 
171
  "repo_name": "DeepSpeed",
172
  "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
173
  "homepage_link": "https://www.deepspeed.ai/",
174
+ "repo_tag": "gpu"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
175
  },
176
  {
177
  "github_repo_link": "https://github.com/triton-inference-server/server",
178
  "repo_name": "server",
179
  "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
180
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
181
+ "repo_tag": "inference"
 
 
 
 
 
 
 
 
182
  },
183
  {
184
  "github_repo_link": "https://github.com/ROCm/ROCm",
185
  "repo_name": "ROCm",
186
  "repo_description": "AMD ROCm™ Software - GitHub Home",
187
  "homepage_link": "https://rocm.docs.amd.com",
188
+ "repo_tag": "documentation"
 
 
189
  },
190
  {
191
  "github_repo_link": "https://github.com/llvm/llvm-project",
192
  "repo_name": "llvm-project",
193
  "repo_description": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
194
  "homepage_link": "http://llvm.org",
 
195
  "category": "compiler"
196
  },
197
  {
198
  "github_repo_link": "https://github.com/cwpearson/cupti",
199
  "repo_name": "cupti",
200
  "repo_description": "Profile how CUDA applications create and modify data in memory.",
 
 
201
  "category": "profiler"
202
  },
203
  {
 
205
  "repo_name": "hatchet",
206
  "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
207
  "homepage_link": "https://llnl-hatchet.readthedocs.io",
208
+ "repo_tag": "performance",
 
 
 
 
 
 
 
 
 
 
 
209
  "category": "profiler"
210
  },
211
  {
 
213
  "repo_name": "triton-runner",
214
  "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
215
  "homepage_link": "https://triton-runner.org",
216
+ "repo_tag": "triton"
 
 
 
 
 
217
  },
218
  {
219
  "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
220
  "repo_name": "Triton-distributed",
221
  "repo_description": "Distributed Compiler based on Triton for Parallel Systems",
222
  "homepage_link": "https://triton-distributed.readthedocs.io/en/latest/",
 
223
  "category": "model compiler"
224
  },
225
  {
 
227
  "repo_name": "Liger-Kernel",
228
  "repo_description": "Efficient Triton Kernels for LLM Training",
229
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
230
+ "repo_tag": "triton",
 
 
 
 
 
 
 
 
 
 
 
 
231
  "category": "kernels"
232
  },
233
  {
234
  "github_repo_link": "https://github.com/thunlp/TritonBench",
235
  "repo_name": "TritonBench",
236
  "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
 
 
237
  "category": "benchmark"
238
  },
239
  {
 
241
  "repo_name": "tritonparse",
242
  "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
243
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
244
+ "repo_tag": "triton"
 
 
 
 
 
 
 
 
 
 
245
  },
246
  {
247
  "github_repo_link": "https://github.com/elastic/elasticsearch",
248
  "repo_name": "elasticsearch",
249
  "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
250
  "homepage_link": "https://www.elastic.co/products/elasticsearch",
251
+ "repo_tag": "search-engine",
 
 
 
 
252
  "category": "search engine"
253
  },
254
  {
 
256
  "repo_name": "kubernetes",
257
  "repo_description": "Production-Grade Container Scheduling and Management",
258
  "homepage_link": "https://kubernetes.io",
259
+ "repo_tag": "containers"
 
 
 
 
 
260
  },
261
  {
262
  "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
263
  "repo_name": "modelcontextprotocol",
264
  "repo_description": "Specification and documentation for the Model Context Protocol",
265
+ "homepage_link": "https://modelcontextprotocol.io"
 
266
  },
267
  {
268
  "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
269
  "repo_name": "mcp-agent",
270
  "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
271
+ "repo_tag": "ai-agents"
 
 
 
 
 
 
 
 
 
 
272
  },
273
  {
274
  "github_repo_link": "https://github.com/milvus-io/milvus",
275
  "repo_name": "milvus",
276
  "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
277
  "homepage_link": "https://milvus.io",
278
+ "repo_tag": "vector-search",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
279
  "category": "vector databse"
280
  },
281
  {
 
283
  "repo_name": "RaBitQ",
284
  "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
285
  "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
286
+ "repo_tag": "nearest-neighbor-search"
 
 
 
 
287
  },
288
  {
289
  "github_repo_link": "https://github.com/Airtable/airtable.js",
290
  "repo_name": "airtable.js",
291
+ "repo_description": "Airtable javascript client"
 
 
292
  },
293
  {
294
  "github_repo_link": "https://github.com/mistralai/mistral-inference",
295
  "repo_name": "mistral-inference",
296
  "repo_description": "Official inference library for Mistral models",
297
  "homepage_link": "https://mistral.ai/",
298
+ "repo_tag": "llm-inference",
 
 
 
 
299
  "category": "inference engine"
300
  },
301
  {
 
327
  "repo_name": "numpy",
328
  "repo_description": "The fundamental package for scientific computing with Python.",
329
  "homepage_link": "https://numpy.org",
330
+ "repo_tag": "python",
 
 
 
331
  "category": "python library"
332
  },
333
  {
 
335
  "repo_name": "scipy",
336
  "repo_description": "SciPy library main repository",
337
  "homepage_link": "https://scipy.org",
338
+ "repo_tag": "python",
 
 
 
 
 
 
339
  "category": "python library"
340
  },
341
  {
 
343
  "repo_name": "numba",
344
  "repo_description": "NumPy aware dynamic Python compiler using LLVM",
345
  "homepage_link": "https://numba.pydata.org/",
346
+ "repo_tag": "compiler"
 
 
 
 
 
 
 
 
347
  },
348
  {
349
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
350
  "repo_name": "torchdendrite",
351
  "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
352
+ "repo_tag": "scr-3078",
 
 
 
353
  "category": "machine learning framework"
354
  },
355
  {
356
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
357
  "repo_name": "lightning-thunder",
358
+ "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own."
 
 
 
 
 
 
 
 
 
 
 
 
 
359
  },
360
  {
361
  "github_repo_link": "https://github.com/pytorch/torchdynamo",
362
  "repo_name": "torchdynamo",
363
+ "repo_description": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster."
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
364
  },
365
  {
366
  "github_repo_link": "https://github.com/microsoft/TileIR",
367
  "repo_name": "TileIR",
 
 
 
368
  "category": "dsl"
369
  },
370
  {
371
  "github_repo_link": "https://github.com/pytorch/torchtitan",
372
  "repo_name": "torchtitan",
373
+ "repo_description": "A PyTorch native platform for training generative AI models"
 
 
374
  },
375
  {
376
  "github_repo_link": "https://github.com/NVIDIA/cudnn-frontend",
377
  "repo_name": "cudnn-frontend",
378
+ "repo_description": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it"
 
 
379
  },
380
  {
381
  "github_repo_link": "https://github.com/pytorch/ort",
382
  "repo_name": "ort",
383
+ "repo_description": "Accelerate PyTorch models with ONNX Runtime"
 
 
384
  },
385
  {
386
  "github_repo_link": "https://github.com/NVIDIA/nccl",
387
  "repo_name": "nccl",
388
  "repo_description": "Optimized primitives for collective multi-GPU communication",
389
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
390
+ "repo_tag": "cuda"
 
 
 
 
 
 
 
391
  },
392
  {
393
  "github_repo_link": "https://github.com/sgl-project/ome",
 
413
  "github_repo_link": "https://github.com/volcengine/verl",
414
  "repo_name": "verl",
415
  "repo_description": "verl: Volcano Engine Reinforcement Learning for LLMs",
416
+ "homepage_link": "https://verl.readthedocs.io/en/latest/index.html"
 
417
  },
418
  {
419
  "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
420
  "repo_name": "neuronx-distributed-inference",
 
 
 
421
  "category": "inference engine"
422
  },
423
  {
424
  "github_repo_link": "https://github.com/meta-pytorch/monarch",
425
  "repo_name": "monarch",
426
  "repo_description": "PyTorch Single Controller",
427
+ "homepage_link": "https://meta-pytorch.org/monarch"
 
428
  },
429
  {
430
  "github_repo_link": "https://github.com/ai-dynamo/nixl",
431
  "repo_name": "nixl",
432
+ "repo_description": "NVIDIA Inference Xfer Library (NIXL)"
 
 
433
  },
434
  {
435
  "github_repo_link": "https://github.com/LMCache/LMCache",
436
  "repo_name": "LMCache",
437
  "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
438
  "homepage_link": "https://lmcache.ai/",
439
+ "repo_tag": "inference"
 
 
 
 
 
 
 
 
 
 
 
440
  },
441
  {
442
  "github_repo_link": "https://github.com/linux-rdma/rdma-core",
 
458
  "repo_name": "TensorRT",
459
  "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
460
  "homepage_link": "https://developer.nvidia.com/tensorrt",
461
+ "repo_tag": "inference"
 
 
 
 
 
 
462
  },
463
  {
464
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
465
  "repo_name": "FTorch",
466
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
467
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
468
+ "repo_tag": "deep-learning"
 
 
 
 
 
 
 
 
 
 
469
  },
470
  {
471
  "github_repo_link": "https://github.com/facebook/hhvm",
472
  "repo_name": "hhvm",
473
  "repo_description": "A virtual machine for executing programs written in Hack.",
474
  "homepage_link": "https://hhvm.com",
475
+ "repo_tag": "hack"
 
 
 
 
 
476
  },
477
  {
478
  "github_repo_link": "https://github.com/vosen/ZLUDA",
479
  "repo_name": "ZLUDA",
480
  "repo_description": "CUDA on non-NVIDIA GPUs",
481
  "homepage_link": "https://vosen.github.io/ZLUDA/",
482
+ "repo_tag": "cuda"
 
 
 
483
  },
484
  {
485
  "github_repo_link": "https://github.com/vtsynergy/CU2CL",
486
  "repo_name": "CU2CL",
487
  "repo_description": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
488
+ "homepage_link": "http://chrec.cs.vt.edu/cu2cl"
 
489
  },
490
  {
491
  "github_repo_link": "https://github.com/pocl/pocl",
492
  "repo_name": "pocl",
493
  "repo_description": "pocl - Portable Computing Language",
494
  "homepage_link": "https://portablecl.org",
495
+ "repo_tag": "opencl"
 
 
 
496
  },
497
  {
498
  "github_repo_link": "https://github.com/apache/spark",
499
  "repo_name": "spark",
500
  "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
501
  "homepage_link": "https://spark.apache.org/",
502
+ "repo_tag": "big-data"
 
 
 
 
 
 
 
 
 
503
  },
504
  {
505
  "github_repo_link": "https://github.com/codelion/openevolve",
 
533
  "repo_name": "hipBLAS",
534
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
535
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
536
+ "repo_tag": "hip"
 
 
 
 
 
537
  },
538
  {
539
  "github_repo_link": "https://github.com/ROCm/roctracer",
540
  "repo_name": "roctracer",
541
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-systems repo ",
542
+ "homepage_link": "https://github.com/ROCm/rocm-systems"
 
543
  },
544
  {
545
  "github_repo_link": "https://github.com/huggingface/peft",
 
564
  "repo_name": "hip",
565
  "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
566
  "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
567
+ "repo_tag": "hip"
 
 
 
 
 
 
 
568
  },
569
  {
570
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
571
  "repo_name": "composable_kernel",
572
  "repo_description": "Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators",
573
+ "homepage_link": "https://rocm.docs.amd.com/projects/composable_kernel/en/latest/"
 
574
  },
575
  {
576
  "github_repo_link": "https://github.com/ROCm/aiter",
577
  "repo_name": "aiter",
578
+ "repo_description": "AI Tensor Engine for ROCm"
 
 
579
  },
580
  {
581
  "github_repo_link": "https://github.com/AMDResearch/intelliperf",
582
  "repo_name": "intelliperf",
583
  "repo_description": "Automated bottleneck detection and solution orchestration",
 
584
  "repo_tags": [
585
  "amd",
586
  "genai",
 
595
  {
596
  "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
597
  "repo_name": "GEAK-agent",
598
+ "repo_description": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically."
 
 
599
  },
600
  {
601
  "github_repo_link": "https://github.com/AMD-AGI/torchtitan",
602
  "repo_name": "torchtitan",
603
+ "repo_description": "A PyTorch native platform for training generative AI models"
 
 
604
  },
605
  {
606
  "github_repo_link": "https://github.com/AMD-AGI/hipBLASLt",
607
  "repo_name": "hipBLASLt",
608
  "repo_description": "hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditional BLAS library",
609
+ "homepage_link": "https://rocm.docs.amd.com/projects/hipBLASLt/en/latest/index.html"
 
610
  },
611
  {
612
  "github_repo_link": "https://github.com/AMD-AGI/rocm-torchtitan",
613
+ "repo_name": "rocm-torchtitan"
 
 
 
614
  },
615
  {
616
  "github_repo_link": "https://github.com/HazyResearch/Megakernels",
617
  "repo_name": "Megakernels",
618
+ "repo_description": "kernels, of the mega variety"
 
 
619
  },
620
  {
621
  "github_repo_link": "https://github.com/huggingface/kernels",
622
  "repo_name": "kernels",
623
  "repo_description": "Load compute kernels from the Hub",
 
 
624
  "category": "kernels"
625
  },
626
  {
 
628
  "repo_name": "tilelang",
629
  "repo_description": " Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
630
  "homepage_link": "https://tilelang.com/",
 
631
  "category": "dsl"
632
  },
633
  {
 
635
  "repo_name": "opencv",
636
  "repo_description": "Open Source Computer Vision Library",
637
  "homepage_link": "https://opencv.org",
638
+ "repo_tag": "image-processing"
 
 
 
 
 
 
639
  },
640
  {
641
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
642
  "repo_name": "lightning-thunder",
643
+ "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own."
 
 
644
  },
645
  {
646
  "github_repo_link": "https://github.com/tracel-ai/burn",
 
672
  "github_repo_link": "https://github.com/huggingface/kernels-community",
673
  "repo_name": "kernels-community",
674
  "repo_description": "Kernel sources for https://huggingface.co/kernels-community",
 
 
675
  "category": "kernels"
676
  },
677
  {
 
679
  "repo_name": "flashinfer-bench",
680
  "repo_description": "Building the Virtuous Cycle for AI-driven LLM Systems",
681
  "homepage_link": "https://bench.flashinfer.ai",
 
682
  "category": "benchmark"
683
  },
684
  {
 
716
  "repo_name": "KernelBench",
717
  "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
718
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
719
+ "repo_tags": "benchmark",
 
 
 
 
 
720
  "category": "benchmark"
721
  },
722
  {
723
  "github_repo_link": "https://github.com/thunlp/TritonBench",
724
  "repo_name": "TritonBench",
725
  "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
 
 
726
  "category": "benchmark"
727
  },
728
  {
729
  "github_repo_link": "https://github.com/AutomataLab/cuJSON",
730
  "repo_name": "cuJSON",
731
+ "repo_description": "cuJSON: A Highly Parallel JSON Parser for GPUs"
 
 
732
  },
733
  {
734
  "github_repo_link": "https://github.com/Netflix/metaflow",
 
760
  },
761
  {
762
  "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
763
+ "repo_name": "IMO2025"
 
 
 
764
  },
765
  {
766
  "github_repo_link": "https://github.com/leanprover/lean4",
767
  "repo_name": "lean4",
768
  "repo_description": "Lean 4 programming language and theorem prover",
769
  "homepage_link": "https://lean-lang.org",
770
+ "repo_tag": "lean"
 
 
 
771
  },
772
  {
773
  "github_repo_link": "https://github.com/NVIDIA/warp",
 
788
  "github_repo_link": "https://github.com/NVIDIA/cuda-python",
789
  "repo_name": "cuda-python",
790
  "repo_description": "CUDA Python: Performance meets Productivity",
791
+ "homepage_link": "https://nvidia.github.io/cuda-python/"
 
792
  },
793
  {
794
  "github_repo_link": "https://github.com/basetenlabs/truss",
 
815
  "repo_name": "terminal-bench",
816
  "repo_description": "A benchmark for LLMs on complicated tasks in the terminal",
817
  "homepage_link": "https://www.tbench.ai",
 
818
  "category": "benchmark"
819
  },
820
  {
 
822
  "repo_name": "goose",
823
  "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
824
  "homepage_link": "https://block.github.io/goose/",
825
+ "repo_tag": "mcp",
 
 
 
826
  "category": "agent"
827
  },
828
  {
 
856
  "github_repo_link": "https://github.com/Dao-AILab/quack",
857
  "repo_name": "quack",
858
  "repo_description": "A Quirky Assortment of CuTe Kernels",
 
 
859
  "category": "kernels"
860
  },
861
  {
862
  "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
863
  "repo_name": "SYCL-Docs",
864
+ "repo_description": "SYCL Open Source Specification"
 
 
865
  },
866
  {
867
  "github_repo_link": "https://github.com/triSYCL/triSYCL",
868
  "repo_name": "triSYCL",
869
  "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
 
870
  "repo_tags": [
871
  "cpp",
872
  "cpp20",
 
884
  "repo_name": "pybind11",
885
  "repo_description": "Seamless operability between C++11 and Python",
886
  "homepage_link": "https://pybind11.readthedocs.io/",
887
+ "repo_tag": "bindings"
 
 
 
888
  },
889
  {
890
  "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
891
  "repo_name": "nvcc4jupyter",
892
  "repo_description": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
 
 
893
  "category": "compiler"
894
  },
895
  {
 
897
  "repo_name": "rocSOLVER",
898
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
899
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
900
+ "repo_tag": "rocm"
 
 
 
 
901
  },
902
  {
903
  "github_repo_link": "https://github.com/ROCm/Tensile",
 
930
  "repo_name": "rocPRIM",
931
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
932
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
933
+ "repo_tag": "hip"
 
 
 
 
 
 
 
 
934
  },
935
  {
936
  "github_repo_link": "https://github.com/ROCm/hipCUB",
937
  "repo_name": "hipCUB",
938
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
939
+ "homepage_link": "https://github.com/ROCm/rocm-libraries"
 
940
  },
941
  {
942
  "github_repo_link": "https://github.com/ROCm/rocFFT",
943
  "repo_name": "rocFFT",
944
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
945
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
946
+ "repo_tag": "hip"
 
 
 
 
 
 
 
 
 
947
  },
948
  {
949
  "github_repo_link": "https://github.com/ROCm/rocSPARSE",
950
  "repo_name": "rocSPARSE",
951
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
952
+ "homepage_link": "https://github.com/ROCm/rocm-libraries"
 
953
  },
954
  {
955
  "github_repo_link": "https://github.com/ROCm/rocRAND",
956
  "repo_name": "rocRAND",
957
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
958
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
959
+ "repo_tag": "hip"
 
 
 
 
 
 
 
960
  },
961
  {
962
  "github_repo_link": "https://github.com/ROCm/MIOpen",
963
  "repo_name": "MIOpen",
964
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
965
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
 
966
  },
967
  {
968
  "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
969
  "repo_name": "lapack",
970
  "repo_description": "LAPACK development repository",
 
971
  "repo_tags": [
972
  "blas",
973
  "eigenvalues",
 
1028
  {
1029
  "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
1030
  "repo_name": "OpenCL-SDK",
1031
+ "repo_description": "OpenCL SDK"
 
 
1032
  },
1033
  {
1034
  "github_repo_link": "https://github.com/meta-llama/synthetic-data-kit",
1035
  "repo_name": "synthetic-data-kit",
1036
  "repo_description": "Tool for generating high quality Synthetic datasets",
1037
  "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
1038
+ "repo_tag": "generation"
 
 
 
 
 
 
1039
  },
1040
  {
1041
  "github_repo_link": "https://github.com/unslothai/unsloth",
1042
  "repo_name": "unsloth",
1043
  "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
1044
  "homepage_link": "https://docs.unsloth.ai/",
1045
+ "repo_tag": "unsloth"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1046
  },
1047
  {
1048
  "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
1049
  "repo_name": "Vulkan-Docs",
1050
+ "repo_description": "The Vulkan API Specification and related tools"
 
 
1051
  },
1052
  {
1053
  "github_repo_link": "https://github.com/tensorflow/tflite-micro",
1054
  "repo_name": "tflite-micro",
1055
+ "repo_description": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors)."
 
 
1056
  },
1057
  {
1058
  "github_repo_link": "https://github.com/Wan-Video/Wan2.2",
1059
  "repo_name": "Wan2.2",
1060
  "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
1061
  "homepage_link": "https://wan.video",
1062
+ "repo_tag": "video-generation"
 
 
 
1063
  },
1064
  {
1065
  "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
1066
+ "repo_name": "Primus-Turbo"
 
 
 
1067
  }
1068
  ]