TylerHilbert commited on
Commit
d413412
·
1 Parent(s): e21714c

Renamed github_topic_closest_fit

Browse files
PyTorchConference2025_GithubRepos.json CHANGED
@@ -4,7 +4,7 @@
4
  "repo_name": "pytorch",
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
- "closest_github_tag": "machine-learning",
8
  "category": "machine learning framework"
9
  },
10
  {
@@ -12,7 +12,7 @@
12
  "repo_name": "vllm",
13
  "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
14
  "homepage_link": "https://docs.vllm.ai",
15
- "closest_github_tag": "inference",
16
  "category": "inference engine"
17
  },
18
  {
@@ -20,7 +20,7 @@
20
  "repo_name": "ollama",
21
  "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
22
  "homepage_link": "https://ollama.com",
23
- "closest_github_tag": "llms",
24
  "category": "inference engine"
25
  },
26
  {
@@ -28,14 +28,14 @@
28
  "repo_name": "sglang",
29
  "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
30
  "homepage_link": "https://docs.sglang.ai/",
31
- "closest_github_tag": "inference",
32
  "category": "inference engine"
33
  },
34
  {
35
  "github_repo_link": "https://github.com/ggml-org/llama.cpp",
36
  "repo_name": "llama.cpp",
37
  "repo_description": "LLM inference in C/C++",
38
- "closest_github_tag": "ggml",
39
  "category": "inference engine"
40
  },
41
  {
@@ -68,14 +68,14 @@
68
  "repo_name": "cutlass",
69
  "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
70
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
71
- "closest_github_tag": "cuda"
72
  },
73
  {
74
  "github_repo_link": "https://github.com/tensorflow/tensorflow",
75
  "repo_name": "tensorflow",
76
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
77
  "homepage_link": "https://tensorflow.org",
78
- "closest_github_tag": "deep-learning",
79
  "category": "machine learning framework"
80
  },
81
  {
@@ -88,7 +88,7 @@
88
  "repo_name": "executorch",
89
  "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
90
  "homepage_link": "https://executorch.ai",
91
- "closest_github_tag": "mobile",
92
  "category": "model compiler"
93
  },
94
  {
@@ -96,21 +96,21 @@
96
  "repo_name": "onnx",
97
  "repo_description": "Open standard for machine learning interoperability",
98
  "homepage_link": "https://onnx.ai/",
99
- "closest_github_tag": "deep-learning"
100
  },
101
  {
102
  "github_repo_link": "https://github.com/ray-project/ray",
103
  "repo_name": "ray",
104
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
105
  "homepage_link": "https://ray.io",
106
- "closest_github_tag": "deep-learning"
107
  },
108
  {
109
  "github_repo_link": "https://github.com/jax-ml/jax",
110
  "repo_name": "jax",
111
  "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
112
  "homepage_link": "https://docs.jax.dev",
113
- "closest_github_tag": "jax"
114
  },
115
  {
116
  "github_repo_link": "https://github.com/llvm/llvm-project",
@@ -124,14 +124,14 @@
124
  "repo_name": "TensorRT",
125
  "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
126
  "homepage_link": "https://developer.nvidia.com/tensorrt",
127
- "closest_github_tag": "inference"
128
  },
129
  {
130
  "github_repo_link": "https://github.com/pytorch/ao",
131
  "repo_name": "ao",
132
  "repo_description": "PyTorch native quantization and sparsity for training and inference",
133
  "homepage_link": "https://pytorch.org/ao/stable/index.html",
134
- "closest_github_tag": "quantization"
135
  },
136
  {
137
  "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
@@ -143,14 +143,14 @@
143
  "repo_name": "goose",
144
  "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
145
  "homepage_link": "https://block.github.io/goose/",
146
- "closest_github_tag": "mcp",
147
  "category": "agent"
148
  },
149
  {
150
  "github_repo_link": "https://github.com/codelion/openevolve",
151
  "repo_name": "openevolve",
152
  "repo_description": "Open-source implementation of AlphaEvolve",
153
- "closest_github_tag": "genetic-algorithm"
154
  },
155
  {
156
  "github_repo_link": "https://github.com/volcengine/verl",
@@ -163,7 +163,7 @@
163
  "repo_name": "peft",
164
  "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
165
  "homepage_link": "https://huggingface.co/docs/peft",
166
- "closest_github_tag": "lora"
167
  },
168
  {
169
  "github_repo_link": "https://github.com/Dao-AILab/quack",
@@ -175,34 +175,34 @@
175
  "github_repo_link": "https://github.com/AMDResearch/intelliperf",
176
  "repo_name": "intelliperf",
177
  "repo_description": "Automated bottleneck detection and solution orchestration",
178
- "closest_github_tag": "performance"
179
  },
180
  {
181
  "github_repo_link": "https://github.com/letta-ai/letta",
182
  "repo_name": "letta",
183
  "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
184
  "homepage_link": "https://docs.letta.com/",
185
- "closest_github_tag": "ai-agents"
186
  },
187
  {
188
  "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
189
  "repo_name": "mcp-agent",
190
  "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
191
- "closest_github_tag": "ai-agents"
192
  },
193
  {
194
  "github_repo_link": "https://github.com/modular/modular",
195
  "repo_name": "modular",
196
  "repo_description": "The Modular Platform (includes MAX & Mojo)",
197
  "homepage_link": "https://docs.modular.com/",
198
- "closest_github_tag": "mojo"
199
  },
200
  {
201
  "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
202
  "repo_name": "KernelBench",
203
  "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
204
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
205
- "closest_github_tag": "benchmark",
206
  "category": "benchmark"
207
  },
208
  {
@@ -230,7 +230,7 @@
230
  "repo_name": "SWE-bench",
231
  "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
232
  "homepage_link": "https://www.swebench.com",
233
- "closest_github_tag": "benchmark",
234
  "category": "benchmark"
235
  },
236
  {
@@ -238,7 +238,7 @@
238
  "repo_name": "reference-kernels",
239
  "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
240
  "homepage_link": "https://gpumode.com",
241
- "closest_github_tag": "gpu",
242
  "category": "kernels"
243
  },
244
  {
@@ -246,7 +246,7 @@
246
  "repo_name": "Liger-Kernel",
247
  "repo_description": "Efficient Triton Kernels for LLM Training",
248
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
249
- "closest_github_tag": "triton",
250
  "category": "kernels"
251
  },
252
  {
@@ -266,14 +266,14 @@
266
  "repo_name": "unsloth",
267
  "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
268
  "homepage_link": "https://docs.unsloth.ai/",
269
- "closest_github_tag": "unsloth"
270
  },
271
  {
272
  "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
273
  "repo_name": "jupyterlab",
274
  "repo_description": "JupyterLab computational environment.",
275
  "homepage_link": "https://jupyterlab.readthedocs.io/",
276
- "closest_github_tag": "jupyter",
277
  "category": "ui"
278
  },
279
  {
@@ -286,28 +286,28 @@
286
  "repo_name": "hip",
287
  "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
288
  "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
289
- "closest_github_tag": "hip"
290
  },
291
  {
292
  "github_repo_link": "https://github.com/ROCm/ROCm",
293
  "repo_name": "ROCm",
294
  "repo_description": "AMD ROCm™ Software - GitHub Home",
295
  "homepage_link": "https://rocm.docs.amd.com",
296
- "closest_github_tag": "documentation"
297
  },
298
  {
299
  "github_repo_link": "https://github.com/ROCm/omnitrace",
300
  "repo_name": "omnitrace",
301
  "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
302
  "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
303
- "closest_github_tag": "performance-analysis"
304
  },
305
  {
306
  "github_repo_link": "https://github.com/vosen/ZLUDA",
307
  "repo_name": "ZLUDA",
308
  "repo_description": "CUDA on non-NVIDIA GPUs",
309
  "homepage_link": "https://vosen.github.io/ZLUDA/",
310
- "closest_github_tag": "cuda"
311
  },
312
  {
313
  "github_repo_link": "https://github.com/vtsynergy/CU2CL",
@@ -320,7 +320,7 @@
320
  "repo_name": "pocl",
321
  "repo_description": "pocl - Portable Computing Language",
322
  "homepage_link": "https://portablecl.org",
323
- "closest_github_tag": "opencl"
324
  },
325
  {
326
  "github_repo_link": "https://github.com/cwpearson/cupti",
@@ -333,7 +333,7 @@
333
  "repo_name": "hatchet",
334
  "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
335
  "homepage_link": "https://llnl-hatchet.readthedocs.io",
336
- "closest_github_tag": "performance",
337
  "category": "profiler"
338
  },
339
  {
@@ -341,7 +341,7 @@
341
  "repo_name": "triton-runner",
342
  "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
343
  "homepage_link": "https://triton-runner.org",
344
- "closest_github_tag": "triton"
345
  },
346
  {
347
  "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
@@ -355,14 +355,14 @@
355
  "repo_name": "tritonparse",
356
  "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
357
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
358
- "closest_github_tag": "triton"
359
  },
360
  {
361
  "github_repo_link": "https://github.com/numpy/numpy",
362
  "repo_name": "numpy",
363
  "repo_description": "The fundamental package for scientific computing with Python.",
364
  "homepage_link": "https://numpy.org",
365
- "closest_github_tag": "python",
366
  "category": "python library"
367
  },
368
  {
@@ -370,7 +370,7 @@
370
  "repo_name": "scipy",
371
  "repo_description": "SciPy library main repository",
372
  "homepage_link": "https://scipy.org",
373
- "closest_github_tag": "python",
374
  "category": "python library"
375
  },
376
  {
@@ -378,7 +378,7 @@
378
  "repo_name": "numba",
379
  "repo_description": "NumPy aware dynamic Python compiler using LLVM",
380
  "homepage_link": "https://numba.pydata.org/",
381
- "closest_github_tag": "compiler"
382
  },
383
  {
384
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
@@ -395,7 +395,7 @@
395
  "repo_name": "nccl",
396
  "repo_description": "Optimized primitives for collective multi-GPU communication",
397
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
398
- "closest_github_tag": "cuda"
399
  },
400
  {
401
  "github_repo_link": "https://github.com/ai-dynamo/nixl",
@@ -418,7 +418,7 @@
418
  "repo_name": "ComfyUI",
419
  "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
420
  "homepage_link": "https://www.comfy.org/",
421
- "closest_github_tag": "stable-diffusion"
422
  },
423
  {
424
  "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
@@ -431,21 +431,21 @@
431
  "repo_name": "DeepSpeed",
432
  "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
433
  "homepage_link": "https://www.deepspeed.ai/",
434
- "closest_github_tag": "gpu"
435
  },
436
  {
437
  "github_repo_link": "https://github.com/triton-inference-server/server",
438
  "repo_name": "server",
439
  "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
440
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
441
- "closest_github_tag": "inference"
442
  },
443
  {
444
  "github_repo_link": "https://github.com/elastic/elasticsearch",
445
  "repo_name": "elasticsearch",
446
  "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
447
  "homepage_link": "https://www.elastic.co/products/elasticsearch",
448
- "closest_github_tag": "search-engine",
449
  "category": "search engine"
450
  },
451
  {
@@ -453,7 +453,7 @@
453
  "repo_name": "kubernetes",
454
  "repo_description": "Production-Grade Container Scheduling and Management",
455
  "homepage_link": "https://kubernetes.io",
456
- "closest_github_tag": "containers"
457
  },
458
  {
459
  "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
@@ -466,7 +466,7 @@
466
  "repo_name": "milvus",
467
  "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
468
  "homepage_link": "https://milvus.io",
469
- "closest_github_tag": "vector-search",
470
  "category": "vector databse"
471
  },
472
  {
@@ -474,7 +474,7 @@
474
  "repo_name": "RaBitQ",
475
  "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
476
  "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
477
- "closest_github_tag": "nearest-neighbor-search"
478
  },
479
  {
480
  "github_repo_link": "https://github.com/Airtable/airtable.js",
@@ -486,7 +486,7 @@
486
  "repo_name": "mistral-inference",
487
  "repo_description": "Official inference library for Mistral models",
488
  "homepage_link": "https://mistral.ai/",
489
- "closest_github_tag": "llm-inference",
490
  "category": "inference engine"
491
  },
492
  {
@@ -494,13 +494,13 @@
494
  "repo_name": "dstack",
495
  "repo_description": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs—across hyperscalers, neoclouds, or on-prem.",
496
  "homepage_link": "https://dstack.ai",
497
- "closest_github_tag": "orchestration"
498
  },
499
  {
500
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
501
  "repo_name": "torchdendrite",
502
  "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
503
- "closest_github_tag": "scr-3078",
504
  "category": "machine learning framework"
505
  },
506
  {
@@ -523,7 +523,7 @@
523
  "repo_name": "ome",
524
  "repo_description": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
525
  "homepage_link": "http://docs.sglang.ai/ome/",
526
- "closest_github_tag": "k8s"
527
  },
528
  {
529
  "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
@@ -541,34 +541,34 @@
541
  "repo_name": "LMCache",
542
  "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
543
  "homepage_link": "https://lmcache.ai/",
544
- "closest_github_tag": "inference"
545
  },
546
  {
547
  "github_repo_link": "https://github.com/linux-rdma/rdma-core",
548
  "repo_name": "rdma-core",
549
  "repo_description": "RDMA core userspace libraries and daemons",
550
- "closest_github_tag": "linux-kernel"
551
  },
552
  {
553
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
554
  "repo_name": "FTorch",
555
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
556
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
557
- "closest_github_tag": "deep-learning"
558
  },
559
  {
560
  "github_repo_link": "https://github.com/facebook/hhvm",
561
  "repo_name": "hhvm",
562
  "repo_description": "A virtual machine for executing programs written in Hack.",
563
  "homepage_link": "https://hhvm.com",
564
- "closest_github_tag": "hack"
565
  },
566
  {
567
  "github_repo_link": "https://github.com/apache/spark",
568
  "repo_name": "spark",
569
  "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
570
  "homepage_link": "https://spark.apache.org/",
571
- "closest_github_tag": "big-data"
572
  },
573
  {
574
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
@@ -606,7 +606,7 @@
606
  "repo_name": "opencv",
607
  "repo_description": "Open Source Computer Vision Library",
608
  "homepage_link": "https://opencv.org",
609
- "closest_github_tag": "image-processing"
610
  },
611
  {
612
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
@@ -618,21 +618,21 @@
618
  "repo_name": "burn",
619
  "repo_description": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
620
  "homepage_link": "https://burn.dev",
621
- "closest_github_tag": "machine-learning"
622
  },
623
  {
624
  "github_repo_link": "https://github.com/OSC/ondemand",
625
  "repo_name": "ondemand",
626
  "repo_description": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
627
  "homepage_link": "https://openondemand.org/",
628
- "closest_github_tag": "hpc"
629
  },
630
  {
631
  "github_repo_link": "https://github.com/flashinfer-ai/flashinfer",
632
  "repo_name": "flashinfer",
633
  "repo_description": "FlashInfer: Kernel Library for LLM Serving",
634
  "homepage_link": "https://flashinfer.ai",
635
- "closest_github_tag": "attention"
636
  },
637
  {
638
  "github_repo_link": "https://github.com/AutomataLab/cuJSON",
@@ -644,7 +644,7 @@
644
  "repo_name": "metaflow",
645
  "repo_description": "Build, Manage and Deploy AI/ML Systems",
646
  "homepage_link": "https://metaflow.org",
647
- "closest_github_tag": "machine-learning"
648
  },
649
  {
650
  "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
@@ -655,14 +655,14 @@
655
  "repo_name": "lean4",
656
  "repo_description": "Lean 4 programming language and theorem prover",
657
  "homepage_link": "https://lean-lang.org",
658
- "closest_github_tag": "lean"
659
  },
660
  {
661
  "github_repo_link": "https://github.com/NVIDIA/warp",
662
  "repo_name": "warp",
663
  "repo_description": "A Python framework for accelerated simulation, data generation and spatial computing.",
664
  "homepage_link": "https://nvidia.github.io/warp/",
665
- "closest_github_tag": "gpu"
666
  },
667
  {
668
  "github_repo_link": "https://github.com/NVIDIA/cuda-python",
@@ -675,14 +675,14 @@
675
  "repo_name": "truss",
676
  "repo_description": "The simplest way to serve AI/ML models in production",
677
  "homepage_link": "https://truss.baseten.co",
678
- "closest_github_tag": "machine-learning"
679
  },
680
  {
681
  "github_repo_link": "https://github.com/kvcache-ai/Mooncake",
682
  "repo_name": "Mooncake",
683
  "repo_description": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
684
  "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
685
- "closest_github_tag": "inference"
686
  },
687
  {
688
  "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
@@ -693,14 +693,14 @@
693
  "github_repo_link": "https://github.com/triSYCL/triSYCL",
694
  "repo_name": "triSYCL",
695
  "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
696
- "closest_github_tag": "opencl"
697
  },
698
  {
699
  "github_repo_link": "https://github.com/pybind/pybind11",
700
  "repo_name": "pybind11",
701
  "repo_description": "Seamless operability between C++11 and Python",
702
  "homepage_link": "https://pybind11.readthedocs.io/",
703
- "closest_github_tag": "bindings"
704
  },
705
  {
706
  "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
@@ -712,14 +712,14 @@
712
  "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
713
  "repo_name": "lapack",
714
  "repo_description": "LAPACK development repository",
715
- "closest_github_tag": "linear-algebra"
716
  },
717
  {
718
  "github_repo_link": "https://github.com/ccache/ccache",
719
  "repo_name": "ccache",
720
  "repo_description": "ccache – a fast compiler cache",
721
  "homepage_link": "https://ccache.dev",
722
- "closest_github_tag": "compiler",
723
  "category": "compiler"
724
  },
725
  {
@@ -732,7 +732,7 @@
732
  "repo_name": "synthetic-data-kit",
733
  "repo_description": "Tool for generating high quality Synthetic datasets",
734
  "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
735
- "closest_github_tag": "generation"
736
  },
737
  {
738
  "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
@@ -749,7 +749,7 @@
749
  "repo_name": "Wan2.2",
750
  "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
751
  "homepage_link": "https://wan.video",
752
- "closest_github_tag": "video-generation"
753
  },
754
  {
755
  "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
@@ -760,7 +760,7 @@
760
  "repo_name": "hipBLAS",
761
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
762
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
763
- "closest_github_tag": "hip"
764
  },
765
  {
766
  "github_repo_link": "https://github.com/ROCm/roctracer",
@@ -773,21 +773,21 @@
773
  "repo_name": "rocSOLVER",
774
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
775
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
776
- "closest_github_tag": "rocm"
777
  },
778
  {
779
  "github_repo_link": "https://github.com/ROCm/Tensile",
780
  "repo_name": "Tensile",
781
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
782
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
783
- "closest_github_tag": "gpu"
784
  },
785
  {
786
  "github_repo_link": "https://github.com/ROCm/rocPRIM",
787
  "repo_name": "rocPRIM",
788
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
789
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
790
- "closest_github_tag": "hip"
791
  },
792
  {
793
  "github_repo_link": "https://github.com/ROCm/hipCUB",
@@ -800,7 +800,7 @@
800
  "repo_name": "rocFFT",
801
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
802
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
803
- "closest_github_tag": "hip"
804
  },
805
  {
806
  "github_repo_link": "https://github.com/ROCm/rocSPARSE",
@@ -813,7 +813,7 @@
813
  "repo_name": "rocRAND",
814
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
815
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
816
- "closest_github_tag": "hip"
817
  },
818
  {
819
  "github_repo_link": "https://github.com/ROCm/MIOpen",
 
4
  "repo_name": "pytorch",
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
+ "github_topic_closest_fit": "machine-learning",
8
  "category": "machine learning framework"
9
  },
10
  {
 
12
  "repo_name": "vllm",
13
  "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
14
  "homepage_link": "https://docs.vllm.ai",
15
+ "github_topic_closest_fit": "inference",
16
  "category": "inference engine"
17
  },
18
  {
 
20
  "repo_name": "ollama",
21
  "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
22
  "homepage_link": "https://ollama.com",
23
+ "github_topic_closest_fit": "llms",
24
  "category": "inference engine"
25
  },
26
  {
 
28
  "repo_name": "sglang",
29
  "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
30
  "homepage_link": "https://docs.sglang.ai/",
31
+ "github_topic_closest_fit": "inference",
32
  "category": "inference engine"
33
  },
34
  {
35
  "github_repo_link": "https://github.com/ggml-org/llama.cpp",
36
  "repo_name": "llama.cpp",
37
  "repo_description": "LLM inference in C/C++",
38
+ "github_topic_closest_fit": "ggml",
39
  "category": "inference engine"
40
  },
41
  {
 
68
  "repo_name": "cutlass",
69
  "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
70
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
71
+ "github_topic_closest_fit": "cuda"
72
  },
73
  {
74
  "github_repo_link": "https://github.com/tensorflow/tensorflow",
75
  "repo_name": "tensorflow",
76
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
77
  "homepage_link": "https://tensorflow.org",
78
+ "github_topic_closest_fit": "deep-learning",
79
  "category": "machine learning framework"
80
  },
81
  {
 
88
  "repo_name": "executorch",
89
  "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
90
  "homepage_link": "https://executorch.ai",
91
+ "github_topic_closest_fit": "mobile",
92
  "category": "model compiler"
93
  },
94
  {
 
96
  "repo_name": "onnx",
97
  "repo_description": "Open standard for machine learning interoperability",
98
  "homepage_link": "https://onnx.ai/",
99
+ "github_topic_closest_fit": "deep-learning"
100
  },
101
  {
102
  "github_repo_link": "https://github.com/ray-project/ray",
103
  "repo_name": "ray",
104
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
105
  "homepage_link": "https://ray.io",
106
+ "github_topic_closest_fit": "deep-learning"
107
  },
108
  {
109
  "github_repo_link": "https://github.com/jax-ml/jax",
110
  "repo_name": "jax",
111
  "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
112
  "homepage_link": "https://docs.jax.dev",
113
+ "github_topic_closest_fit": "jax"
114
  },
115
  {
116
  "github_repo_link": "https://github.com/llvm/llvm-project",
 
124
  "repo_name": "TensorRT",
125
  "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
126
  "homepage_link": "https://developer.nvidia.com/tensorrt",
127
+ "github_topic_closest_fit": "inference"
128
  },
129
  {
130
  "github_repo_link": "https://github.com/pytorch/ao",
131
  "repo_name": "ao",
132
  "repo_description": "PyTorch native quantization and sparsity for training and inference",
133
  "homepage_link": "https://pytorch.org/ao/stable/index.html",
134
+ "github_topic_closest_fit": "quantization"
135
  },
136
  {
137
  "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
 
143
  "repo_name": "goose",
144
  "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
145
  "homepage_link": "https://block.github.io/goose/",
146
+ "github_topic_closest_fit": "mcp",
147
  "category": "agent"
148
  },
149
  {
150
  "github_repo_link": "https://github.com/codelion/openevolve",
151
  "repo_name": "openevolve",
152
  "repo_description": "Open-source implementation of AlphaEvolve",
153
+ "github_topic_closest_fit": "genetic-algorithm"
154
  },
155
  {
156
  "github_repo_link": "https://github.com/volcengine/verl",
 
163
  "repo_name": "peft",
164
  "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
165
  "homepage_link": "https://huggingface.co/docs/peft",
166
+ "github_topic_closest_fit": "lora"
167
  },
168
  {
169
  "github_repo_link": "https://github.com/Dao-AILab/quack",
 
175
  "github_repo_link": "https://github.com/AMDResearch/intelliperf",
176
  "repo_name": "intelliperf",
177
  "repo_description": "Automated bottleneck detection and solution orchestration",
178
+ "github_topic_closest_fit": "performance"
179
  },
180
  {
181
  "github_repo_link": "https://github.com/letta-ai/letta",
182
  "repo_name": "letta",
183
  "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
184
  "homepage_link": "https://docs.letta.com/",
185
+ "github_topic_closest_fit": "ai-agents"
186
  },
187
  {
188
  "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
189
  "repo_name": "mcp-agent",
190
  "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
191
+ "github_topic_closest_fit": "ai-agents"
192
  },
193
  {
194
  "github_repo_link": "https://github.com/modular/modular",
195
  "repo_name": "modular",
196
  "repo_description": "The Modular Platform (includes MAX & Mojo)",
197
  "homepage_link": "https://docs.modular.com/",
198
+ "github_topic_closest_fit": "mojo"
199
  },
200
  {
201
  "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
202
  "repo_name": "KernelBench",
203
  "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
204
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
205
+ "github_topic_closest_fit": "benchmark",
206
  "category": "benchmark"
207
  },
208
  {
 
230
  "repo_name": "SWE-bench",
231
  "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
232
  "homepage_link": "https://www.swebench.com",
233
+ "github_topic_closest_fit": "benchmark",
234
  "category": "benchmark"
235
  },
236
  {
 
238
  "repo_name": "reference-kernels",
239
  "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
240
  "homepage_link": "https://gpumode.com",
241
+ "github_topic_closest_fit": "gpu",
242
  "category": "kernels"
243
  },
244
  {
 
246
  "repo_name": "Liger-Kernel",
247
  "repo_description": "Efficient Triton Kernels for LLM Training",
248
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
249
+ "github_topic_closest_fit": "triton",
250
  "category": "kernels"
251
  },
252
  {
 
266
  "repo_name": "unsloth",
267
  "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
268
  "homepage_link": "https://docs.unsloth.ai/",
269
+ "github_topic_closest_fit": "unsloth"
270
  },
271
  {
272
  "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
273
  "repo_name": "jupyterlab",
274
  "repo_description": "JupyterLab computational environment.",
275
  "homepage_link": "https://jupyterlab.readthedocs.io/",
276
+ "github_topic_closest_fit": "jupyter",
277
  "category": "ui"
278
  },
279
  {
 
286
  "repo_name": "hip",
287
  "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
288
  "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
289
+ "github_topic_closest_fit": "hip"
290
  },
291
  {
292
  "github_repo_link": "https://github.com/ROCm/ROCm",
293
  "repo_name": "ROCm",
294
  "repo_description": "AMD ROCm™ Software - GitHub Home",
295
  "homepage_link": "https://rocm.docs.amd.com",
296
+ "github_topic_closest_fit": "documentation"
297
  },
298
  {
299
  "github_repo_link": "https://github.com/ROCm/omnitrace",
300
  "repo_name": "omnitrace",
301
  "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
302
  "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
303
+ "github_topic_closest_fit": "performance-analysis"
304
  },
305
  {
306
  "github_repo_link": "https://github.com/vosen/ZLUDA",
307
  "repo_name": "ZLUDA",
308
  "repo_description": "CUDA on non-NVIDIA GPUs",
309
  "homepage_link": "https://vosen.github.io/ZLUDA/",
310
+ "github_topic_closest_fit": "cuda"
311
  },
312
  {
313
  "github_repo_link": "https://github.com/vtsynergy/CU2CL",
 
320
  "repo_name": "pocl",
321
  "repo_description": "pocl - Portable Computing Language",
322
  "homepage_link": "https://portablecl.org",
323
+ "github_topic_closest_fit": "opencl"
324
  },
325
  {
326
  "github_repo_link": "https://github.com/cwpearson/cupti",
 
333
  "repo_name": "hatchet",
334
  "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
335
  "homepage_link": "https://llnl-hatchet.readthedocs.io",
336
+ "github_topic_closest_fit": "performance",
337
  "category": "profiler"
338
  },
339
  {
 
341
  "repo_name": "triton-runner",
342
  "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
343
  "homepage_link": "https://triton-runner.org",
344
+ "github_topic_closest_fit": "triton"
345
  },
346
  {
347
  "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
 
355
  "repo_name": "tritonparse",
356
  "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
357
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
358
+ "github_topic_closest_fit": "triton"
359
  },
360
  {
361
  "github_repo_link": "https://github.com/numpy/numpy",
362
  "repo_name": "numpy",
363
  "repo_description": "The fundamental package for scientific computing with Python.",
364
  "homepage_link": "https://numpy.org",
365
+ "github_topic_closest_fit": "python",
366
  "category": "python library"
367
  },
368
  {
 
370
  "repo_name": "scipy",
371
  "repo_description": "SciPy library main repository",
372
  "homepage_link": "https://scipy.org",
373
+ "github_topic_closest_fit": "python",
374
  "category": "python library"
375
  },
376
  {
 
378
  "repo_name": "numba",
379
  "repo_description": "NumPy aware dynamic Python compiler using LLVM",
380
  "homepage_link": "https://numba.pydata.org/",
381
+ "github_topic_closest_fit": "compiler"
382
  },
383
  {
384
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
 
395
  "repo_name": "nccl",
396
  "repo_description": "Optimized primitives for collective multi-GPU communication",
397
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
398
+ "github_topic_closest_fit": "cuda"
399
  },
400
  {
401
  "github_repo_link": "https://github.com/ai-dynamo/nixl",
 
418
  "repo_name": "ComfyUI",
419
  "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
420
  "homepage_link": "https://www.comfy.org/",
421
+ "github_topic_closest_fit": "stable-diffusion"
422
  },
423
  {
424
  "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
 
431
  "repo_name": "DeepSpeed",
432
  "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
433
  "homepage_link": "https://www.deepspeed.ai/",
434
+ "github_topic_closest_fit": "gpu"
435
  },
436
  {
437
  "github_repo_link": "https://github.com/triton-inference-server/server",
438
  "repo_name": "server",
439
  "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
440
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
441
+ "github_topic_closest_fit": "inference"
442
  },
443
  {
444
  "github_repo_link": "https://github.com/elastic/elasticsearch",
445
  "repo_name": "elasticsearch",
446
  "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
447
  "homepage_link": "https://www.elastic.co/products/elasticsearch",
448
+ "github_topic_closest_fit": "search-engine",
449
  "category": "search engine"
450
  },
451
  {
 
453
  "repo_name": "kubernetes",
454
  "repo_description": "Production-Grade Container Scheduling and Management",
455
  "homepage_link": "https://kubernetes.io",
456
+ "github_topic_closest_fit": "containers"
457
  },
458
  {
459
  "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
 
466
  "repo_name": "milvus",
467
  "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
468
  "homepage_link": "https://milvus.io",
469
+ "github_topic_closest_fit": "vector-search",
470
  "category": "vector databse"
471
  },
472
  {
 
474
  "repo_name": "RaBitQ",
475
  "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
476
  "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
477
+ "github_topic_closest_fit": "nearest-neighbor-search"
478
  },
479
  {
480
  "github_repo_link": "https://github.com/Airtable/airtable.js",
 
486
  "repo_name": "mistral-inference",
487
  "repo_description": "Official inference library for Mistral models",
488
  "homepage_link": "https://mistral.ai/",
489
+ "github_topic_closest_fit": "llm-inference",
490
  "category": "inference engine"
491
  },
492
  {
 
494
  "repo_name": "dstack",
495
  "repo_description": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs—across hyperscalers, neoclouds, or on-prem.",
496
  "homepage_link": "https://dstack.ai",
497
+ "github_topic_closest_fit": "orchestration"
498
  },
499
  {
500
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
501
  "repo_name": "torchdendrite",
502
  "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
503
+ "github_topic_closest_fit": "scr-3078",
504
  "category": "machine learning framework"
505
  },
506
  {
 
523
  "repo_name": "ome",
524
  "repo_description": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
525
  "homepage_link": "http://docs.sglang.ai/ome/",
526
+ "github_topic_closest_fit": "k8s"
527
  },
528
  {
529
  "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
 
541
  "repo_name": "LMCache",
542
  "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
543
  "homepage_link": "https://lmcache.ai/",
544
+ "github_topic_closest_fit": "inference"
545
  },
546
  {
547
  "github_repo_link": "https://github.com/linux-rdma/rdma-core",
548
  "repo_name": "rdma-core",
549
  "repo_description": "RDMA core userspace libraries and daemons",
550
+ "github_topic_closest_fit": "linux-kernel"
551
  },
552
  {
553
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
554
  "repo_name": "FTorch",
555
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
556
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
557
+ "github_topic_closest_fit": "deep-learning"
558
  },
559
  {
560
  "github_repo_link": "https://github.com/facebook/hhvm",
561
  "repo_name": "hhvm",
562
  "repo_description": "A virtual machine for executing programs written in Hack.",
563
  "homepage_link": "https://hhvm.com",
564
+ "github_topic_closest_fit": "hack"
565
  },
566
  {
567
  "github_repo_link": "https://github.com/apache/spark",
568
  "repo_name": "spark",
569
  "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
570
  "homepage_link": "https://spark.apache.org/",
571
+ "github_topic_closest_fit": "big-data"
572
  },
573
  {
574
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
 
606
  "repo_name": "opencv",
607
  "repo_description": "Open Source Computer Vision Library",
608
  "homepage_link": "https://opencv.org",
609
+ "github_topic_closest_fit": "image-processing"
610
  },
611
  {
612
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
 
618
  "repo_name": "burn",
619
  "repo_description": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
620
  "homepage_link": "https://burn.dev",
621
+ "github_topic_closest_fit": "machine-learning"
622
  },
623
  {
624
  "github_repo_link": "https://github.com/OSC/ondemand",
625
  "repo_name": "ondemand",
626
  "repo_description": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
627
  "homepage_link": "https://openondemand.org/",
628
+ "github_topic_closest_fit": "hpc"
629
  },
630
  {
631
  "github_repo_link": "https://github.com/flashinfer-ai/flashinfer",
632
  "repo_name": "flashinfer",
633
  "repo_description": "FlashInfer: Kernel Library for LLM Serving",
634
  "homepage_link": "https://flashinfer.ai",
635
+ "github_topic_closest_fit": "attention"
636
  },
637
  {
638
  "github_repo_link": "https://github.com/AutomataLab/cuJSON",
 
644
  "repo_name": "metaflow",
645
  "repo_description": "Build, Manage and Deploy AI/ML Systems",
646
  "homepage_link": "https://metaflow.org",
647
+ "github_topic_closest_fit": "machine-learning"
648
  },
649
  {
650
  "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
 
655
  "repo_name": "lean4",
656
  "repo_description": "Lean 4 programming language and theorem prover",
657
  "homepage_link": "https://lean-lang.org",
658
+ "github_topic_closest_fit": "lean"
659
  },
660
  {
661
  "github_repo_link": "https://github.com/NVIDIA/warp",
662
  "repo_name": "warp",
663
  "repo_description": "A Python framework for accelerated simulation, data generation and spatial computing.",
664
  "homepage_link": "https://nvidia.github.io/warp/",
665
+ "github_topic_closest_fit": "gpu"
666
  },
667
  {
668
  "github_repo_link": "https://github.com/NVIDIA/cuda-python",
 
675
  "repo_name": "truss",
676
  "repo_description": "The simplest way to serve AI/ML models in production",
677
  "homepage_link": "https://truss.baseten.co",
678
+ "github_topic_closest_fit": "machine-learning"
679
  },
680
  {
681
  "github_repo_link": "https://github.com/kvcache-ai/Mooncake",
682
  "repo_name": "Mooncake",
683
  "repo_description": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
684
  "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
685
+ "github_topic_closest_fit": "inference"
686
  },
687
  {
688
  "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
 
693
  "github_repo_link": "https://github.com/triSYCL/triSYCL",
694
  "repo_name": "triSYCL",
695
  "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
696
+ "github_topic_closest_fit": "opencl"
697
  },
698
  {
699
  "github_repo_link": "https://github.com/pybind/pybind11",
700
  "repo_name": "pybind11",
701
  "repo_description": "Seamless operability between C++11 and Python",
702
  "homepage_link": "https://pybind11.readthedocs.io/",
703
+ "github_topic_closest_fit": "bindings"
704
  },
705
  {
706
  "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
 
712
  "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
713
  "repo_name": "lapack",
714
  "repo_description": "LAPACK development repository",
715
+ "github_topic_closest_fit": "linear-algebra"
716
  },
717
  {
718
  "github_repo_link": "https://github.com/ccache/ccache",
719
  "repo_name": "ccache",
720
  "repo_description": "ccache – a fast compiler cache",
721
  "homepage_link": "https://ccache.dev",
722
+ "github_topic_closest_fit": "compiler",
723
  "category": "compiler"
724
  },
725
  {
 
732
  "repo_name": "synthetic-data-kit",
733
  "repo_description": "Tool for generating high quality Synthetic datasets",
734
  "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
735
+ "github_topic_closest_fit": "generation"
736
  },
737
  {
738
  "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
 
749
  "repo_name": "Wan2.2",
750
  "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
751
  "homepage_link": "https://wan.video",
752
+ "github_topic_closest_fit": "video-generation"
753
  },
754
  {
755
  "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
 
760
  "repo_name": "hipBLAS",
761
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
762
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
763
+ "github_topic_closest_fit": "hip"
764
  },
765
  {
766
  "github_repo_link": "https://github.com/ROCm/roctracer",
 
773
  "repo_name": "rocSOLVER",
774
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
775
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
776
+ "github_topic_closest_fit": "rocm"
777
  },
778
  {
779
  "github_repo_link": "https://github.com/ROCm/Tensile",
780
  "repo_name": "Tensile",
781
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
782
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
783
+ "github_topic_closest_fit": "gpu"
784
  },
785
  {
786
  "github_repo_link": "https://github.com/ROCm/rocPRIM",
787
  "repo_name": "rocPRIM",
788
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
789
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
790
+ "github_topic_closest_fit": "hip"
791
  },
792
  {
793
  "github_repo_link": "https://github.com/ROCm/hipCUB",
 
800
  "repo_name": "rocFFT",
801
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
802
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
803
+ "github_topic_closest_fit": "hip"
804
  },
805
  {
806
  "github_repo_link": "https://github.com/ROCm/rocSPARSE",
 
813
  "repo_name": "rocRAND",
814
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
815
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
816
+ "github_topic_closest_fit": "hip"
817
  },
818
  {
819
  "github_repo_link": "https://github.com/ROCm/MIOpen",