TylerHilbert commited on
Commit
9de99b9
·
verified ·
1 Parent(s): 9469924

Added PyTorch Conference 2025 GitHub Repos dataset

Browse files
Files changed (1) hide show
  1. PyTorchConference2025_GithubRepos.json +1722 -0
PyTorchConference2025_GithubRepos.json ADDED
@@ -0,0 +1,1722 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "github_repo_link": "https://github.com/pytorch/pytorch",
4
+ "repo_name": "pytorch",
5
+ "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
+ "homepage_link": "https://pytorch.org",
7
+ "repo_tags": [
8
+ "autograd",
9
+ "deep-learning",
10
+ "gpu",
11
+ "machine-learning",
12
+ "neural-network",
13
+ "numpy",
14
+ "python",
15
+ "tensor"
16
+ ]
17
+ },
18
+ {
19
+ "github_repo_link": "https://github.com/pytorch/executorch",
20
+ "repo_name": "executorch",
21
+ "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
22
+ "homepage_link": "https://executorch.ai",
23
+ "repo_tags": [
24
+ "deep-learning",
25
+ "embedded",
26
+ "gpu",
27
+ "machine-learning",
28
+ "mobile",
29
+ "neural-network",
30
+ "tensor"
31
+ ]
32
+ },
33
+ {
34
+ "github_repo_link": "https://github.com/ggml-org/llama.cpp",
35
+ "repo_name": "llama.cpp",
36
+ "repo_description": "LLM inference in C/C++",
37
+ "homepage_link": "",
38
+ "repo_tags": [
39
+ "ggml"
40
+ ]
41
+ },
42
+ {
43
+ "github_repo_link": "https://github.com/tensorflow/tflite-micro",
44
+ "repo_name": "tflite-micro",
45
+ "repo_description": "Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors).",
46
+ "homepage_link": "",
47
+ "repo_tags": []
48
+ },
49
+ {
50
+ "github_repo_link": "https://github.com/onnx/onnx",
51
+ "repo_name": "onnx",
52
+ "repo_description": "Open standard for machine learning interoperability",
53
+ "homepage_link": "https://onnx.ai/",
54
+ "repo_tags": [
55
+ "deep-learning",
56
+ "deep-neural-networks",
57
+ "dnn",
58
+ "keras",
59
+ "machine-learning",
60
+ "ml",
61
+ "neural-network",
62
+ "onnx",
63
+ "pytorch",
64
+ "scikit-learn",
65
+ "tensorflow"
66
+ ]
67
+ },
68
+ {
69
+ "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
70
+ "repo_name": "OpenCL-SDK",
71
+ "repo_description": "OpenCL SDK",
72
+ "homepage_link": "",
73
+ "repo_tags": []
74
+ },
75
+ {
76
+ "github_repo_link": "https://github.com/meta-llama/synthetic-data-kit",
77
+ "repo_name": "synthetic-data-kit",
78
+ "repo_description": "Tool for generating high quality Synthetic datasets",
79
+ "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
80
+ "repo_tags": [
81
+ "data",
82
+ "generation",
83
+ "llm",
84
+ "python",
85
+ "synthetic"
86
+ ]
87
+ },
88
+ {
89
+ "github_repo_link": "https://github.com/unslothai/unsloth",
90
+ "repo_name": "unsloth",
91
+ "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
92
+ "homepage_link": "https://docs.unsloth.ai/",
93
+ "repo_tags": [
94
+ "agent",
95
+ "deepseek",
96
+ "deepseek-r1",
97
+ "fine-tuning",
98
+ "gemma",
99
+ "gemma3",
100
+ "gpt-oss",
101
+ "llama",
102
+ "llama3",
103
+ "llm",
104
+ "llms",
105
+ "mistral",
106
+ "openai",
107
+ "qwen",
108
+ "qwen3",
109
+ "reinforcement-learning",
110
+ "text-to-speech",
111
+ "tts",
112
+ "unsloth",
113
+ "voice-cloning"
114
+ ]
115
+ },
116
+ {
117
+ "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
118
+ "repo_name": "Vulkan-Docs",
119
+ "repo_description": "The Vulkan API Specification and related tools",
120
+ "homepage_link": null,
121
+ "repo_tags": []
122
+ },
123
+ {
124
+ "github_repo_link": "https://github.com/ray-project/ray",
125
+ "repo_name": "ray",
126
+ "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
127
+ "homepage_link": "https://ray.io",
128
+ "repo_tags": [
129
+ "data-science",
130
+ "deep-learning",
131
+ "deployment",
132
+ "distributed",
133
+ "hyperparameter-optimization",
134
+ "hyperparameter-search",
135
+ "large-language-models",
136
+ "llm",
137
+ "llm-inference",
138
+ "llm-serving",
139
+ "machine-learning",
140
+ "optimization",
141
+ "parallel",
142
+ "python",
143
+ "pytorch",
144
+ "ray",
145
+ "reinforcement-learning",
146
+ "rllib",
147
+ "serving",
148
+ "tensorflow"
149
+ ]
150
+ },
151
+ {
152
+ "github_repo_link": "https://github.com/vllm-project/vllm",
153
+ "repo_name": "vllm",
154
+ "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
155
+ "homepage_link": "https://docs.vllm.ai",
156
+ "repo_tags": [
157
+ "amd",
158
+ "blackwell",
159
+ "cuda",
160
+ "deepseek",
161
+ "deepseek-v3",
162
+ "gpt",
163
+ "gpt-oss",
164
+ "inference",
165
+ "kimi",
166
+ "llama",
167
+ "llm",
168
+ "llm-serving",
169
+ "model-serving",
170
+ "moe",
171
+ "openai",
172
+ "pytorch",
173
+ "qwen",
174
+ "qwen3",
175
+ "tpu",
176
+ "transformer"
177
+ ]
178
+ },
179
+ {
180
+ "github_repo_link": "https://github.com/ollama/ollama",
181
+ "repo_name": "ollama",
182
+ "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
183
+ "homepage_link": "https://ollama.com",
184
+ "repo_tags": [
185
+ "deepseek",
186
+ "gemma",
187
+ "gemma3",
188
+ "gemma3n",
189
+ "go",
190
+ "golang",
191
+ "gpt-oss",
192
+ "llama",
193
+ "llama2",
194
+ "llama3",
195
+ "llava",
196
+ "llm",
197
+ "llms",
198
+ "mistral",
199
+ "ollama",
200
+ "phi4",
201
+ "qwen"
202
+ ]
203
+ },
204
+ {
205
+ "github_repo_link": "https://github.com/sgl-project/sglang",
206
+ "repo_name": "sglang",
207
+ "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
208
+ "homepage_link": "https://docs.sglang.ai/",
209
+ "repo_tags": [
210
+ "blackwell",
211
+ "cuda",
212
+ "deepseek",
213
+ "deepseek-r1",
214
+ "deepseek-v3",
215
+ "deepseek-v3-2",
216
+ "gpt-oss",
217
+ "inference",
218
+ "kimi",
219
+ "llama",
220
+ "llama3",
221
+ "llava",
222
+ "llm",
223
+ "llm-serving",
224
+ "moe",
225
+ "openai",
226
+ "pytorch",
227
+ "qwen3",
228
+ "transformer",
229
+ "vlm"
230
+ ]
231
+ },
232
+ {
233
+ "github_repo_link": "https://github.com/Wan-Video/Wan2.2",
234
+ "repo_name": "Wan2.2",
235
+ "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
236
+ "homepage_link": "https://wan.video",
237
+ "repo_tags": [
238
+ "aigc",
239
+ "video-generation"
240
+ ]
241
+ },
242
+ {
243
+ "github_repo_link": "https://github.com/pytorch/ao",
244
+ "repo_name": "ao",
245
+ "repo_description": "PyTorch native quantization and sparsity for training and inference",
246
+ "homepage_link": "https://pytorch.org/ao/stable/index.html",
247
+ "repo_tags": [
248
+ "brrr",
249
+ "cuda",
250
+ "dtypes",
251
+ "float8",
252
+ "inference",
253
+ "llama",
254
+ "mx",
255
+ "offloading",
256
+ "optimizer",
257
+ "pytorch",
258
+ "quantization",
259
+ "sparsity",
260
+ "training",
261
+ "transformer"
262
+ ]
263
+ },
264
+ {
265
+ "github_repo_link": "https://github.com/triton-lang/triton",
266
+ "repo_name": "triton",
267
+ "repo_description": "Development repository for the Triton language and compiler",
268
+ "homepage_link": "https://triton-lang.org/",
269
+ "repo_tags": []
270
+ },
271
+ {
272
+ "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
273
+ "repo_name": "Primus-Turbo",
274
+ "repo_description": null,
275
+ "homepage_link": null,
276
+ "repo_tags": []
277
+ },
278
+ {
279
+ "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
280
+ "repo_name": "ThunderKittens",
281
+ "repo_description": "Tile primitives for speedy kernels",
282
+ "homepage_link": null,
283
+ "repo_tags": []
284
+ },
285
+ {
286
+ "github_repo_link": "https://github.com/gpu-mode/reference-kernels",
287
+ "repo_name": "reference-kernels",
288
+ "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
289
+ "homepage_link": "https://gpumode.com",
290
+ "repo_tags": [
291
+ "cuda",
292
+ "gpu",
293
+ "leaderboard",
294
+ "triton"
295
+ ]
296
+ },
297
+ {
298
+ "github_repo_link": "https://github.com/guandeh17/Self-Forcing",
299
+ "repo_name": "Self-Forcing",
300
+ "repo_description": "Official codebase for \"Self Forcing: Bridging Training and Inference in Autoregressive Video Diffusion\" (NeurIPS 2025 Spotlight)",
301
+ "homepage_link": "",
302
+ "repo_tags": []
303
+ },
304
+ {
305
+ "github_repo_link": "https://github.com/chenfengxu714/StreamDiffusionV2",
306
+ "repo_name": "StreamDiffusionV2",
307
+ "repo_description": "StreamDiffusion, Live Stream APP",
308
+ "homepage_link": "",
309
+ "repo_tags": []
310
+ },
311
+ {
312
+ "github_repo_link": "https://github.com/cumulo-autumn/StreamDiffusion",
313
+ "repo_name": "StreamDiffusion",
314
+ "repo_description": "StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation",
315
+ "homepage_link": "",
316
+ "repo_tags": []
317
+ },
318
+ {
319
+ "github_repo_link": "https://github.com/comfyanonymous/ComfyUI",
320
+ "repo_name": "ComfyUI",
321
+ "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
322
+ "homepage_link": "https://www.comfy.org/",
323
+ "repo_tags": [
324
+ "ai",
325
+ "comfy",
326
+ "comfyui",
327
+ "python",
328
+ "pytorch",
329
+ "stable-diffusion"
330
+ ]
331
+ },
332
+ {
333
+ "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
334
+ "repo_name": "streamv2v",
335
+ "repo_description": "Official Pytorch implementation of StreamV2V. ",
336
+ "homepage_link": "https://jeff-liangf.github.io/projects/streamv2v/",
337
+ "repo_tags": []
338
+ },
339
+ {
340
+ "github_repo_link": "https://github.com/letta-ai/letta",
341
+ "repo_name": "letta",
342
+ "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
343
+ "homepage_link": "https://docs.letta.com/",
344
+ "repo_tags": [
345
+ "ai",
346
+ "ai-agents",
347
+ "llm",
348
+ "llm-agent"
349
+ ]
350
+ },
351
+ {
352
+ "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
353
+ "repo_name": "jupyterlab",
354
+ "repo_description": "JupyterLab computational environment.",
355
+ "homepage_link": "https://jupyterlab.readthedocs.io/",
356
+ "repo_tags": [
357
+ "jupyter",
358
+ "jupyterlab"
359
+ ]
360
+ },
361
+ {
362
+ "github_repo_link": "https://github.com/ROCm/rocm-systems",
363
+ "repo_name": "rocm-systems",
364
+ "repo_description": "super repo for rocm systems projects",
365
+ "homepage_link": "",
366
+ "repo_tags": []
367
+ },
368
+ {
369
+ "github_repo_link": "https://github.com/NVIDIA/cutlass",
370
+ "repo_name": "cutlass",
371
+ "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
372
+ "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
373
+ "repo_tags": [
374
+ "cpp",
375
+ "cuda",
376
+ "deep-learning",
377
+ "deep-learning-library",
378
+ "gpu",
379
+ "nvidia",
380
+ "python"
381
+ ]
382
+ },
383
+ {
384
+ "github_repo_link": "https://github.com/pytorch/helion",
385
+ "repo_name": "helion",
386
+ "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
387
+ "homepage_link": null,
388
+ "repo_tags": []
389
+ },
390
+ {
391
+ "github_repo_link": "https://github.com/jax-ml/jax",
392
+ "repo_name": "jax",
393
+ "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
394
+ "homepage_link": "https://docs.jax.dev",
395
+ "repo_tags": [
396
+ "jax"
397
+ ]
398
+ },
399
+ {
400
+ "github_repo_link": "https://github.com/tensorflow/tensorflow",
401
+ "repo_name": "tensorflow",
402
+ "repo_description": "An Open Source Machine Learning Framework for Everyone",
403
+ "homepage_link": "https://tensorflow.org",
404
+ "repo_tags": [
405
+ "deep-learning",
406
+ "deep-neural-networks",
407
+ "distributed",
408
+ "machine-learning",
409
+ "ml",
410
+ "neural-network",
411
+ "python",
412
+ "tensorflow"
413
+ ]
414
+ },
415
+ {
416
+ "github_repo_link": "https://github.com/deepspeedai/DeepSpeed",
417
+ "repo_name": "DeepSpeed",
418
+ "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
419
+ "homepage_link": "https://www.deepspeed.ai/",
420
+ "repo_tags": [
421
+ "billion-parameters",
422
+ "compression",
423
+ "data-parallelism",
424
+ "deep-learning",
425
+ "gpu",
426
+ "inference",
427
+ "machine-learning",
428
+ "mixture-of-experts",
429
+ "model-parallelism",
430
+ "pipeline-parallelism",
431
+ "pytorch",
432
+ "trillion-parameters",
433
+ "zero"
434
+ ]
435
+ },
436
+ {
437
+ "github_repo_link": "https://github.com/triton-inference-server/server",
438
+ "repo_name": "server",
439
+ "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
440
+ "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
441
+ "repo_tags": [
442
+ "cloud",
443
+ "datacenter",
444
+ "deep-learning",
445
+ "edge",
446
+ "gpu",
447
+ "inference",
448
+ "machine-learning"
449
+ ]
450
+ },
451
+ {
452
+ "github_repo_link": "https://github.com/ROCm/ROCm",
453
+ "repo_name": "ROCm",
454
+ "repo_description": "AMD ROCm™ Software - GitHub Home",
455
+ "homepage_link": "https://rocm.docs.amd.com",
456
+ "repo_tags": [
457
+ "documentation"
458
+ ]
459
+ },
460
+ {
461
+ "github_repo_link": "https://github.com/llvm/llvm-project",
462
+ "repo_name": "llvm-project",
463
+ "repo_description": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
464
+ "homepage_link": "http://llvm.org",
465
+ "repo_tags": []
466
+ },
467
+ {
468
+ "github_repo_link": "https://github.com/cwpearson/cupti",
469
+ "repo_name": "cupti",
470
+ "repo_description": "Profile how CUDA applications create and modify data in memory.",
471
+ "homepage_link": "",
472
+ "repo_tags": []
473
+ },
474
+ {
475
+ "github_repo_link": "https://github.com/LLNL/hatchet",
476
+ "repo_name": "hatchet",
477
+ "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
478
+ "homepage_link": "https://llnl-hatchet.readthedocs.io",
479
+ "repo_tags": [
480
+ "comparative-analysis",
481
+ "data-analytics",
482
+ "graphs",
483
+ "hierarchical-data",
484
+ "hpc",
485
+ "performance",
486
+ "performance-analysis",
487
+ "python",
488
+ "radiuss",
489
+ "trees"
490
+ ]
491
+ },
492
+ {
493
+ "github_repo_link": "https://github.com/toyaix/triton-runner",
494
+ "repo_name": "triton-runner",
495
+ "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
496
+ "homepage_link": "https://triton-runner.org",
497
+ "repo_tags": [
498
+ "ai-infra",
499
+ "cuda",
500
+ "tools",
501
+ "triton"
502
+ ]
503
+ },
504
+ {
505
+ "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
506
+ "repo_name": "Triton-distributed",
507
+ "repo_description": "Distributed Compiler based on Triton for Parallel Systems",
508
+ "homepage_link": "https://triton-distributed.readthedocs.io/en/latest/",
509
+ "repo_tags": []
510
+ },
511
+ {
512
+ "github_repo_link": "https://github.com/linkedin/Liger-Kernel",
513
+ "repo_name": "Liger-Kernel",
514
+ "repo_description": "Efficient Triton Kernels for LLM Training",
515
+ "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
516
+ "repo_tags": [
517
+ "finetuning",
518
+ "gemma2",
519
+ "hacktoberfest",
520
+ "llama",
521
+ "llama3",
522
+ "llm-training",
523
+ "llms",
524
+ "mistral",
525
+ "phi3",
526
+ "triton",
527
+ "triton-kernels"
528
+ ]
529
+ },
530
+ {
531
+ "github_repo_link": "https://github.com/thunlp/TritonBench",
532
+ "repo_name": "TritonBench",
533
+ "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
534
+ "homepage_link": "",
535
+ "repo_tags": []
536
+ },
537
+ {
538
+ "github_repo_link": "https://github.com/meta-pytorch/tritonparse",
539
+ "repo_name": "tritonparse",
540
+ "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
541
+ "homepage_link": "https://meta-pytorch.org/tritonparse/",
542
+ "repo_tags": [
543
+ "compiler",
544
+ "debugging",
545
+ "gpu-kernel",
546
+ "interactive-visualization",
547
+ "ir-analysis",
548
+ "ir-visualization",
549
+ "pytorch",
550
+ "structured-logging",
551
+ "triton"
552
+ ]
553
+ },
554
+ {
555
+ "github_repo_link": "https://github.com/elastic/elasticsearch",
556
+ "repo_name": "elasticsearch",
557
+ "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
558
+ "homepage_link": "https://www.elastic.co/products/elasticsearch",
559
+ "repo_tags": [
560
+ "elasticsearch",
561
+ "java",
562
+ "search-engine"
563
+ ]
564
+ },
565
+ {
566
+ "github_repo_link": "https://github.com/kubernetes/kubernetes",
567
+ "repo_name": "kubernetes",
568
+ "repo_description": "Production-Grade Container Scheduling and Management",
569
+ "homepage_link": "https://kubernetes.io",
570
+ "repo_tags": [
571
+ "cncf",
572
+ "containers",
573
+ "go",
574
+ "kubernetes"
575
+ ]
576
+ },
577
+ {
578
+ "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
579
+ "repo_name": "modelcontextprotocol",
580
+ "repo_description": "Specification and documentation for the Model Context Protocol",
581
+ "homepage_link": "https://modelcontextprotocol.io",
582
+ "repo_tags": []
583
+ },
584
+ {
585
+ "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
586
+ "repo_name": "mcp-agent",
587
+ "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
588
+ "homepage_link": "",
589
+ "repo_tags": [
590
+ "agents",
591
+ "ai",
592
+ "ai-agents",
593
+ "llm",
594
+ "llms",
595
+ "mcp",
596
+ "model-context-protocol",
597
+ "python"
598
+ ]
599
+ },
600
+ {
601
+ "github_repo_link": "https://github.com/milvus-io/milvus",
602
+ "repo_name": "milvus",
603
+ "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
604
+ "homepage_link": "https://milvus.io",
605
+ "repo_tags": [
606
+ "anns",
607
+ "cloud-native",
608
+ "diskann",
609
+ "distributed",
610
+ "embedding-database",
611
+ "embedding-similarity",
612
+ "embedding-store",
613
+ "faiss",
614
+ "golang",
615
+ "hnsw",
616
+ "image-search",
617
+ "llm",
618
+ "nearest-neighbor-search",
619
+ "rag",
620
+ "vector-database",
621
+ "vector-search",
622
+ "vector-similarity",
623
+ "vector-store"
624
+ ]
625
+ },
626
+ {
627
+ "github_repo_link": "https://github.com/gaoj0017/RaBitQ",
628
+ "repo_name": "RaBitQ",
629
+ "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
630
+ "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
631
+ "repo_tags": [
632
+ "high-dimensional-vectors",
633
+ "nearest-neighbor-search",
634
+ "quantization"
635
+ ]
636
+ },
637
+ {
638
+ "github_repo_link": "https://github.com/Airtable/airtable.js",
639
+ "repo_name": "airtable.js",
640
+ "repo_description": "Airtable javascript client",
641
+ "homepage_link": null,
642
+ "repo_tags": []
643
+ },
644
+ {
645
+ "github_repo_link": "https://github.com/mistralai/mistral-inference",
646
+ "repo_name": "mistral-inference",
647
+ "repo_description": "Official inference library for Mistral models",
648
+ "homepage_link": "https://mistral.ai/",
649
+ "repo_tags": [
650
+ "llm",
651
+ "llm-inference",
652
+ "mistralai"
653
+ ]
654
+ },
655
+ {
656
+ "github_repo_link": "https://github.com/dstackai/dstack",
657
+ "repo_name": "dstack",
658
+ "repo_description": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs—across hyperscalers, neoclouds, or on-prem.",
659
+ "homepage_link": "https://dstack.ai",
660
+ "repo_tags": [
661
+ "amd",
662
+ "cloud",
663
+ "containers",
664
+ "docker",
665
+ "fine-tuning",
666
+ "gpu",
667
+ "inference",
668
+ "k8s",
669
+ "kubernetes",
670
+ "llms",
671
+ "machine-learning",
672
+ "nvidia",
673
+ "orchestration",
674
+ "python",
675
+ "slurm",
676
+ "training"
677
+ ]
678
+ },
679
+ {
680
+ "github_repo_link": "https://github.com/numpy/numpy",
681
+ "repo_name": "numpy",
682
+ "repo_description": "The fundamental package for scientific computing with Python.",
683
+ "homepage_link": "https://numpy.org",
684
+ "repo_tags": [
685
+ "numpy",
686
+ "python"
687
+ ]
688
+ },
689
+ {
690
+ "github_repo_link": "https://github.com/scipy/scipy",
691
+ "repo_name": "scipy",
692
+ "repo_description": "SciPy library main repository",
693
+ "homepage_link": "https://scipy.org",
694
+ "repo_tags": [
695
+ "algorithms",
696
+ "closember",
697
+ "python",
698
+ "scientific-computing",
699
+ "scipy"
700
+ ]
701
+ },
702
+ {
703
+ "github_repo_link": "https://github.com/numba/numba",
704
+ "repo_name": "numba",
705
+ "repo_description": "NumPy aware dynamic Python compiler using LLVM",
706
+ "homepage_link": "https://numba.pydata.org/",
707
+ "repo_tags": [
708
+ "compiler",
709
+ "cuda",
710
+ "llvm",
711
+ "numba",
712
+ "numpy",
713
+ "parallel",
714
+ "python"
715
+ ]
716
+ },
717
+ {
718
+ "github_repo_link": "https://github.com/torvalds/linux",
719
+ "repo_name": "linux",
720
+ "repo_description": "Linux kernel source tree",
721
+ "homepage_link": "",
722
+ "repo_tags": []
723
+ },
724
+ {
725
+ "github_repo_link": "https://github.com/google-deepmind/gemma",
726
+ "repo_name": "gemma",
727
+ "repo_description": "Gemma open-weight LLM library, from Google DeepMind",
728
+ "homepage_link": "https://gemma-llm.readthedocs.io",
729
+ "repo_tags": []
730
+ },
731
+ {
732
+ "github_repo_link": "https://github.com/sandialabs/torchdendrite",
733
+ "repo_name": "torchdendrite",
734
+ "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
735
+ "homepage_link": "",
736
+ "repo_tags": [
737
+ "scr-3078"
738
+ ]
739
+ },
740
+ {
741
+ "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
742
+ "repo_name": "lightning-thunder",
743
+ "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.",
744
+ "homepage_link": "",
745
+ "repo_tags": []
746
+ },
747
+ {
748
+ "github_repo_link": "https://github.com/Ascend/pytorch",
749
+ "repo_name": "pytorch",
750
+ "repo_description": "Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch",
751
+ "homepage_link": "https://ascend.github.io/docs/",
752
+ "repo_tags": [
753
+ "ascend",
754
+ "deep-learning",
755
+ "pytorch"
756
+ ]
757
+ },
758
+ {
759
+ "github_repo_link": "https://github.com/pytorch/torchdynamo",
760
+ "repo_name": "torchdynamo",
761
+ "repo_description": "A Python-level JIT compiler designed to make unmodified PyTorch programs faster.",
762
+ "homepage_link": "",
763
+ "repo_tags": []
764
+ },
765
+ {
766
+ "github_repo_link": "https://github.com/modular/modular",
767
+ "repo_name": "modular",
768
+ "repo_description": "The Modular Platform (includes MAX & Mojo)",
769
+ "homepage_link": "https://docs.modular.com/",
770
+ "repo_tags": [
771
+ "ai",
772
+ "language",
773
+ "machine-learning",
774
+ "max",
775
+ "modular",
776
+ "mojo",
777
+ "programming-language"
778
+ ]
779
+ },
780
+ {
781
+ "github_repo_link": "https://github.com/microsoft/TileIR",
782
+ "repo_name": "TileIR",
783
+ "repo_description": null,
784
+ "homepage_link": null,
785
+ "repo_tags": []
786
+ },
787
+ {
788
+ "github_repo_link": "https://github.com/pytorch/torchtitan",
789
+ "repo_name": "torchtitan",
790
+ "repo_description": "A PyTorch native platform for training generative AI models",
791
+ "homepage_link": "",
792
+ "repo_tags": []
793
+ },
794
+ {
795
+ "github_repo_link": "https://github.com/NVIDIA/cudnn-frontend",
796
+ "repo_name": "cudnn-frontend",
797
+ "repo_description": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
798
+ "homepage_link": null,
799
+ "repo_tags": []
800
+ },
801
+ {
802
+ "github_repo_link": "https://github.com/pytorch/ort",
803
+ "repo_name": "ort",
804
+ "repo_description": "Accelerate PyTorch models with ONNX Runtime",
805
+ "homepage_link": "",
806
+ "repo_tags": []
807
+ },
808
+ {
809
+ "github_repo_link": "https://github.com/NVIDIA/nccl",
810
+ "repo_name": "nccl",
811
+ "repo_description": "Optimized primitives for collective multi-GPU communication",
812
+ "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
813
+ "repo_tags": [
814
+ "communications",
815
+ "cpp",
816
+ "cuda",
817
+ "deep-learning",
818
+ "gpu",
819
+ "nvidia"
820
+ ]
821
+ },
822
+ {
823
+ "github_repo_link": "https://github.com/sgl-project/ome",
824
+ "repo_name": "ome",
825
+ "repo_description": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
826
+ "homepage_link": "http://docs.sglang.ai/ome/",
827
+ "repo_tags": [
828
+ "deepseek",
829
+ "k8s",
830
+ "kimi-k2",
831
+ "llama",
832
+ "llm",
833
+ "llm-inference",
834
+ "model-as-a-service",
835
+ "model-serving",
836
+ "multi-node-kubernetes",
837
+ "oracle-cloud",
838
+ "sgalng",
839
+ "sglang"
840
+ ]
841
+ },
842
+ {
843
+ "github_repo_link": "https://github.com/volcengine/verl",
844
+ "repo_name": "verl",
845
+ "repo_description": "verl: Volcano Engine Reinforcement Learning for LLMs",
846
+ "homepage_link": "https://verl.readthedocs.io/en/latest/index.html",
847
+ "repo_tags": []
848
+ },
849
+ {
850
+ "github_repo_link": "https://github.com/aws-neuron/neuronx-distributed-inference",
851
+ "repo_name": "neuronx-distributed-inference",
852
+ "repo_description": null,
853
+ "homepage_link": null,
854
+ "repo_tags": []
855
+ },
856
+ {
857
+ "github_repo_link": "https://github.com/meta-pytorch/monarch",
858
+ "repo_name": "monarch",
859
+ "repo_description": "PyTorch Single Controller",
860
+ "homepage_link": "https://meta-pytorch.org/monarch",
861
+ "repo_tags": []
862
+ },
863
+ {
864
+ "github_repo_link": "https://github.com/ai-dynamo/nixl",
865
+ "repo_name": "nixl",
866
+ "repo_description": "NVIDIA Inference Xfer Library (NIXL)",
867
+ "homepage_link": "",
868
+ "repo_tags": []
869
+ },
870
+ {
871
+ "github_repo_link": "https://github.com/LMCache/LMCache",
872
+ "repo_name": "LMCache",
873
+ "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
874
+ "homepage_link": "https://lmcache.ai/",
875
+ "repo_tags": [
876
+ "amd",
877
+ "cuda",
878
+ "fast",
879
+ "inference",
880
+ "kv-cache",
881
+ "llm",
882
+ "pytorch",
883
+ "rocm",
884
+ "speed",
885
+ "vllm"
886
+ ]
887
+ },
888
+ {
889
+ "github_repo_link": "https://github.com/linux-rdma/rdma-core",
890
+ "repo_name": "rdma-core",
891
+ "repo_description": "RDMA core userspace libraries and daemons",
892
+ "homepage_link": null,
893
+ "repo_tags": [
894
+ "infiniband",
895
+ "iwarp",
896
+ "kernel-rdma-drivers",
897
+ "linux-kernel",
898
+ "rdma",
899
+ "roce",
900
+ "userspace-libraries"
901
+ ]
902
+ },
903
+ {
904
+ "github_repo_link": "https://github.com/NVIDIA/TensorRT",
905
+ "repo_name": "TensorRT",
906
+ "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
907
+ "homepage_link": "https://developer.nvidia.com/tensorrt",
908
+ "repo_tags": [
909
+ "deep-learning",
910
+ "gpu-acceleration",
911
+ "inference",
912
+ "nvidia",
913
+ "tensorrt"
914
+ ]
915
+ },
916
+ {
917
+ "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
918
+ "repo_name": "FTorch",
919
+ "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
920
+ "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
921
+ "repo_tags": [
922
+ "deep-learning",
923
+ "fortran",
924
+ "hacktoberfest",
925
+ "interoperability",
926
+ "libtorch",
927
+ "machine-learning",
928
+ "python",
929
+ "pytorch",
930
+ "torch"
931
+ ]
932
+ },
933
+ {
934
+ "github_repo_link": "https://github.com/facebook/hhvm",
935
+ "repo_name": "hhvm",
936
+ "repo_description": "A virtual machine for executing programs written in Hack.",
937
+ "homepage_link": "https://hhvm.com",
938
+ "repo_tags": [
939
+ "hack",
940
+ "hacklang",
941
+ "hhvm",
942
+ "php"
943
+ ]
944
+ },
945
+ {
946
+ "github_repo_link": "https://github.com/MoonshotAI/Kimi-K2",
947
+ "repo_name": "Kimi-K2",
948
+ "repo_description": "Kimi K2 is the large language model series developed by Moonshot AI team",
949
+ "homepage_link": "",
950
+ "repo_tags": []
951
+ },
952
+ {
953
+ "github_repo_link": "https://github.com/vosen/ZLUDA",
954
+ "repo_name": "ZLUDA",
955
+ "repo_description": "CUDA on non-NVIDIA GPUs",
956
+ "homepage_link": "https://vosen.github.io/ZLUDA/",
957
+ "repo_tags": [
958
+ "cuda",
959
+ "rust"
960
+ ]
961
+ },
962
+ {
963
+ "github_repo_link": "https://github.com/vtsynergy/CU2CL",
964
+ "repo_name": "CU2CL",
965
+ "repo_description": "A prototype CUDA-to-OpenCL source-to-source translator, built on the Clang compiler framework",
966
+ "homepage_link": "http://chrec.cs.vt.edu/cu2cl",
967
+ "repo_tags": []
968
+ },
969
+ {
970
+ "github_repo_link": "https://github.com/pocl/pocl",
971
+ "repo_name": "pocl",
972
+ "repo_description": "pocl - Portable Computing Language",
973
+ "homepage_link": "https://portablecl.org",
974
+ "repo_tags": [
975
+ "heterogeneous-parallel-programming",
976
+ "opencl"
977
+ ]
978
+ },
979
+ {
980
+ "github_repo_link": "https://github.com/apache/spark",
981
+ "repo_name": "spark",
982
+ "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
983
+ "homepage_link": "https://spark.apache.org/",
984
+ "repo_tags": [
985
+ "big-data",
986
+ "java",
987
+ "jdbc",
988
+ "python",
989
+ "r",
990
+ "scala",
991
+ "spark",
992
+ "sql"
993
+ ]
994
+ },
995
+ {
996
+ "github_repo_link": "https://github.com/codelion/openevolve",
997
+ "repo_name": "openevolve",
998
+ "repo_description": "Open-source implementation of AlphaEvolve",
999
+ "homepage_link": "",
1000
+ "repo_tags": [
1001
+ "alpha-evolve",
1002
+ "alphacode",
1003
+ "alphaevolve",
1004
+ "coding-agent",
1005
+ "deepmind",
1006
+ "deepmind-lab",
1007
+ "discovery",
1008
+ "distributed-evolutionary-algorithms",
1009
+ "evolutionary-algorithms",
1010
+ "evolutionary-computation",
1011
+ "genetic-algorithm",
1012
+ "genetic-algorithms",
1013
+ "iterative-methods",
1014
+ "iterative-refinement",
1015
+ "llm-engineering",
1016
+ "llm-ensemble",
1017
+ "llm-inference",
1018
+ "openevolve",
1019
+ "optimize"
1020
+ ]
1021
+ },
1022
+ {
1023
+ "github_repo_link": "https://github.com/PaddlePaddle/ERNIE",
1024
+ "repo_name": "ERNIE",
1025
+ "repo_description": "The official repository for ERNIE 4.5 and ERNIEKit – its industrial-grade development toolkit based on PaddlePaddle.",
1026
+ "homepage_link": "https://ernie.baidu.com",
1027
+ "repo_tags": [
1028
+ "ernie",
1029
+ "ernie-45",
1030
+ "ernie-45-vl",
1031
+ "erniekit",
1032
+ "llm",
1033
+ "vlm"
1034
+ ]
1035
+ },
1036
+ {
1037
+ "github_repo_link": "https://github.com/ROCm/hipBLAS",
1038
+ "repo_name": "hipBLAS",
1039
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1040
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1041
+ "repo_tags": [
1042
+ "blas",
1043
+ "cuda",
1044
+ "hip",
1045
+ "rocm"
1046
+ ]
1047
+ },
1048
+ {
1049
+ "github_repo_link": "https://github.com/ROCm/roctracer",
1050
+ "repo_name": "roctracer",
1051
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-systems repo ",
1052
+ "homepage_link": "https://github.com/ROCm/rocm-systems",
1053
+ "repo_tags": []
1054
+ },
1055
+ {
1056
+ "github_repo_link": "https://github.com/huggingface/peft",
1057
+ "repo_name": "peft",
1058
+ "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
1059
+ "homepage_link": "https://huggingface.co/docs/peft",
1060
+ "repo_tags": [
1061
+ "adapter",
1062
+ "diffusion",
1063
+ "fine-tuning",
1064
+ "llm",
1065
+ "lora",
1066
+ "parameter-efficient-learning",
1067
+ "peft",
1068
+ "python",
1069
+ "pytorch",
1070
+ "transformers"
1071
+ ]
1072
+ },
1073
+ {
1074
+ "github_repo_link": "https://github.com/ROCm/hip",
1075
+ "repo_name": "hip",
1076
+ "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
1077
+ "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
1078
+ "repo_tags": [
1079
+ "cuda",
1080
+ "hip",
1081
+ "hip-kernel-language",
1082
+ "hip-portability",
1083
+ "hip-runtime",
1084
+ "hipify"
1085
+ ]
1086
+ },
1087
+ {
1088
+ "github_repo_link": "https://github.com/ROCm/composable_kernel",
1089
+ "repo_name": "composable_kernel",
1090
+ "repo_description": "Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators",
1091
+ "homepage_link": "https://rocm.docs.amd.com/projects/composable_kernel/en/latest/",
1092
+ "repo_tags": []
1093
+ },
1094
+ {
1095
+ "github_repo_link": "https://github.com/ROCm/aiter",
1096
+ "repo_name": "aiter",
1097
+ "repo_description": "AI Tensor Engine for ROCm",
1098
+ "homepage_link": null,
1099
+ "repo_tags": []
1100
+ },
1101
+ {
1102
+ "github_repo_link": "https://github.com/AMDResearch/intelliperf",
1103
+ "repo_name": "intelliperf",
1104
+ "repo_description": "Automated bottleneck detection and solution orchestration",
1105
+ "homepage_link": "",
1106
+ "repo_tags": [
1107
+ "amd",
1108
+ "genai",
1109
+ "gpu",
1110
+ "hip",
1111
+ "instinct",
1112
+ "llm",
1113
+ "performance",
1114
+ "rocm"
1115
+ ]
1116
+ },
1117
+ {
1118
+ "github_repo_link": "https://github.com/AMD-AGI/Instella",
1119
+ "repo_name": "Instella",
1120
+ "repo_description": "Fully Open Language Models with Stellar Performance",
1121
+ "homepage_link": null,
1122
+ "repo_tags": []
1123
+ },
1124
+ {
1125
+ "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
1126
+ "repo_name": "GEAK-agent",
1127
+ "repo_description": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
1128
+ "homepage_link": null,
1129
+ "repo_tags": []
1130
+ },
1131
+ {
1132
+ "github_repo_link": "https://github.com/AMD-AGI/torchtitan",
1133
+ "repo_name": "torchtitan",
1134
+ "repo_description": "A PyTorch native platform for training generative AI models",
1135
+ "homepage_link": "",
1136
+ "repo_tags": []
1137
+ },
1138
+ {
1139
+ "github_repo_link": "https://github.com/AMD-AGI/hipBLASLt",
1140
+ "repo_name": "hipBLASLt",
1141
+ "repo_description": "hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditional BLAS library",
1142
+ "homepage_link": "https://rocm.docs.amd.com/projects/hipBLASLt/en/latest/index.html",
1143
+ "repo_tags": []
1144
+ },
1145
+ {
1146
+ "github_repo_link": "https://github.com/AMD-AGI/rocm-torchtitan",
1147
+ "repo_name": "rocm-torchtitan",
1148
+ "repo_description": null,
1149
+ "homepage_link": null,
1150
+ "repo_tags": []
1151
+ },
1152
+ {
1153
+ "github_repo_link": "https://github.com/HazyResearch/Megakernels",
1154
+ "repo_name": "Megakernels",
1155
+ "repo_description": "kernels, of the mega variety",
1156
+ "homepage_link": null,
1157
+ "repo_tags": []
1158
+ },
1159
+ {
1160
+ "github_repo_link": "https://github.com/microsoft/DirectXShaderCompiler",
1161
+ "repo_name": "DirectXShaderCompiler",
1162
+ "repo_description": "This repo hosts the source for the DirectX Shader Compiler which is based on LLVM/Clang.",
1163
+ "homepage_link": "",
1164
+ "repo_tags": [
1165
+ "directx-shader-compiler",
1166
+ "dxil",
1167
+ "hlsl",
1168
+ "shader-programs"
1169
+ ]
1170
+ },
1171
+ {
1172
+ "github_repo_link": "https://github.com/microsoft/DirectXTK",
1173
+ "repo_name": "DirectXTK",
1174
+ "repo_description": "The DirectX Tool Kit (aka DirectXTK) is a collection of helper classes for writing DirectX 11.x code in C++",
1175
+ "homepage_link": "https://walbourn.github.io/directxtk/",
1176
+ "repo_tags": [
1177
+ "cpp-library",
1178
+ "desktop",
1179
+ "directx",
1180
+ "directx-11",
1181
+ "directxtk",
1182
+ "graphics",
1183
+ "microsoft",
1184
+ "uwp",
1185
+ "xbox"
1186
+ ]
1187
+ },
1188
+ {
1189
+ "github_repo_link": "https://github.com/huggingface/kernels",
1190
+ "repo_name": "kernels",
1191
+ "repo_description": "Load compute kernels from the Hub",
1192
+ "homepage_link": "",
1193
+ "repo_tags": []
1194
+ },
1195
+ {
1196
+ "github_repo_link": "https://github.com/tile-ai/tilelang",
1197
+ "repo_name": "tilelang",
1198
+ "repo_description": " Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
1199
+ "homepage_link": "https://tilelang.com/",
1200
+ "repo_tags": []
1201
+ },
1202
+ {
1203
+ "github_repo_link": "https://github.com/opencv/opencv",
1204
+ "repo_name": "opencv",
1205
+ "repo_description": "Open Source Computer Vision Library",
1206
+ "homepage_link": "https://opencv.org",
1207
+ "repo_tags": [
1208
+ "c-plus-plus",
1209
+ "computer-vision",
1210
+ "deep-learning",
1211
+ "image-processing",
1212
+ "opencv"
1213
+ ]
1214
+ },
1215
+ {
1216
+ "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
1217
+ "repo_name": "lightning-thunder",
1218
+ "repo_description": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.",
1219
+ "homepage_link": "",
1220
+ "repo_tags": []
1221
+ },
1222
+ {
1223
+ "github_repo_link": "https://github.com/tracel-ai/burn",
1224
+ "repo_name": "burn",
1225
+ "repo_description": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
1226
+ "homepage_link": "https://burn.dev",
1227
+ "repo_tags": [
1228
+ "autodiff",
1229
+ "cross-platform",
1230
+ "cuda",
1231
+ "deep-learning",
1232
+ "kernel-fusion",
1233
+ "machine-learning",
1234
+ "metal",
1235
+ "ndarray",
1236
+ "neural-network",
1237
+ "onnx",
1238
+ "pytorch",
1239
+ "rocm",
1240
+ "rust",
1241
+ "scientific-computing",
1242
+ "tensor",
1243
+ "vulkan",
1244
+ "wasm",
1245
+ "webgpu"
1246
+ ]
1247
+ },
1248
+ {
1249
+ "github_repo_link": "https://github.com/huggingface/kernels-community",
1250
+ "repo_name": "kernels-community",
1251
+ "repo_description": "Kernel sources for https://huggingface.co/kernels-community",
1252
+ "homepage_link": null,
1253
+ "repo_tags": []
1254
+ },
1255
+ {
1256
+ "github_repo_link": "https://github.com/flashinfer-ai/flashinfer-bench",
1257
+ "repo_name": "flashinfer-bench",
1258
+ "repo_description": "Building the Virtuous Cycle for AI-driven LLM Systems",
1259
+ "homepage_link": "https://bench.flashinfer.ai",
1260
+ "repo_tags": []
1261
+ },
1262
+ {
1263
+ "github_repo_link": "https://github.com/OSC/ondemand",
1264
+ "repo_name": "ondemand",
1265
+ "repo_description": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
1266
+ "homepage_link": "https://openondemand.org/",
1267
+ "repo_tags": [
1268
+ "gateway",
1269
+ "hacktoberfest",
1270
+ "hpc",
1271
+ "hpc-applications"
1272
+ ]
1273
+ },
1274
+ {
1275
+ "github_repo_link": "https://github.com/flashinfer-ai/flashinfer",
1276
+ "repo_name": "flashinfer",
1277
+ "repo_description": "FlashInfer: Kernel Library for LLM Serving",
1278
+ "homepage_link": "https://flashinfer.ai",
1279
+ "repo_tags": [
1280
+ "attention",
1281
+ "cuda",
1282
+ "distributed-inference",
1283
+ "gpu",
1284
+ "jit",
1285
+ "large-large-models",
1286
+ "llm-inference",
1287
+ "moe",
1288
+ "nvidia",
1289
+ "pytorch"
1290
+ ]
1291
+ },
1292
+ {
1293
+ "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
1294
+ "repo_name": "KernelBench",
1295
+ "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
1296
+ "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
1297
+ "repo_tags": [
1298
+ "benchmark",
1299
+ "codegen",
1300
+ "evaluation",
1301
+ "gpu"
1302
+ ]
1303
+ },
1304
+ {
1305
+ "github_repo_link": "https://github.com/thunlp/TritonBench",
1306
+ "repo_name": "TritonBench",
1307
+ "repo_description": "TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators",
1308
+ "homepage_link": "",
1309
+ "repo_tags": []
1310
+ },
1311
+ {
1312
+ "github_repo_link": "https://github.com/AutomataLab/cuJSON",
1313
+ "repo_name": "cuJSON",
1314
+ "repo_description": "cuJSON: A Highly Parallel JSON Parser for GPUs",
1315
+ "homepage_link": "",
1316
+ "repo_tags": []
1317
+ },
1318
+ {
1319
+ "github_repo_link": "https://github.com/Netflix/metaflow",
1320
+ "repo_name": "metaflow",
1321
+ "repo_description": "Build, Manage and Deploy AI/ML Systems",
1322
+ "homepage_link": "https://metaflow.org",
1323
+ "repo_tags": [
1324
+ "agents",
1325
+ "ai",
1326
+ "aws",
1327
+ "azure",
1328
+ "cost-optimization",
1329
+ "datascience",
1330
+ "distributed-training",
1331
+ "gcp",
1332
+ "generative-ai",
1333
+ "high-performance-computing",
1334
+ "kubernetes",
1335
+ "llm",
1336
+ "llmops",
1337
+ "machine-learning",
1338
+ "ml",
1339
+ "ml-infrastructure",
1340
+ "ml-platform",
1341
+ "mlops",
1342
+ "model-management",
1343
+ "python"
1344
+ ]
1345
+ },
1346
+ {
1347
+ "github_repo_link": "https://github.com/docker/compose",
1348
+ "repo_name": "compose",
1349
+ "repo_description": "Define and run multi-container applications with Docker",
1350
+ "homepage_link": "https://docs.docker.com/compose/",
1351
+ "repo_tags": [
1352
+ "docker",
1353
+ "docker-compose",
1354
+ "go",
1355
+ "golang",
1356
+ "orchestration"
1357
+ ]
1358
+ },
1359
+ {
1360
+ "github_repo_link": "https://github.com/docker/cli",
1361
+ "repo_name": "cli",
1362
+ "repo_description": "The Docker CLI",
1363
+ "homepage_link": null,
1364
+ "repo_tags": [
1365
+ "cli",
1366
+ "docker"
1367
+ ]
1368
+ },
1369
+ {
1370
+ "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
1371
+ "repo_name": "IMO2025",
1372
+ "repo_description": null,
1373
+ "homepage_link": null,
1374
+ "repo_tags": []
1375
+ },
1376
+ {
1377
+ "github_repo_link": "https://github.com/leanprover/lean4",
1378
+ "repo_name": "lean4",
1379
+ "repo_description": "Lean 4 programming language and theorem prover",
1380
+ "homepage_link": "https://lean-lang.org",
1381
+ "repo_tags": [
1382
+ "lean",
1383
+ "lean4"
1384
+ ]
1385
+ },
1386
+ {
1387
+ "github_repo_link": "https://github.com/NVIDIA/warp",
1388
+ "repo_name": "warp",
1389
+ "repo_description": "A Python framework for accelerated simulation, data generation and spatial computing.",
1390
+ "homepage_link": "https://nvidia.github.io/warp/",
1391
+ "repo_tags": [
1392
+ "cuda",
1393
+ "differentiable-programming",
1394
+ "gpu",
1395
+ "gpu-acceleration",
1396
+ "nvidia",
1397
+ "nvidia-warp",
1398
+ "python"
1399
+ ]
1400
+ },
1401
+ {
1402
+ "github_repo_link": "https://github.com/NVIDIA/cuda-python",
1403
+ "repo_name": "cuda-python",
1404
+ "repo_description": "CUDA Python: Performance meets Productivity",
1405
+ "homepage_link": "https://nvidia.github.io/cuda-python/",
1406
+ "repo_tags": []
1407
+ },
1408
+ {
1409
+ "github_repo_link": "https://github.com/basetenlabs/truss",
1410
+ "repo_name": "truss",
1411
+ "repo_description": "The simplest way to serve AI/ML models in production",
1412
+ "homepage_link": "https://truss.baseten.co",
1413
+ "repo_tags": [
1414
+ "artificial-intelligence",
1415
+ "easy-to-use",
1416
+ "falcon",
1417
+ "inference-api",
1418
+ "inference-server",
1419
+ "machine-learning",
1420
+ "model-serving",
1421
+ "open-source",
1422
+ "packaging",
1423
+ "stable-diffusion",
1424
+ "whisper",
1425
+ "wizardlm"
1426
+ ]
1427
+ },
1428
+ {
1429
+ "github_repo_link": "https://github.com/laude-institute/terminal-bench",
1430
+ "repo_name": "terminal-bench",
1431
+ "repo_description": "A benchmark for LLMs on complicated tasks in the terminal",
1432
+ "homepage_link": "https://www.tbench.ai",
1433
+ "repo_tags": []
1434
+ },
1435
+ {
1436
+ "github_repo_link": "https://github.com/block/goose",
1437
+ "repo_name": "goose",
1438
+ "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
1439
+ "homepage_link": "https://block.github.io/goose/",
1440
+ "repo_tags": [
1441
+ "hacktoberfest",
1442
+ "mcp"
1443
+ ]
1444
+ },
1445
+ {
1446
+ "github_repo_link": "https://github.com/kvcache-ai/Mooncake",
1447
+ "repo_name": "Mooncake",
1448
+ "repo_description": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
1449
+ "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
1450
+ "repo_tags": [
1451
+ "disaggregation",
1452
+ "inference",
1453
+ "kvcache",
1454
+ "llm",
1455
+ "rdma",
1456
+ "sglang",
1457
+ "vllm"
1458
+ ]
1459
+ },
1460
+ {
1461
+ "github_repo_link": "https://github.com/SWE-bench/SWE-bench",
1462
+ "repo_name": "SWE-bench",
1463
+ "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
1464
+ "homepage_link": "https://www.swebench.com",
1465
+ "repo_tags": [
1466
+ "benchmark",
1467
+ "language-model",
1468
+ "software-engineering"
1469
+ ]
1470
+ },
1471
+ {
1472
+ "github_repo_link": "https://github.com/Dao-AILab/quack",
1473
+ "repo_name": "quack",
1474
+ "repo_description": "A Quirky Assortment of CuTe Kernels",
1475
+ "homepage_link": "",
1476
+ "repo_tags": []
1477
+ },
1478
+ {
1479
+ "github_repo_link": "https://github.com/KhronosGroup/SYCL-Docs",
1480
+ "repo_name": "SYCL-Docs",
1481
+ "repo_description": "SYCL Open Source Specification",
1482
+ "homepage_link": null,
1483
+ "repo_tags": []
1484
+ },
1485
+ {
1486
+ "github_repo_link": "https://github.com/triSYCL/triSYCL",
1487
+ "repo_name": "triSYCL",
1488
+ "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
1489
+ "homepage_link": "",
1490
+ "repo_tags": [
1491
+ "cpp",
1492
+ "cpp20",
1493
+ "fpga",
1494
+ "gpu-computing",
1495
+ "heterogeneous-parallel-programming",
1496
+ "opencl",
1497
+ "spir",
1498
+ "sycl",
1499
+ "trisycl"
1500
+ ]
1501
+ },
1502
+ {
1503
+ "github_repo_link": "https://github.com/pybind/pybind11",
1504
+ "repo_name": "pybind11",
1505
+ "repo_description": "Seamless operability between C++11 and Python",
1506
+ "homepage_link": "https://pybind11.readthedocs.io/",
1507
+ "repo_tags": [
1508
+ "bindings",
1509
+ "python"
1510
+ ]
1511
+ },
1512
+ {
1513
+ "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
1514
+ "repo_name": "nvcc4jupyter",
1515
+ "repo_description": "A plugin for Jupyter Notebook to run CUDA C/C++ code",
1516
+ "homepage_link": null,
1517
+ "repo_tags": []
1518
+ },
1519
+ {
1520
+ "github_repo_link": "https://github.com/ROCm/rocSOLVER",
1521
+ "repo_name": "rocSOLVER",
1522
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1523
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1524
+ "repo_tags": [
1525
+ "lapack",
1526
+ "linear-algebra",
1527
+ "rocm"
1528
+ ]
1529
+ },
1530
+ {
1531
+ "github_repo_link": "https://github.com/ROCm/Tensile",
1532
+ "repo_name": "Tensile",
1533
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1534
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1535
+ "repo_tags": [
1536
+ "amd",
1537
+ "assembly",
1538
+ "auto-tuning",
1539
+ "blas",
1540
+ "dnn",
1541
+ "gemm",
1542
+ "gpu",
1543
+ "gpu-acceleration",
1544
+ "gpu-computing",
1545
+ "hip",
1546
+ "machine-learning",
1547
+ "matrix-multiplication",
1548
+ "neural-networks",
1549
+ "opencl",
1550
+ "python",
1551
+ "radeon",
1552
+ "tensor-contraction",
1553
+ "tensors"
1554
+ ]
1555
+ },
1556
+ {
1557
+ "github_repo_link": "https://github.com/ROCm/rocPRIM",
1558
+ "repo_name": "rocPRIM",
1559
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
1560
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1561
+ "repo_tags": [
1562
+ "amd",
1563
+ "cuda",
1564
+ "gpu",
1565
+ "hip",
1566
+ "parallel",
1567
+ "primitive",
1568
+ "rocm"
1569
+ ]
1570
+ },
1571
+ {
1572
+ "github_repo_link": "https://github.com/ROCm/hipCUB",
1573
+ "repo_name": "hipCUB",
1574
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
1575
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1576
+ "repo_tags": []
1577
+ },
1578
+ {
1579
+ "github_repo_link": "https://github.com/ROCm/rocFFT",
1580
+ "repo_name": "rocFFT",
1581
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1582
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1583
+ "repo_tags": [
1584
+ "amd",
1585
+ "fast",
1586
+ "fft",
1587
+ "fourier",
1588
+ "gpu",
1589
+ "hip",
1590
+ "rocm",
1591
+ "transform"
1592
+ ]
1593
+ },
1594
+ {
1595
+ "github_repo_link": "https://github.com/ROCm/rocSPARSE",
1596
+ "repo_name": "rocSPARSE",
1597
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1598
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1599
+ "repo_tags": []
1600
+ },
1601
+ {
1602
+ "github_repo_link": "https://github.com/ROCm/rocRAND",
1603
+ "repo_name": "rocRAND",
1604
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
1605
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1606
+ "repo_tags": [
1607
+ "cuda",
1608
+ "gpu",
1609
+ "hip",
1610
+ "random",
1611
+ "rng",
1612
+ "rocm"
1613
+ ]
1614
+ },
1615
+ {
1616
+ "github_repo_link": "https://github.com/ROCm/MIOpen",
1617
+ "repo_name": "MIOpen",
1618
+ "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
1619
+ "homepage_link": "https://github.com/ROCm/rocm-libraries",
1620
+ "repo_tags": []
1621
+ },
1622
+ {
1623
+ "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
1624
+ "repo_name": "lapack",
1625
+ "repo_description": "LAPACK development repository",
1626
+ "homepage_link": "",
1627
+ "repo_tags": [
1628
+ "blas",
1629
+ "eigenvalues",
1630
+ "eigenvectors",
1631
+ "lapack",
1632
+ "lapacke",
1633
+ "linear-algebra",
1634
+ "linear-equations",
1635
+ "matrix-factorization",
1636
+ "singular-values",
1637
+ "svd"
1638
+ ]
1639
+ },
1640
+ {
1641
+ "github_repo_link": "https://github.com/ccache/ccache",
1642
+ "repo_name": "ccache",
1643
+ "repo_description": "ccache – a fast compiler cache",
1644
+ "homepage_link": "https://ccache.dev",
1645
+ "repo_tags": [
1646
+ "c",
1647
+ "c-plus-plus",
1648
+ "cache",
1649
+ "ccache",
1650
+ "clang",
1651
+ "compiler",
1652
+ "cplusplus",
1653
+ "cpp",
1654
+ "gcc",
1655
+ "msvc"
1656
+ ]
1657
+ },
1658
+ {
1659
+ "github_repo_link": "https://github.com/ROCm/omnitrace",
1660
+ "repo_name": "omnitrace",
1661
+ "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
1662
+ "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
1663
+ "repo_tags": [
1664
+ "binary-instrumentation",
1665
+ "code-coverage",
1666
+ "cpu-profiler",
1667
+ "dynamic-instrumentation",
1668
+ "gpu-profiler",
1669
+ "hardware-counters",
1670
+ "instrumentation-profiler",
1671
+ "linux",
1672
+ "performance-analysis",
1673
+ "performance-metrics",
1674
+ "performance-monitoring",
1675
+ "profiler",
1676
+ "profiling",
1677
+ "python",
1678
+ "python-profiler",
1679
+ "sampling-profiler",
1680
+ "tracing"
1681
+ ]
1682
+ },
1683
+ {
1684
+ "github_repo_link": "https://github.com/python/cpython",
1685
+ "repo_name": "cpython",
1686
+ "repo_description": "The Python programming language",
1687
+ "homepage_link": "https://www.python.org",
1688
+ "repo_tags": []
1689
+ },
1690
+ {
1691
+ "github_repo_link": "https://github.com/rust-lang/rust",
1692
+ "repo_name": "rust",
1693
+ "repo_description": "Empowering everyone to build reliable and efficient software.",
1694
+ "homepage_link": "https://www.rust-lang.org",
1695
+ "repo_tags": [
1696
+ "compiler",
1697
+ "language",
1698
+ "rust"
1699
+ ]
1700
+ },
1701
+ {
1702
+ "github_repo_link": "https://github.com/tailscale/tailscale",
1703
+ "repo_name": "tailscale",
1704
+ "repo_description": "The easiest, most secure way to use WireGuard and 2FA.",
1705
+ "homepage_link": "https://tailscale.com",
1706
+ "repo_tags": [
1707
+ "2fa",
1708
+ "oauth",
1709
+ "sso",
1710
+ "tailscale",
1711
+ "vpn",
1712
+ "wireguard"
1713
+ ]
1714
+ },
1715
+ {
1716
+ "github_repo_link": "https://github.com/WireGuard/wireguard-linux",
1717
+ "repo_name": "wireguard-linux",
1718
+ "repo_description": "Mirror only. Official repository is at https://git.zx2c4.com/wireguard-linux",
1719
+ "homepage_link": "https://www.wireguard.com",
1720
+ "repo_tags": []
1721
+ }
1722
+ ]