TylerHilbert commited on
Commit
bb96ee5
·
1 Parent(s): 6146163

Progress on categories and reorder

Browse files
PyTorchConference2025_GithubRepos.json CHANGED
@@ -1,109 +1,126 @@
1
  [
2
  {
3
- "github_repo_link": "https://github.com/pytorch/pytorch",
4
  "repo_name": "pytorch",
 
 
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
- "github_topic_closest_fit": "machine-learning",
8
- "category": "machine learning framework"
9
- },
10
- {
11
- "github_repo_link": "https://github.com/vllm-project/vllm",
12
- "repo_name": "vllm",
13
- "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
14
- "homepage_link": "https://docs.vllm.ai",
15
- "github_topic_closest_fit": "inference",
16
- "category": "inference engine"
17
  },
18
  {
19
- "github_repo_link": "https://github.com/ollama/ollama",
20
- "repo_name": "ollama",
21
- "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
22
- "homepage_link": "https://ollama.com",
23
- "github_topic_closest_fit": "llms",
24
- "category": "inference engine"
25
  },
26
  {
27
- "github_repo_link": "https://github.com/sgl-project/sglang",
28
- "repo_name": "sglang",
29
- "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
30
- "homepage_link": "https://docs.sglang.ai/",
31
- "github_topic_closest_fit": "inference",
32
- "category": "inference engine"
33
  },
34
  {
35
- "github_repo_link": "https://github.com/ggml-org/llama.cpp",
36
- "repo_name": "llama.cpp",
37
- "repo_description": "LLM inference in C/C++",
38
- "github_topic_closest_fit": "ggml",
39
- "category": "inference engine"
 
40
  },
41
  {
42
- "github_repo_link": "https://github.com/triton-lang/triton",
43
- "repo_name": "triton",
44
- "repo_description": "Development repository for the Triton language and compiler",
45
- "homepage_link": "https://triton-lang.org/",
46
- "category": "dsl"
 
47
  },
48
  {
49
- "github_repo_link": "https://github.com/pytorch/helion",
50
  "repo_name": "helion",
 
 
51
  "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
52
- "category": "dsl"
 
53
  },
54
  {
55
- "github_repo_link": "https://github.com/microsoft/TileIR",
56
  "repo_name": "TileIR",
57
- "category": "dsl"
58
- },
59
- {
60
- "github_repo_link": "https://github.com/tile-ai/tilelang",
61
- "repo_name": "tilelang",
62
- "repo_description": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
63
- "homepage_link": "https://tilelang.com/",
64
- "category": "dsl"
65
  },
66
  {
67
- "github_repo_link": "https://github.com/NVIDIA/cutlass",
68
- "repo_name": "cutlass",
69
- "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
70
- "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
71
- "github_topic_closest_fit": "cuda"
72
- },
73
  {
74
- "github_repo_link": "https://github.com/tensorflow/tensorflow",
75
  "repo_name": "tensorflow",
 
 
76
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
77
  "homepage_link": "https://tensorflow.org",
78
- "github_topic_closest_fit": "deep-learning",
79
- "category": "machine learning framework"
80
  },
81
  {
82
- "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
83
- "repo_name": "ThunderKittens",
84
- "repo_description": "Tile primitives for speedy kernels"
 
 
 
85
  },
86
  {
87
- "github_repo_link": "https://github.com/pytorch/executorch",
88
- "repo_name": "executorch",
89
- "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
90
- "homepage_link": "https://executorch.ai",
91
- "github_topic_closest_fit": "mobile",
92
- "category": "model compiler"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
93
  },
94
  {
95
- "github_repo_link": "https://github.com/onnx/onnx",
96
  "repo_name": "onnx",
 
 
97
  "repo_description": "Open standard for machine learning interoperability",
98
  "homepage_link": "https://onnx.ai/",
99
- "github_topic_closest_fit": "deep-learning"
 
 
 
 
 
 
 
 
100
  },
101
  {
102
  "github_repo_link": "https://github.com/ray-project/ray",
103
  "repo_name": "ray",
104
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
105
  "homepage_link": "https://ray.io",
106
- "github_topic_closest_fit": "deep-learning"
107
  },
108
  {
109
  "github_repo_link": "https://github.com/jax-ml/jax",
@@ -553,7 +570,7 @@
553
  "repo_name": "FTorch",
554
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
555
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
556
- "github_topic_closest_fit": "deep-learning"
557
  },
558
  {
559
  "github_repo_link": "https://github.com/facebook/hhvm",
 
1
  [
2
  {
 
3
  "repo_name": "pytorch",
4
+ "github_repo_link": "https://github.com/pytorch/pytorch",
5
+ "category": "machine learning framework",
6
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
7
  "homepage_link": "https://pytorch.org",
8
+ "github_topic_closest_fit": "machine-learning"
 
 
 
 
 
 
 
 
 
9
  },
10
  {
11
+ "repo_name": "triton",
12
+ "github_repo_link": "https://github.com/triton-lang/triton",
13
+ "category": "parallel computing dsl",
14
+ "repo_description": "Development repository for the Triton language and compiler",
15
+ "homepage_link": "https://triton-lang.org/",
16
+ "github_topic_closest_fit": "parallel-programming"
17
  },
18
  {
19
+ "repo_name": "cutlass",
20
+ "github_repo_link": "https://github.com/NVIDIA/cutlass",
21
+ "category": "parallel computing",
22
+ "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
23
+ "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
24
+ "github_topic_closest_fit": "parallel-programming"
25
  },
26
  {
27
+ "repo_name": "tilelang",
28
+ "github_repo_link": "https://github.com/tile-ai/tilelang",
29
+ "category": "parallel computing dsl",
30
+ "repo_description": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
31
+ "homepage_link": "https://tilelang.com",
32
+ "github_topic_closest_fit": "parallel-programming"
33
  },
34
  {
35
+ "repo_name": "ThunderKittens",
36
+ "github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
37
+ "category": "parallel computing",
38
+ "repo_description": "Tile primitives for speedy kernels",
39
+ "homepage_link": "https://hazyresearch.stanford.edu/blog/2024-10-29-tk2",
40
+ "github_topic_closest_fit": "parallel-programming"
41
  },
42
  {
 
43
  "repo_name": "helion",
44
+ "github_repo_link": "https://github.com/pytorch/helion",
45
+ "category": "parallel computing dsl",
46
  "repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
47
+ "homepage_link": "https://helionlang.com",
48
+ "github_topic_closest_fit": "parallel-programming"
49
  },
50
  {
 
51
  "repo_name": "TileIR",
52
+ "github_repo_link": "https://github.com/microsoft/TileIR",
53
+ "category": "parallel computing dsl",
54
+ "repo_description": "TileIR (tile-ir) is a concise domain-specific IR designed to streamline the development of high-performance GPU/CPU kernels (e.g., GEMM, Dequant GEMM, FlashAttention, LinearAttention). By employing a Pythonic syntax with an underlying compiler infrastructure on top of TVM, TileIR allows developers to focus on productivity without sacrificing the low-level optimizations necessary for state-of-the-art performance.",
55
+ "github_topic_closest_fit": "parallel-programming"
 
 
 
 
56
  },
57
  {
58
+ "repo_name": "BitBLAS",
59
+ "github_repo_link": "https://github.com/microsoft/BitBLAS",
60
+ "repo_description": "BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment."
61
+ },
 
 
62
  {
 
63
  "repo_name": "tensorflow",
64
+ "github_repo_link": "https://github.com/tensorflow/tensorflow",
65
+ "category": "machine learning framework",
66
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
67
  "homepage_link": "https://tensorflow.org",
68
+ "github_topic_closest_fit": "machine-learning"
 
69
  },
70
  {
71
+ "repo_name": "vllm",
72
+ "github_repo_link": "https://github.com/vllm-project/vllm",
73
+ "category": "inference engine",
74
+ "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
75
+ "homepage_link": "https://docs.vllm.ai",
76
+ "github_topic_closest_fit": "inference"
77
  },
78
  {
79
+ "repo_name": "ollama",
80
+ "github_repo_link": "https://github.com/ollama/ollama",
81
+ "category": "inference engine",
82
+ "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
83
+ "homepage_link": "https://ollama.com",
84
+ "github_topic_closest_fit": "inference"
85
+ },
86
+ {
87
+ "repo_name": "llama.cpp",
88
+ "github_repo_link": "https://github.com/ggml-org/llama.cpp",
89
+ "category": "inference engine",
90
+ "repo_description": "LLM inference in C/C++",
91
+ "homepage_link": "https://ggml.ai",
92
+ "github_topic_closest_fit": "inference"
93
+ },
94
+ {
95
+ "repo_name": "sglang",
96
+ "github_repo_link": "https://github.com/sgl-project/sglang",
97
+ "category": "inference engine",
98
+ "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
99
+ "homepage_link": "https://docs.sglang.ai",
100
+ "github_topic_closest_fit": "inference"
101
  },
102
  {
 
103
  "repo_name": "onnx",
104
+ "github_repo_link": "https://github.com/onnx/onnx",
105
+ "category": "machine learning framework",
106
  "repo_description": "Open standard for machine learning interoperability",
107
  "homepage_link": "https://onnx.ai/",
108
+ "github_topic_closest_fit": "onnx"
109
+ },
110
+ {
111
+ "repo_name": "executorch",
112
+ "github_repo_link": "https://github.com/pytorch/executorch",
113
+ "category": "model compiler",
114
+ "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
115
+ "homepage_link": "https://executorch.ai",
116
+ "github_topic_closest_fit": "compiler"
117
  },
118
  {
119
  "github_repo_link": "https://github.com/ray-project/ray",
120
  "repo_name": "ray",
121
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
122
  "homepage_link": "https://ray.io",
123
+ "github_topic_closest_fit": "machine-learning"
124
  },
125
  {
126
  "github_repo_link": "https://github.com/jax-ml/jax",
 
570
  "repo_name": "FTorch",
571
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
572
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
573
+ "github_topic_closest_fit": "machine-learning"
574
  },
575
  {
576
  "github_repo_link": "https://github.com/facebook/hhvm",