Commit
·
bb96ee5
1
Parent(s):
6146163
Progress on categories and reorder
Browse files
PyTorchConference2025_GithubRepos.json
CHANGED
|
@@ -1,109 +1,126 @@
|
|
| 1 |
[
|
| 2 |
{
|
| 3 |
-
"github_repo_link": "https://github.com/pytorch/pytorch",
|
| 4 |
"repo_name": "pytorch",
|
|
|
|
|
|
|
| 5 |
"repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
|
| 6 |
"homepage_link": "https://pytorch.org",
|
| 7 |
-
"github_topic_closest_fit": "machine-learning"
|
| 8 |
-
"category": "machine learning framework"
|
| 9 |
-
},
|
| 10 |
-
{
|
| 11 |
-
"github_repo_link": "https://github.com/vllm-project/vllm",
|
| 12 |
-
"repo_name": "vllm",
|
| 13 |
-
"repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
|
| 14 |
-
"homepage_link": "https://docs.vllm.ai",
|
| 15 |
-
"github_topic_closest_fit": "inference",
|
| 16 |
-
"category": "inference engine"
|
| 17 |
},
|
| 18 |
{
|
| 19 |
-
"
|
| 20 |
-
"
|
| 21 |
-
"
|
| 22 |
-
"
|
| 23 |
-
"
|
| 24 |
-
"
|
| 25 |
},
|
| 26 |
{
|
| 27 |
-
"
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"
|
| 31 |
-
"
|
| 32 |
-
"
|
| 33 |
},
|
| 34 |
{
|
| 35 |
-
"
|
| 36 |
-
"
|
| 37 |
-
"
|
| 38 |
-
"
|
| 39 |
-
"
|
|
|
|
| 40 |
},
|
| 41 |
{
|
| 42 |
-
"
|
| 43 |
-
"
|
| 44 |
-
"
|
| 45 |
-
"
|
| 46 |
-
"
|
|
|
|
| 47 |
},
|
| 48 |
{
|
| 49 |
-
"github_repo_link": "https://github.com/pytorch/helion",
|
| 50 |
"repo_name": "helion",
|
|
|
|
|
|
|
| 51 |
"repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
|
| 52 |
-
"
|
|
|
|
| 53 |
},
|
| 54 |
{
|
| 55 |
-
"github_repo_link": "https://github.com/microsoft/TileIR",
|
| 56 |
"repo_name": "TileIR",
|
| 57 |
-
"
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
"
|
| 61 |
-
"repo_name": "tilelang",
|
| 62 |
-
"repo_description": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
|
| 63 |
-
"homepage_link": "https://tilelang.com/",
|
| 64 |
-
"category": "dsl"
|
| 65 |
},
|
| 66 |
{
|
| 67 |
-
"
|
| 68 |
-
"
|
| 69 |
-
"repo_description": "
|
| 70 |
-
|
| 71 |
-
"github_topic_closest_fit": "cuda"
|
| 72 |
-
},
|
| 73 |
{
|
| 74 |
-
"github_repo_link": "https://github.com/tensorflow/tensorflow",
|
| 75 |
"repo_name": "tensorflow",
|
|
|
|
|
|
|
| 76 |
"repo_description": "An Open Source Machine Learning Framework for Everyone",
|
| 77 |
"homepage_link": "https://tensorflow.org",
|
| 78 |
-
"github_topic_closest_fit": "
|
| 79 |
-
"category": "machine learning framework"
|
| 80 |
},
|
| 81 |
{
|
| 82 |
-
"
|
| 83 |
-
"
|
| 84 |
-
"
|
|
|
|
|
|
|
|
|
|
| 85 |
},
|
| 86 |
{
|
| 87 |
-
"
|
| 88 |
-
"
|
| 89 |
-
"
|
| 90 |
-
"
|
| 91 |
-
"
|
| 92 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 93 |
},
|
| 94 |
{
|
| 95 |
-
"github_repo_link": "https://github.com/onnx/onnx",
|
| 96 |
"repo_name": "onnx",
|
|
|
|
|
|
|
| 97 |
"repo_description": "Open standard for machine learning interoperability",
|
| 98 |
"homepage_link": "https://onnx.ai/",
|
| 99 |
-
"github_topic_closest_fit": "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 100 |
},
|
| 101 |
{
|
| 102 |
"github_repo_link": "https://github.com/ray-project/ray",
|
| 103 |
"repo_name": "ray",
|
| 104 |
"repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
|
| 105 |
"homepage_link": "https://ray.io",
|
| 106 |
-
"github_topic_closest_fit": "
|
| 107 |
},
|
| 108 |
{
|
| 109 |
"github_repo_link": "https://github.com/jax-ml/jax",
|
|
@@ -553,7 +570,7 @@
|
|
| 553 |
"repo_name": "FTorch",
|
| 554 |
"repo_description": "A library for directly calling PyTorch ML models from Fortran.",
|
| 555 |
"homepage_link": "https://cambridge-iccs.github.io/FTorch/",
|
| 556 |
-
"github_topic_closest_fit": "
|
| 557 |
},
|
| 558 |
{
|
| 559 |
"github_repo_link": "https://github.com/facebook/hhvm",
|
|
|
|
| 1 |
[
|
| 2 |
{
|
|
|
|
| 3 |
"repo_name": "pytorch",
|
| 4 |
+
"github_repo_link": "https://github.com/pytorch/pytorch",
|
| 5 |
+
"category": "machine learning framework",
|
| 6 |
"repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
|
| 7 |
"homepage_link": "https://pytorch.org",
|
| 8 |
+
"github_topic_closest_fit": "machine-learning"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
},
|
| 10 |
{
|
| 11 |
+
"repo_name": "triton",
|
| 12 |
+
"github_repo_link": "https://github.com/triton-lang/triton",
|
| 13 |
+
"category": "parallel computing dsl",
|
| 14 |
+
"repo_description": "Development repository for the Triton language and compiler",
|
| 15 |
+
"homepage_link": "https://triton-lang.org/",
|
| 16 |
+
"github_topic_closest_fit": "parallel-programming"
|
| 17 |
},
|
| 18 |
{
|
| 19 |
+
"repo_name": "cutlass",
|
| 20 |
+
"github_repo_link": "https://github.com/NVIDIA/cutlass",
|
| 21 |
+
"category": "parallel computing",
|
| 22 |
+
"repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
|
| 23 |
+
"homepage_link": "https://docs.nvidia.com/cutlass/index.html",
|
| 24 |
+
"github_topic_closest_fit": "parallel-programming"
|
| 25 |
},
|
| 26 |
{
|
| 27 |
+
"repo_name": "tilelang",
|
| 28 |
+
"github_repo_link": "https://github.com/tile-ai/tilelang",
|
| 29 |
+
"category": "parallel computing dsl",
|
| 30 |
+
"repo_description": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
|
| 31 |
+
"homepage_link": "https://tilelang.com",
|
| 32 |
+
"github_topic_closest_fit": "parallel-programming"
|
| 33 |
},
|
| 34 |
{
|
| 35 |
+
"repo_name": "ThunderKittens",
|
| 36 |
+
"github_repo_link": "https://github.com/HazyResearch/ThunderKittens",
|
| 37 |
+
"category": "parallel computing",
|
| 38 |
+
"repo_description": "Tile primitives for speedy kernels",
|
| 39 |
+
"homepage_link": "https://hazyresearch.stanford.edu/blog/2024-10-29-tk2",
|
| 40 |
+
"github_topic_closest_fit": "parallel-programming"
|
| 41 |
},
|
| 42 |
{
|
|
|
|
| 43 |
"repo_name": "helion",
|
| 44 |
+
"github_repo_link": "https://github.com/pytorch/helion",
|
| 45 |
+
"category": "parallel computing dsl",
|
| 46 |
"repo_description": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
|
| 47 |
+
"homepage_link": "https://helionlang.com",
|
| 48 |
+
"github_topic_closest_fit": "parallel-programming"
|
| 49 |
},
|
| 50 |
{
|
|
|
|
| 51 |
"repo_name": "TileIR",
|
| 52 |
+
"github_repo_link": "https://github.com/microsoft/TileIR",
|
| 53 |
+
"category": "parallel computing dsl",
|
| 54 |
+
"repo_description": "TileIR (tile-ir) is a concise domain-specific IR designed to streamline the development of high-performance GPU/CPU kernels (e.g., GEMM, Dequant GEMM, FlashAttention, LinearAttention). By employing a Pythonic syntax with an underlying compiler infrastructure on top of TVM, TileIR allows developers to focus on productivity without sacrificing the low-level optimizations necessary for state-of-the-art performance.",
|
| 55 |
+
"github_topic_closest_fit": "parallel-programming"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 56 |
},
|
| 57 |
{
|
| 58 |
+
"repo_name": "BitBLAS",
|
| 59 |
+
"github_repo_link": "https://github.com/microsoft/BitBLAS",
|
| 60 |
+
"repo_description": "BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment."
|
| 61 |
+
},
|
|
|
|
|
|
|
| 62 |
{
|
|
|
|
| 63 |
"repo_name": "tensorflow",
|
| 64 |
+
"github_repo_link": "https://github.com/tensorflow/tensorflow",
|
| 65 |
+
"category": "machine learning framework",
|
| 66 |
"repo_description": "An Open Source Machine Learning Framework for Everyone",
|
| 67 |
"homepage_link": "https://tensorflow.org",
|
| 68 |
+
"github_topic_closest_fit": "machine-learning"
|
|
|
|
| 69 |
},
|
| 70 |
{
|
| 71 |
+
"repo_name": "vllm",
|
| 72 |
+
"github_repo_link": "https://github.com/vllm-project/vllm",
|
| 73 |
+
"category": "inference engine",
|
| 74 |
+
"repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
|
| 75 |
+
"homepage_link": "https://docs.vllm.ai",
|
| 76 |
+
"github_topic_closest_fit": "inference"
|
| 77 |
},
|
| 78 |
{
|
| 79 |
+
"repo_name": "ollama",
|
| 80 |
+
"github_repo_link": "https://github.com/ollama/ollama",
|
| 81 |
+
"category": "inference engine",
|
| 82 |
+
"repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
|
| 83 |
+
"homepage_link": "https://ollama.com",
|
| 84 |
+
"github_topic_closest_fit": "inference"
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"repo_name": "llama.cpp",
|
| 88 |
+
"github_repo_link": "https://github.com/ggml-org/llama.cpp",
|
| 89 |
+
"category": "inference engine",
|
| 90 |
+
"repo_description": "LLM inference in C/C++",
|
| 91 |
+
"homepage_link": "https://ggml.ai",
|
| 92 |
+
"github_topic_closest_fit": "inference"
|
| 93 |
+
},
|
| 94 |
+
{
|
| 95 |
+
"repo_name": "sglang",
|
| 96 |
+
"github_repo_link": "https://github.com/sgl-project/sglang",
|
| 97 |
+
"category": "inference engine",
|
| 98 |
+
"repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
|
| 99 |
+
"homepage_link": "https://docs.sglang.ai",
|
| 100 |
+
"github_topic_closest_fit": "inference"
|
| 101 |
},
|
| 102 |
{
|
|
|
|
| 103 |
"repo_name": "onnx",
|
| 104 |
+
"github_repo_link": "https://github.com/onnx/onnx",
|
| 105 |
+
"category": "machine learning framework",
|
| 106 |
"repo_description": "Open standard for machine learning interoperability",
|
| 107 |
"homepage_link": "https://onnx.ai/",
|
| 108 |
+
"github_topic_closest_fit": "onnx"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"repo_name": "executorch",
|
| 112 |
+
"github_repo_link": "https://github.com/pytorch/executorch",
|
| 113 |
+
"category": "model compiler",
|
| 114 |
+
"repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
|
| 115 |
+
"homepage_link": "https://executorch.ai",
|
| 116 |
+
"github_topic_closest_fit": "compiler"
|
| 117 |
},
|
| 118 |
{
|
| 119 |
"github_repo_link": "https://github.com/ray-project/ray",
|
| 120 |
"repo_name": "ray",
|
| 121 |
"repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
|
| 122 |
"homepage_link": "https://ray.io",
|
| 123 |
+
"github_topic_closest_fit": "machine-learning"
|
| 124 |
},
|
| 125 |
{
|
| 126 |
"github_repo_link": "https://github.com/jax-ml/jax",
|
|
|
|
| 570 |
"repo_name": "FTorch",
|
| 571 |
"repo_description": "A library for directly calling PyTorch ML models from Fortran.",
|
| 572 |
"homepage_link": "https://cambridge-iccs.github.io/FTorch/",
|
| 573 |
+
"github_topic_closest_fit": "machine-learning"
|
| 574 |
},
|
| 575 |
{
|
| 576 |
"github_repo_link": "https://github.com/facebook/hhvm",
|