Commit ·
0c39f5f
1
Parent(s): d1a41fe
Primus-Turbo
Browse files
PyTorchConference2025_GithubRepos.json
CHANGED
|
@@ -985,6 +985,8 @@
|
|
| 985 |
{
|
| 986 |
"repo_name": "Primus-Turbo",
|
| 987 |
"repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
|
|
|
|
|
|
|
| 988 |
"contributors_all": 12,
|
| 989 |
"contributors_2025": 12,
|
| 990 |
"contributors_2024": 0,
|
|
@@ -1005,7 +1007,7 @@
|
|
| 1005 |
{
|
| 1006 |
"repo_name": "FTorch",
|
| 1007 |
"repo_link": "https://github.com/Cambridge-ICCS/FTorch",
|
| 1008 |
-
"category": "
|
| 1009 |
"github_about_section": "A library for directly calling PyTorch ML models from Fortran.",
|
| 1010 |
"homepage_link": "https://cambridge-iccs.github.io/FTorch",
|
| 1011 |
"github_topic_closest_fit": "machine-learning",
|
|
|
|
| 985 |
{
|
| 986 |
"repo_name": "Primus-Turbo",
|
| 987 |
"repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
|
| 988 |
+
"category": "training framework",
|
| 989 |
+
"github_about_section": "Primus-Turbo is a high-performance acceleration library dedicated to large-scale model training on AMD GPUs. Built and optimized for the AMD ROCm platform, it covers the full training stack — including core compute operators (GEMM, Attention, GroupedGEMM), communication primitives, optimizer modules, low-precision computation (FP8), and compute–communication overlap kernels.",
|
| 990 |
"contributors_all": 12,
|
| 991 |
"contributors_2025": 12,
|
| 992 |
"contributors_2024": 0,
|
|
|
|
| 1007 |
{
|
| 1008 |
"repo_name": "FTorch",
|
| 1009 |
"repo_link": "https://github.com/Cambridge-ICCS/FTorch",
|
| 1010 |
+
"category": "middleware",
|
| 1011 |
"github_about_section": "A library for directly calling PyTorch ML models from Fortran.",
|
| 1012 |
"homepage_link": "https://cambridge-iccs.github.io/FTorch",
|
| 1013 |
"github_topic_closest_fit": "machine-learning",
|