TylerHilbert commited on
Commit
eb16c60
·
1 Parent(s): 9f15961

Updated contributors* fields

Browse files
Files changed (1) hide show
  1. PyTorchConference2025_GithubRepos.json +141 -101
PyTorchConference2025_GithubRepos.json CHANGED
@@ -6,7 +6,7 @@
6
  "github_about_section": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
7
  "homepage_link": "http://llvm.org",
8
  "github_topic_closest_fit": "compiler",
9
- "contributors_all": 6680,
10
  "contributors_2025": 2378,
11
  "contributors_2024": 2130,
12
  "contributors_2023": 1920
@@ -18,7 +18,7 @@
18
  "github_about_section": "A high-throughput and memory-efficient inference and serving engine for LLMs",
19
  "homepage_link": "https://docs.vllm.ai",
20
  "github_topic_closest_fit": "inference",
21
- "contributors_all": 1885,
22
  "contributors_2025": 1369,
23
  "contributors_2024": 579,
24
  "contributors_2023": 145
@@ -30,7 +30,7 @@
30
  "github_about_section": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
31
  "homepage_link": "https://pytorch.org",
32
  "github_topic_closest_fit": "machine-learning",
33
- "contributors_all": 5434,
34
  "contributors_2025": 1187,
35
  "contributors_2024": 1090,
36
  "contributors_2023": 1024
@@ -42,7 +42,7 @@
42
  "github_about_section": "Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.",
43
  "homepage_link": "https://huggingface.co/transformers",
44
  "github_topic_closest_fit": "machine-learning",
45
- "contributors_all": 3582,
46
  "contributors_2025": 860,
47
  "contributors_2024": 769,
48
  "contributors_2023": 758
@@ -54,7 +54,7 @@
54
  "github_about_section": "SGLang is a fast serving framework for large language models and vision language models.",
55
  "homepage_link": "https://docs.sglang.ai",
56
  "github_topic_closest_fit": "inference",
57
- "contributors_all": 937,
58
  "contributors_2025": 796,
59
  "contributors_2024": 189,
60
  "contributors_2023": 1
@@ -66,7 +66,7 @@
66
  "github_about_section": "A virtual machine for executing programs written in Hack.",
67
  "homepage_link": "https://hhvm.com",
68
  "github_topic_closest_fit": "virtual-machine",
69
- "contributors_all": 2624,
70
  "contributors_2025": 692,
71
  "contributors_2024": 648,
72
  "contributors_2023": 604
@@ -78,7 +78,7 @@
78
  "github_about_section": "LLM inference in C/C++",
79
  "homepage_link": "https://ggml.ai",
80
  "github_topic_closest_fit": "inference",
81
- "contributors_all": 1374,
82
  "contributors_2025": 535,
83
  "contributors_2024": 575,
84
  "contributors_2023": 461
@@ -90,9 +90,9 @@
90
  "github_about_section": "Production-Grade Container Scheduling and Management",
91
  "homepage_link": "https://kubernetes.io",
92
  "github_topic_closest_fit": "kubernetes",
93
- "contributors_all": 5041,
94
- "contributors_2025": 509,
95
- "contributors_2024": 498,
96
  "contributors_2023": 565
97
  },
98
  {
@@ -102,8 +102,8 @@
102
  "github_about_section": "An Open Source Machine Learning Framework for Everyone",
103
  "homepage_link": "https://tensorflow.org",
104
  "github_topic_closest_fit": "machine-learning",
105
- "contributors_all": 4618,
106
- "contributors_2025": 500,
107
  "contributors_2024": 523,
108
  "contributors_2023": 630
109
  },
@@ -114,7 +114,7 @@
114
  "github_about_section": "verl: Volcano Engine Reinforcement Learning for LLMs",
115
  "homepage_link": "https://verl.readthedocs.io",
116
  "github_topic_closest_fit": "deep-reinforcement-learning",
117
- "contributors_all": 462,
118
  "contributors_2025": 454,
119
  "contributors_2024": 10,
120
  "contributors_2023": 0
@@ -126,10 +126,10 @@
126
  "github_about_section": "super repo for rocm systems projects",
127
  "homepage_link": "https://amd.com/en/products/software/rocm.html",
128
  "github_topic_closest_fit": "amd",
129
- "contributors_all": 1032,
130
- "contributors_2025": 440,
131
- "contributors_2024": 323,
132
- "contributors_2023": 204
133
  },
134
  {
135
  "repo_name": "ray",
@@ -138,7 +138,7 @@
138
  "github_about_section": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
139
  "homepage_link": "https://ray.io",
140
  "github_topic_closest_fit": "machine-learning",
141
- "contributors_all": 1381,
142
  "contributors_2025": 397,
143
  "contributors_2024": 223,
144
  "contributors_2023": 230
@@ -150,7 +150,7 @@
150
  "github_about_section": "Apache Spark - A unified analytics engine for large-scale data processing",
151
  "homepage_link": "https://spark.apache.org",
152
  "github_topic_closest_fit": "data-processing",
153
- "contributors_all": 3083,
154
  "contributors_2025": 322,
155
  "contributors_2024": 300,
156
  "contributors_2023": 336
@@ -162,7 +162,7 @@
162
  "github_about_section": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
163
  "homepage_link": "https://block.github.io/goose",
164
  "github_topic_closest_fit": "ai-agents",
165
- "contributors_all": 332,
166
  "contributors_2025": 319,
167
  "contributors_2024": 32,
168
  "contributors_2023": 0
@@ -174,7 +174,7 @@
174
  "github_about_section": "Free and Open Source, Distributed, RESTful Search Engine",
175
  "homepage_link": "https://elastic.co/products/elasticsearch",
176
  "github_topic_closest_fit": "search-engine",
177
- "contributors_all": 2297,
178
  "contributors_2025": 316,
179
  "contributors_2024": 284,
180
  "contributors_2023": 270
@@ -186,8 +186,8 @@
186
  "github_about_section": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
187
  "homepage_link": "https://docs.jax.dev",
188
  "github_topic_closest_fit": "scientific-computing",
189
- "contributors_all": 997,
190
- "contributors_2025": 312,
191
  "contributors_2024": 280,
192
  "contributors_2023": 202
193
  },
@@ -198,8 +198,8 @@
198
  "github_about_section": "Specification and documentation for the Model Context Protocol",
199
  "homepage_link": "https://modelcontextprotocol.io",
200
  "github_topic_closest_fit": "mcp",
201
- "contributors_all": 327,
202
- "contributors_2025": 298,
203
  "contributors_2024": 42,
204
  "contributors_2023": 0
205
  },
@@ -210,7 +210,7 @@
210
  "github_about_section": "On-device AI across mobile, embedded and edge for PyTorch",
211
  "homepage_link": "https://executorch.ai",
212
  "github_topic_closest_fit": "inference",
213
- "contributors_all": 437,
214
  "contributors_2025": 267,
215
  "contributors_2024": 243,
216
  "contributors_2023": 77
@@ -222,8 +222,8 @@
222
  "github_about_section": "The fundamental package for scientific computing with Python.",
223
  "homepage_link": "https://numpy.org",
224
  "github_topic_closest_fit": "scientific-computing",
225
- "contributors_all": 2172,
226
- "contributors_2025": 235,
227
  "contributors_2024": 233,
228
  "contributors_2023": 252
229
  },
@@ -234,7 +234,7 @@
234
  "github_about_section": "Development repository for the Triton language and compiler",
235
  "homepage_link": "https://triton-lang.org",
236
  "github_topic_closest_fit": "parallel-programming",
237
- "contributors_all": 522,
238
  "contributors_2025": 233,
239
  "contributors_2024": 206,
240
  "contributors_2023": 159
@@ -246,7 +246,7 @@
246
  "github_about_section": "The Modular Platform (includes MAX & Mojo)",
247
  "homepage_link": "https://docs.modular.com",
248
  "github_topic_closest_fit": "parallel-programming",
249
- "contributors_all": 366,
250
  "contributors_2025": 222,
251
  "contributors_2024": 205,
252
  "contributors_2023": 99
@@ -258,8 +258,8 @@
258
  "github_about_section": "SciPy library main repository",
259
  "homepage_link": "https://scipy.org",
260
  "github_topic_closest_fit": "scientific-computing",
261
- "contributors_all": 1973,
262
- "contributors_2025": 210,
263
  "contributors_2024": 251,
264
  "contributors_2023": 245
265
  },
@@ -270,7 +270,7 @@
270
  "github_about_section": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
271
  "homepage_link": "https://ollama.com",
272
  "github_topic_closest_fit": "inference",
273
- "contributors_all": 574,
274
  "contributors_2025": 202,
275
  "contributors_2024": 314,
276
  "contributors_2023": 97
@@ -282,7 +282,7 @@
282
  "github_about_section": "Train transformer language models with reinforcement learning.",
283
  "homepage_link": "http://hf.co/docs/trl",
284
  "github_topic_closest_fit": "reinforcement-learning",
285
- "contributors_all": 433,
286
  "contributors_2025": 189,
287
  "contributors_2024": 154,
288
  "contributors_2023": 122
@@ -294,7 +294,7 @@
294
  "github_about_section": "FlashInfer: Kernel Library for LLM Serving",
295
  "homepage_link": "https://flashinfer.ai",
296
  "github_topic_closest_fit": "attention",
297
- "contributors_all": 205,
298
  "contributors_2025": 158,
299
  "contributors_2024": 50,
300
  "contributors_2023": 11
@@ -305,7 +305,7 @@
305
  "category": "gpu kernels",
306
  "github_about_section": "AI Tensor Engine for ROCm",
307
  "homepage_link": "https://rocm.blogs.amd.com/software-tools-optimization/aiter-ai-tensor-engine/README.html",
308
- "contributors_all": 151,
309
  "contributors_2025": 145,
310
  "contributors_2024": 10,
311
  "contributors_2023": 0
@@ -316,7 +316,7 @@
316
  "category": "inference",
317
  "github_about_section": "Supercharge Your LLM with the Fastest KV Cache Layer",
318
  "homepage_link": "https://lmcache.ai",
319
- "contributors_all": 152,
320
  "contributors_2025": 144,
321
  "contributors_2024": 18,
322
  "contributors_2023": 0
@@ -328,7 +328,7 @@
328
  "github_about_section": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
329
  "homepage_link": "https://kvcache-ai.github.io/Mooncake",
330
  "github_topic_closest_fit": "inference",
331
- "contributors_all": 138,
332
  "contributors_2025": 133,
333
  "contributors_2024": 13,
334
  "contributors_2023": 0
@@ -339,7 +339,7 @@
339
  "category": "training framework",
340
  "github_about_section": "A PyTorch native platform for training generative AI models",
341
  "homepage_link": "https://arxiv.org/abs/2410.06511",
342
- "contributors_all": 145,
343
  "contributors_2025": 119,
344
  "contributors_2024": 43,
345
  "contributors_2023": 1
@@ -351,7 +351,7 @@
351
  "github_about_section": "PyTorch native quantization and sparsity for training and inference",
352
  "homepage_link": "https://pytorch.org/ao",
353
  "github_topic_closest_fit": "quantization",
354
- "contributors_all": 178,
355
  "contributors_2025": 114,
356
  "contributors_2024": 100,
357
  "contributors_2023": 5
@@ -363,7 +363,7 @@
363
  "github_about_section": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
364
  "homepage_link": "https://comfy.org",
365
  "github_topic_closest_fit": "stable-diffusion",
366
- "contributors_all": 278,
367
  "contributors_2025": 108,
368
  "contributors_2024": 119,
369
  "contributors_2023": 94
@@ -375,9 +375,9 @@
375
  "github_about_section": "Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
376
  "homepage_link": "https://docs.unsloth.ai",
377
  "github_topic_closest_fit": "fine-tuning",
378
- "contributors_all": 127,
379
- "contributors_2025": 102,
380
- "contributors_2024": 27,
381
  "contributors_2023": 3
382
  },
383
  {
@@ -386,7 +386,7 @@
386
  "category": "training framework",
387
  "github_about_section": "A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support.",
388
  "homepage_link": "https://huggingface.co/docs/accelerate",
389
- "contributors_all": 392,
390
  "contributors_2025": 97,
391
  "contributors_2024": 124,
392
  "contributors_2023": 149
@@ -409,7 +409,7 @@
409
  "category": "training framework",
410
  "github_about_section": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
411
  "homepage_link": "https://deepspeed.ai",
412
- "contributors_all": 442,
413
  "contributors_2025": 96,
414
  "contributors_2024": 134,
415
  "contributors_2023": 165
@@ -421,7 +421,7 @@
421
  "github_about_section": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
422
  "homepage_link": "https://milvus.io",
423
  "github_topic_closest_fit": "vector-search",
424
- "contributors_all": 387,
425
  "contributors_2025": 95,
426
  "contributors_2024": 84,
427
  "contributors_2023": 72
@@ -433,7 +433,7 @@
433
  "github_about_section": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
434
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
435
  "github_topic_closest_fit": "parallel-programming",
436
- "contributors_all": 238,
437
  "contributors_2025": 94,
438
  "contributors_2024": 64,
439
  "contributors_2023": 66
@@ -445,7 +445,7 @@
445
  "github_about_section": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
446
  "homepage_link": "https://tilelang.com",
447
  "github_topic_closest_fit": "parallel-programming",
448
- "contributors_all": 90,
449
  "contributors_2025": 89,
450
  "contributors_2024": 1,
451
  "contributors_2023": 0
@@ -456,7 +456,7 @@
456
  "category": "distributed computing",
457
  "github_about_section": "PyTorch Single Controller",
458
  "homepage_link": "https://meta-pytorch.org/monarch",
459
- "contributors_all": 85,
460
  "contributors_2025": 85,
461
  "contributors_2024": 0,
462
  "contributors_2023": 0
@@ -468,7 +468,7 @@
468
  "github_about_section": "Efficient Triton Kernels for LLM Training",
469
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
470
  "github_topic_closest_fit": "triton",
471
- "contributors_all": 120,
472
  "contributors_2025": 78,
473
  "contributors_2024": 61,
474
  "contributors_2023": 0
@@ -491,7 +491,7 @@
491
  "category": "fine tuning",
492
  "github_about_section": "PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
493
  "homepage_link": "https://huggingface.co/docs/peft",
494
- "contributors_all": 272,
495
  "contributors_2025": 69,
496
  "contributors_2024": 111,
497
  "contributors_2023": 115
@@ -502,7 +502,7 @@
502
  "category": "multi-purpose library",
503
  "github_about_section": "AMD ROCm Software - GitHub Home",
504
  "homepage_link": "https://rocm.docs.amd.com",
505
- "contributors_all": 166,
506
  "contributors_2025": 67,
507
  "contributors_2024": 61,
508
  "contributors_2023": 44
@@ -513,7 +513,7 @@
513
  "category": "mcp",
514
  "github_about_section": "Build effective agents using Model Context Protocol and simple workflow patterns",
515
  "github_topic_closest_fit": "mcp",
516
- "contributors_all": 63,
517
  "contributors_2025": 63,
518
  "contributors_2024": 1,
519
  "contributors_2023": 0
@@ -523,8 +523,8 @@
523
  "repo_link": "https://github.com/linux-rdma/rdma-core",
524
  "category": "systems level code",
525
  "github_about_section": "RDMA core userspace libraries and daemons",
526
- "contributors_all": 437,
527
- "contributors_2025": 58,
528
  "contributors_2024": 61,
529
  "contributors_2023": 66
530
  },
@@ -535,7 +535,7 @@
535
  "github_about_section": "Open standard for machine learning interoperability",
536
  "homepage_link": "https://onnx.ai",
537
  "github_topic_closest_fit": "onnx",
538
- "contributors_all": 370,
539
  "contributors_2025": 56,
540
  "contributors_2024": 45,
541
  "contributors_2023": 61
@@ -547,8 +547,8 @@
547
  "github_about_section": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
548
  "homepage_link": "https://docs.letta.com",
549
  "github_topic_closest_fit": "ai-agents",
550
- "contributors_all": 157,
551
- "contributors_2025": 56,
552
  "contributors_2024": 75,
553
  "contributors_2023": 47
554
  },
@@ -559,7 +559,7 @@
559
  "github_about_section": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
560
  "homepage_link": "https://helionlang.com",
561
  "github_topic_closest_fit": "parallel-programming",
562
- "contributors_all": 49,
563
  "contributors_2025": 49,
564
  "contributors_2024": 0,
565
  "contributors_2023": 0
@@ -570,7 +570,7 @@
570
  "category": "evolutionary algorithm",
571
  "github_about_section": "Open-source implementation of AlphaEvolve",
572
  "github_topic_closest_fit": "genetic-algorithm",
573
- "contributors_all": 46,
574
  "contributors_2025": 46,
575
  "contributors_2024": 0,
576
  "contributors_2023": 0
@@ -580,7 +580,7 @@
580
  "repo_link": "https://github.com/Lightning-AI/lightning-thunder",
581
  "category": "model compiler",
582
  "github_about_section": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.",
583
- "contributors_all": 76,
584
  "contributors_2025": 44,
585
  "contributors_2024": 47,
586
  "contributors_2023": 29
@@ -592,7 +592,7 @@
592
  "github_about_section": "The simplest way to serve AI/ML models in production",
593
  "homepage_link": "https://truss.baseten.co",
594
  "github_topic_closest_fit": "inference",
595
- "contributors_all": 72,
596
  "contributors_2025": 44,
597
  "contributors_2024": 30,
598
  "contributors_2023": 21
@@ -604,7 +604,7 @@
604
  "github_about_section": "CUDA Python: Performance meets Productivity",
605
  "homepage_link": "https://nvidia.github.io/cuda-python",
606
  "github_topic_closest_fit": "parallel-programming",
607
- "contributors_all": 48,
608
  "contributors_2025": 41,
609
  "contributors_2024": 12,
610
  "contributors_2023": 1
@@ -616,7 +616,7 @@
616
  "github_about_section": "A Python framework for accelerated simulation, data generation and spatial computing.",
617
  "homepage_link": "https://nvidia.github.io/warp",
618
  "github_topic_closest_fit": "physics-simulation",
619
- "contributors_all": 79,
620
  "contributors_2025": 40,
621
  "contributors_2024": 29,
622
  "contributors_2023": 17
@@ -627,7 +627,7 @@
627
  "category": "container orchestration",
628
  "github_about_section": "Build, Manage and Deploy AI/ML Systems",
629
  "homepage_link": "https://metaflow.org",
630
- "contributors_all": 121,
631
  "contributors_2025": 37,
632
  "contributors_2024": 35,
633
  "contributors_2023": 28
@@ -638,8 +638,8 @@
638
  "category": "compiler",
639
  "github_about_section": "NumPy aware dynamic Python compiler using LLVM",
640
  "homepage_link": "https://numba.pydata.org",
641
- "contributors_all": 430,
642
- "contributors_2025": 36,
643
  "contributors_2024": 32,
644
  "contributors_2023": 55
645
  },
@@ -661,7 +661,7 @@
661
  "category": "distributed computing",
662
  "github_about_section": "Distributed Compiler based on Triton for Parallel Systems",
663
  "homepage_link": "https://triton-distributed.readthedocs.io",
664
- "contributors_all": 30,
665
  "contributors_2025": 30,
666
  "contributors_2024": 0,
667
  "contributors_2023": 0
@@ -673,7 +673,7 @@
673
  "github_about_section": "Tile primitives for speedy kernels",
674
  "homepage_link": "https://hazyresearch.stanford.edu/blog/2024-10-29-tk2",
675
  "github_topic_closest_fit": "parallel-programming",
676
- "contributors_all": 34,
677
  "contributors_2025": 29,
678
  "contributors_2024": 13,
679
  "contributors_2023": 0
@@ -697,7 +697,7 @@
697
  "github_about_section": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
698
  "homepage_link": "http://docs.sglang.ai/ome",
699
  "github_topic_closest_fit": "k8s",
700
- "contributors_all": 28,
701
  "contributors_2025": 28,
702
  "contributors_2024": 0,
703
  "contributors_2023": 0
@@ -709,7 +709,7 @@
709
  "github_about_section": "The Triton Inference Server provides an optimized cloud and edge inferencing solution.",
710
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
711
  "github_topic_closest_fit": "inference",
712
- "contributors_all": 147,
713
  "contributors_2025": 24,
714
  "contributors_2024": 36,
715
  "contributors_2023": 34
@@ -720,7 +720,7 @@
720
  "category": "compiler",
721
  "github_about_section": "ccache - a fast compiler cache",
722
  "homepage_link": "https://ccache.dev",
723
- "contributors_all": 218,
724
  "contributors_2025": 20,
725
  "contributors_2024": 28,
726
  "contributors_2023": 22
@@ -732,9 +732,9 @@
732
  "github_about_section": "LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.",
733
  "homepage_link": "https://netlib.org/lapack",
734
  "github_topic_closest_fit": "linear-algebra",
735
- "contributors_all": 178,
736
- "contributors_2025": 20,
737
- "contributors_2024": 24,
738
  "contributors_2023": 42
739
  },
740
  {
@@ -742,7 +742,7 @@
742
  "repo_link": "https://github.com/Dao-AILab/quack",
743
  "category": "kernel examples",
744
  "github_about_section": "A Quirky Assortment of CuTe Kernels",
745
- "contributors_all": 17,
746
  "contributors_2025": 17,
747
  "contributors_2024": 0,
748
  "contributors_2023": 0
@@ -754,7 +754,7 @@
754
  "github_about_section": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
755
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench",
756
  "github_topic_closest_fit": "benchmark",
757
- "contributors_all": 19,
758
  "contributors_2025": 16,
759
  "contributors_2024": 3,
760
  "contributors_2023": 0
@@ -765,7 +765,7 @@
765
  "category": "kernel examples",
766
  "github_about_section": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
767
  "homepage_link": "https://gpumode.com",
768
- "contributors_all": 16,
769
  "contributors_2025": 16,
770
  "contributors_2024": 0,
771
  "contributors_2023": 0
@@ -788,7 +788,7 @@
788
  "category": "performance testing",
789
  "github_about_section": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
790
  "homepage_link": "https://meta-pytorch.org/tritonparse",
791
- "contributors_all": 15,
792
  "contributors_2025": 15,
793
  "contributors_2024": 0,
794
  "contributors_2023": 0
@@ -798,7 +798,7 @@
798
  "repo_link": "https://github.com/huggingface/kernels",
799
  "category": "gpu kernels",
800
  "github_about_section": "Load compute kernels from the Hub",
801
- "contributors_all": 15,
802
  "contributors_2025": 14,
803
  "contributors_2024": 2,
804
  "contributors_2023": 0
@@ -810,7 +810,7 @@
810
  "github_about_section": "Wan: Open and Advanced Large-Scale Video Generative Models",
811
  "homepage_link": "https://wan.video",
812
  "github_topic_closest_fit": "diffusion-models",
813
- "contributors_all": 14,
814
  "contributors_2025": 14,
815
  "contributors_2024": 0,
816
  "contributors_2023": 0
@@ -820,7 +820,7 @@
820
  "repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
821
  "category": "training framework",
822
  "github_about_section": "Primus-Turbo is a high-performance acceleration library dedicated to large-scale model training on AMD GPUs. Built and optimized for the AMD ROCm platform, it covers the full training stack — including core compute operators (GEMM, Attention, GroupedGEMM), communication primitives, optimizer modules, low-precision computation (FP8), and compute–communication overlap kernels.",
823
- "contributors_all": 12,
824
  "contributors_2025": 12,
825
  "contributors_2024": 0,
826
  "contributors_2023": 0
@@ -832,7 +832,7 @@
832
  "github_about_section": "Building the Virtuous Cycle for AI-driven LLM Systems",
833
  "homepage_link": "https://bench.flashinfer.ai",
834
  "github_topic_closest_fit": "benchmark",
835
- "contributors_all": 12,
836
  "contributors_2025": 11,
837
  "contributors_2024": 0,
838
  "contributors_2023": 0
@@ -844,8 +844,8 @@
844
  "github_about_section": "A library for directly calling PyTorch ML models from Fortran.",
845
  "homepage_link": "https://cambridge-iccs.github.io/FTorch",
846
  "github_topic_closest_fit": "machine-learning",
847
- "contributors_all": 20,
848
- "contributors_2025": 11,
849
  "contributors_2024": 8,
850
  "contributors_2023": 9
851
  },
@@ -877,7 +877,7 @@
877
  "category": "gpu kernels",
878
  "homepage_link": "https://huggingface.co/kernels-community",
879
  "github_about_section": "Kernel sources for https://huggingface.co/kernels-community",
880
- "contributors_all": 9,
881
  "contributors_2025": 9,
882
  "contributors_2024": 0,
883
  "contributors_2023": 0
@@ -888,7 +888,7 @@
888
  "category": "agent",
889
  "github_about_section": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
890
  "github_topic_closest_fit": "ai-agents",
891
- "contributors_all": 9,
892
  "contributors_2025": 9,
893
  "contributors_2024": 0,
894
  "contributors_2023": 0
@@ -912,7 +912,7 @@
912
  "github_about_section": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
913
  "homepage_link": "https://developer.nvidia.com/cudnn",
914
  "github_topic_closest_fit": "parallel-programming",
915
- "contributors_all": 12,
916
  "contributors_2025": 6,
917
  "contributors_2024": 5,
918
  "contributors_2023": 1
@@ -983,7 +983,7 @@
983
  "github_about_section": "Official inference library for Mistral models",
984
  "homepage_link": "https://mistral.ai",
985
  "github_topic_closest_fit": "inference",
986
- "contributors_all": 29,
987
  "contributors_2025": 2,
988
  "contributors_2024": 17,
989
  "contributors_2023": 14
@@ -1040,7 +1040,7 @@
1040
  "category": "debugger",
1041
  "github_about_section": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
1042
  "homepage_link": "https://triton-runner.org",
1043
- "contributors_all": 1,
1044
  "contributors_2025": 1,
1045
  "contributors_2024": 0,
1046
  "contributors_2023": 0
@@ -1074,66 +1074,106 @@
1074
  "repo_link": "https://github.com/wandb/wandb",
1075
  "category": "ml visualization",
1076
  "github_about_section": "The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models from experimentation to production.",
1077
- "homepage_link": "https://wandb.ai"
 
 
 
 
1078
  },
1079
  {
1080
  "repo_name": "aws-neuron-sdk",
1081
  "repo_link": "https://github.com/aws-neuron/aws-neuron-sdk",
1082
  "category": "sdk",
1083
  "github_about_section": "Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services",
1084
- "homepage_link": "https://aws.amazon.com/ai/machine-learning/neuron"
 
 
 
 
1085
  },
1086
  {
1087
  "repo_name": "onnxruntime",
1088
  "repo_link": "https://github.com/microsoft/onnxruntime",
1089
  "category": "machine learning interoperability",
1090
  "github_about_section": "ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator",
1091
- "homepage_link": "https://onnxruntime.ai"
 
 
 
 
1092
  },
1093
  {
1094
  "repo_name": "ort",
1095
  "repo_link": "https://github.com/pykeio/ort",
1096
  "category": "machine learning interoperability",
1097
  "github_about_section": "Fast ML inference & training for ONNX models in Rust",
1098
- "homepage_link": "https://ort.pyke.io"
 
 
 
 
1099
  },
1100
  {
1101
  "repo_name": "Triton-distributed",
1102
  "repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
1103
  "category": "distributed computing",
1104
  "github_about_section": "Distributed Compiler based on Triton for Parallel Systems",
1105
- "homepage_link": "https://triton-distributed.readthedocs.io"
 
 
 
 
1106
  },
1107
  {
1108
  "repo_name": "gemlite",
1109
  "repo_link": "https://github.com/dropbox/gemlite",
1110
  "category": "gpu kernels",
1111
- "github_about_section": "Fast low-bit matmul kernels in Triton"
 
 
 
 
1112
  },
1113
  {
1114
  "repo_name": "cutile-python",
1115
  "repo_link": "https://github.com/NVIDIA/cutile-python",
1116
  "category": "parallel computing",
1117
  "github_about_section": "cuTile is a programming model for writing parallel kernels for NVIDIA GPUs",
1118
- "homepage_link": "https://docs.nvidia.com/cuda/cutile-python"
 
 
 
 
1119
  },
1120
  {
1121
  "repo_name": "tilus",
1122
  "repo_link": "https://github.com/NVIDIA/tilus",
1123
  "category": "parallel computing",
1124
  "github_about_section": "Tilus is a tile-level kernel programming language with explicit control over shared memory and registers.",
1125
- "homepage_link": "https://nvidia.github.io/tilus"
 
 
 
 
1126
  },
1127
  {
1128
  "repo_name": "triton-windows",
1129
  "repo_link": "https://github.com/woct0rdho/triton-windows",
1130
  "category": "parallel computing dsl",
1131
- "github_about_section": "Fork of the Triton language and compiler for Windows support and easy installation"
 
 
 
 
1132
  },
1133
  {
1134
  "repo_name": "flash-linear-attention",
1135
  "repo_link": "https://github.com/fla-org/flash-linear-attention",
1136
  "category": "gpu kernels",
1137
- "github_about_section": "Efficient implementations of state-of-the-art linear attention models"
 
 
 
 
1138
  }
1139
  ]
 
6
  "github_about_section": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
7
  "homepage_link": "http://llvm.org",
8
  "github_topic_closest_fit": "compiler",
9
+ "contributors_all": 7043,
10
  "contributors_2025": 2378,
11
  "contributors_2024": 2130,
12
  "contributors_2023": 1920
 
18
  "github_about_section": "A high-throughput and memory-efficient inference and serving engine for LLMs",
19
  "homepage_link": "https://docs.vllm.ai",
20
  "github_topic_closest_fit": "inference",
21
+ "contributors_all": 2302,
22
  "contributors_2025": 1369,
23
  "contributors_2024": 579,
24
  "contributors_2023": 145
 
30
  "github_about_section": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
31
  "homepage_link": "https://pytorch.org",
32
  "github_topic_closest_fit": "machine-learning",
33
+ "contributors_all": 5662,
34
  "contributors_2025": 1187,
35
  "contributors_2024": 1090,
36
  "contributors_2023": 1024
 
42
  "github_about_section": "Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.",
43
  "homepage_link": "https://huggingface.co/transformers",
44
  "github_topic_closest_fit": "machine-learning",
45
+ "contributors_all": 3728,
46
  "contributors_2025": 860,
47
  "contributors_2024": 769,
48
  "contributors_2023": 758
 
54
  "github_about_section": "SGLang is a fast serving framework for large language models and vision language models.",
55
  "homepage_link": "https://docs.sglang.ai",
56
  "github_topic_closest_fit": "inference",
57
+ "contributors_all": 1229,
58
  "contributors_2025": 796,
59
  "contributors_2024": 189,
60
  "contributors_2023": 1
 
66
  "github_about_section": "A virtual machine for executing programs written in Hack.",
67
  "homepage_link": "https://hhvm.com",
68
  "github_topic_closest_fit": "virtual-machine",
69
+ "contributors_all": 2773,
70
  "contributors_2025": 692,
71
  "contributors_2024": 648,
72
  "contributors_2023": 604
 
78
  "github_about_section": "LLM inference in C/C++",
79
  "homepage_link": "https://ggml.ai",
80
  "github_topic_closest_fit": "inference",
81
+ "contributors_all": 1540,
82
  "contributors_2025": 535,
83
  "contributors_2024": 575,
84
  "contributors_2023": 461
 
90
  "github_about_section": "Production-Grade Container Scheduling and Management",
91
  "homepage_link": "https://kubernetes.io",
92
  "github_topic_closest_fit": "kubernetes",
93
+ "contributors_all": 5157,
94
+ "contributors_2025": 541,
95
+ "contributors_2024": 499,
96
  "contributors_2023": 565
97
  },
98
  {
 
102
  "github_about_section": "An Open Source Machine Learning Framework for Everyone",
103
  "homepage_link": "https://tensorflow.org",
104
  "github_topic_closest_fit": "machine-learning",
105
+ "contributors_all": 4672,
106
+ "contributors_2025": 506,
107
  "contributors_2024": 523,
108
  "contributors_2023": 630
109
  },
 
114
  "github_about_section": "verl: Volcano Engine Reinforcement Learning for LLMs",
115
  "homepage_link": "https://verl.readthedocs.io",
116
  "github_topic_closest_fit": "deep-reinforcement-learning",
117
+ "contributors_all": 567,
118
  "contributors_2025": 454,
119
  "contributors_2024": 10,
120
  "contributors_2023": 0
 
126
  "github_about_section": "super repo for rocm systems projects",
127
  "homepage_link": "https://amd.com/en/products/software/rocm.html",
128
  "github_topic_closest_fit": "amd",
129
+ "contributors_all": 1142,
130
+ "contributors_2025": 486,
131
+ "contributors_2024": 351,
132
+ "contributors_2023": 213
133
  },
134
  {
135
  "repo_name": "ray",
 
138
  "github_about_section": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
139
  "homepage_link": "https://ray.io",
140
  "github_topic_closest_fit": "machine-learning",
141
+ "contributors_all": 1458,
142
  "contributors_2025": 397,
143
  "contributors_2024": 223,
144
  "contributors_2023": 230
 
150
  "github_about_section": "Apache Spark - A unified analytics engine for large-scale data processing",
151
  "homepage_link": "https://spark.apache.org",
152
  "github_topic_closest_fit": "data-processing",
153
+ "contributors_all": 3129,
154
  "contributors_2025": 322,
155
  "contributors_2024": 300,
156
  "contributors_2023": 336
 
162
  "github_about_section": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
163
  "homepage_link": "https://block.github.io/goose",
164
  "github_topic_closest_fit": "ai-agents",
165
+ "contributors_all": 417,
166
  "contributors_2025": 319,
167
  "contributors_2024": 32,
168
  "contributors_2023": 0
 
174
  "github_about_section": "Free and Open Source, Distributed, RESTful Search Engine",
175
  "homepage_link": "https://elastic.co/products/elasticsearch",
176
  "github_topic_closest_fit": "search-engine",
177
+ "contributors_all": 2337,
178
  "contributors_2025": 316,
179
  "contributors_2024": 284,
180
  "contributors_2023": 270
 
186
  "github_about_section": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
187
  "homepage_link": "https://docs.jax.dev",
188
  "github_topic_closest_fit": "scientific-computing",
189
+ "contributors_all": 1033,
190
+ "contributors_2025": 316,
191
  "contributors_2024": 280,
192
  "contributors_2023": 202
193
  },
 
198
  "github_about_section": "Specification and documentation for the Model Context Protocol",
199
  "homepage_link": "https://modelcontextprotocol.io",
200
  "github_topic_closest_fit": "mcp",
201
+ "contributors_all": 361,
202
+ "contributors_2025": 301,
203
  "contributors_2024": 42,
204
  "contributors_2023": 0
205
  },
 
210
  "github_about_section": "On-device AI across mobile, embedded and edge for PyTorch",
211
  "homepage_link": "https://executorch.ai",
212
  "github_topic_closest_fit": "inference",
213
+ "contributors_all": 493,
214
  "contributors_2025": 267,
215
  "contributors_2024": 243,
216
  "contributors_2023": 77
 
222
  "github_about_section": "The fundamental package for scientific computing with Python.",
223
  "homepage_link": "https://numpy.org",
224
  "github_topic_closest_fit": "scientific-computing",
225
+ "contributors_all": 2210,
226
+ "contributors_2025": 237,
227
  "contributors_2024": 233,
228
  "contributors_2023": 252
229
  },
 
234
  "github_about_section": "Development repository for the Triton language and compiler",
235
  "homepage_link": "https://triton-lang.org",
236
  "github_topic_closest_fit": "parallel-programming",
237
+ "contributors_all": 558,
238
  "contributors_2025": 233,
239
  "contributors_2024": 206,
240
  "contributors_2023": 159
 
246
  "github_about_section": "The Modular Platform (includes MAX & Mojo)",
247
  "homepage_link": "https://docs.modular.com",
248
  "github_topic_closest_fit": "parallel-programming",
249
+ "contributors_all": 418,
250
  "contributors_2025": 222,
251
  "contributors_2024": 205,
252
  "contributors_2023": 99
 
258
  "github_about_section": "SciPy library main repository",
259
  "homepage_link": "https://scipy.org",
260
  "github_topic_closest_fit": "scientific-computing",
261
+ "contributors_all": 2008,
262
+ "contributors_2025": 213,
263
  "contributors_2024": 251,
264
  "contributors_2023": 245
265
  },
 
270
  "github_about_section": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
271
  "homepage_link": "https://ollama.com",
272
  "github_topic_closest_fit": "inference",
273
+ "contributors_all": 596,
274
  "contributors_2025": 202,
275
  "contributors_2024": 314,
276
  "contributors_2023": 97
 
282
  "github_about_section": "Train transformer language models with reinforcement learning.",
283
  "homepage_link": "http://hf.co/docs/trl",
284
  "github_topic_closest_fit": "reinforcement-learning",
285
+ "contributors_all": 472,
286
  "contributors_2025": 189,
287
  "contributors_2024": 154,
288
  "contributors_2023": 122
 
294
  "github_about_section": "FlashInfer: Kernel Library for LLM Serving",
295
  "homepage_link": "https://flashinfer.ai",
296
  "github_topic_closest_fit": "attention",
297
+ "contributors_all": 259,
298
  "contributors_2025": 158,
299
  "contributors_2024": 50,
300
  "contributors_2023": 11
 
305
  "category": "gpu kernels",
306
  "github_about_section": "AI Tensor Engine for ROCm",
307
  "homepage_link": "https://rocm.blogs.amd.com/software-tools-optimization/aiter-ai-tensor-engine/README.html",
308
+ "contributors_all": 216,
309
  "contributors_2025": 145,
310
  "contributors_2024": 10,
311
  "contributors_2023": 0
 
316
  "category": "inference",
317
  "github_about_section": "Supercharge Your LLM with the Fastest KV Cache Layer",
318
  "homepage_link": "https://lmcache.ai",
319
+ "contributors_all": 174,
320
  "contributors_2025": 144,
321
  "contributors_2024": 18,
322
  "contributors_2023": 0
 
328
  "github_about_section": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
329
  "homepage_link": "https://kvcache-ai.github.io/Mooncake",
330
  "github_topic_closest_fit": "inference",
331
+ "contributors_all": 188,
332
  "contributors_2025": 133,
333
  "contributors_2024": 13,
334
  "contributors_2023": 0
 
339
  "category": "training framework",
340
  "github_about_section": "A PyTorch native platform for training generative AI models",
341
  "homepage_link": "https://arxiv.org/abs/2410.06511",
342
+ "contributors_all": 181,
343
  "contributors_2025": 119,
344
  "contributors_2024": 43,
345
  "contributors_2023": 1
 
351
  "github_about_section": "PyTorch native quantization and sparsity for training and inference",
352
  "homepage_link": "https://pytorch.org/ao",
353
  "github_topic_closest_fit": "quantization",
354
+ "contributors_all": 211,
355
  "contributors_2025": 114,
356
  "contributors_2024": 100,
357
  "contributors_2023": 5
 
363
  "github_about_section": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
364
  "homepage_link": "https://comfy.org",
365
  "github_topic_closest_fit": "stable-diffusion",
366
+ "contributors_all": 309,
367
  "contributors_2025": 108,
368
  "contributors_2024": 119,
369
  "contributors_2023": 94
 
375
  "github_about_section": "Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
376
  "homepage_link": "https://docs.unsloth.ai",
377
  "github_topic_closest_fit": "fine-tuning",
378
+ "contributors_all": 175,
379
+ "contributors_2025": 108,
380
+ "contributors_2024": 29,
381
  "contributors_2023": 3
382
  },
383
  {
 
386
  "category": "training framework",
387
  "github_about_section": "A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support.",
388
  "homepage_link": "https://huggingface.co/docs/accelerate",
389
+ "contributors_all": 410,
390
  "contributors_2025": 97,
391
  "contributors_2024": 124,
392
  "contributors_2023": 149
 
409
  "category": "training framework",
410
  "github_about_section": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
411
  "homepage_link": "https://deepspeed.ai",
412
+ "contributors_all": 458,
413
  "contributors_2025": 96,
414
  "contributors_2024": 134,
415
  "contributors_2023": 165
 
421
  "github_about_section": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
422
  "homepage_link": "https://milvus.io",
423
  "github_topic_closest_fit": "vector-search",
424
+ "contributors_all": 398,
425
  "contributors_2025": 95,
426
  "contributors_2024": 84,
427
  "contributors_2023": 72
 
433
  "github_about_section": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
434
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
435
  "github_topic_closest_fit": "parallel-programming",
436
+ "contributors_all": 264,
437
  "contributors_2025": 94,
438
  "contributors_2024": 64,
439
  "contributors_2023": 66
 
445
  "github_about_section": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
446
  "homepage_link": "https://tilelang.com",
447
  "github_topic_closest_fit": "parallel-programming",
448
+ "contributors_all": 117,
449
  "contributors_2025": 89,
450
  "contributors_2024": 1,
451
  "contributors_2023": 0
 
456
  "category": "distributed computing",
457
  "github_about_section": "PyTorch Single Controller",
458
  "homepage_link": "https://meta-pytorch.org/monarch",
459
+ "contributors_all": 101,
460
  "contributors_2025": 85,
461
  "contributors_2024": 0,
462
  "contributors_2023": 0
 
468
  "github_about_section": "Efficient Triton Kernels for LLM Training",
469
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
470
  "github_topic_closest_fit": "triton",
471
+ "contributors_all": 138,
472
  "contributors_2025": 78,
473
  "contributors_2024": 61,
474
  "contributors_2023": 0
 
491
  "category": "fine tuning",
492
  "github_about_section": "PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
493
  "homepage_link": "https://huggingface.co/docs/peft",
494
+ "contributors_all": 286,
495
  "contributors_2025": 69,
496
  "contributors_2024": 111,
497
  "contributors_2023": 115
 
502
  "category": "multi-purpose library",
503
  "github_about_section": "AMD ROCm Software - GitHub Home",
504
  "homepage_link": "https://rocm.docs.amd.com",
505
+ "contributors_all": 167,
506
  "contributors_2025": 67,
507
  "contributors_2024": 61,
508
  "contributors_2023": 44
 
513
  "category": "mcp",
514
  "github_about_section": "Build effective agents using Model Context Protocol and simple workflow patterns",
515
  "github_topic_closest_fit": "mcp",
516
+ "contributors_all": 64,
517
  "contributors_2025": 63,
518
  "contributors_2024": 1,
519
  "contributors_2023": 0
 
523
  "repo_link": "https://github.com/linux-rdma/rdma-core",
524
  "category": "systems level code",
525
  "github_about_section": "RDMA core userspace libraries and daemons",
526
+ "contributors_all": 441,
527
+ "contributors_2025": 62,
528
  "contributors_2024": 61,
529
  "contributors_2023": 66
530
  },
 
535
  "github_about_section": "Open standard for machine learning interoperability",
536
  "homepage_link": "https://onnx.ai",
537
  "github_topic_closest_fit": "onnx",
538
+ "contributors_all": 380,
539
  "contributors_2025": 56,
540
  "contributors_2024": 45,
541
  "contributors_2023": 61
 
547
  "github_about_section": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
548
  "homepage_link": "https://docs.letta.com",
549
  "github_topic_closest_fit": "ai-agents",
550
+ "contributors_all": 159,
551
+ "contributors_2025": 57,
552
  "contributors_2024": 75,
553
  "contributors_2023": 47
554
  },
 
559
  "github_about_section": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
560
  "homepage_link": "https://helionlang.com",
561
  "github_topic_closest_fit": "parallel-programming",
562
+ "contributors_all": 66,
563
  "contributors_2025": 49,
564
  "contributors_2024": 0,
565
  "contributors_2023": 0
 
570
  "category": "evolutionary algorithm",
571
  "github_about_section": "Open-source implementation of AlphaEvolve",
572
  "github_topic_closest_fit": "genetic-algorithm",
573
+ "contributors_all": 51,
574
  "contributors_2025": 46,
575
  "contributors_2024": 0,
576
  "contributors_2023": 0
 
580
  "repo_link": "https://github.com/Lightning-AI/lightning-thunder",
581
  "category": "model compiler",
582
  "github_about_section": "PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.",
583
+ "contributors_all": 79,
584
  "contributors_2025": 44,
585
  "contributors_2024": 47,
586
  "contributors_2023": 29
 
592
  "github_about_section": "The simplest way to serve AI/ML models in production",
593
  "homepage_link": "https://truss.baseten.co",
594
  "github_topic_closest_fit": "inference",
595
+ "contributors_all": 84,
596
  "contributors_2025": 44,
597
  "contributors_2024": 30,
598
  "contributors_2023": 21
 
604
  "github_about_section": "CUDA Python: Performance meets Productivity",
605
  "homepage_link": "https://nvidia.github.io/cuda-python",
606
  "github_topic_closest_fit": "parallel-programming",
607
+ "contributors_all": 54,
608
  "contributors_2025": 41,
609
  "contributors_2024": 12,
610
  "contributors_2023": 1
 
616
  "github_about_section": "A Python framework for accelerated simulation, data generation and spatial computing.",
617
  "homepage_link": "https://nvidia.github.io/warp",
618
  "github_topic_closest_fit": "physics-simulation",
619
+ "contributors_all": 89,
620
  "contributors_2025": 40,
621
  "contributors_2024": 29,
622
  "contributors_2023": 17
 
627
  "category": "container orchestration",
628
  "github_about_section": "Build, Manage and Deploy AI/ML Systems",
629
  "homepage_link": "https://metaflow.org",
630
+ "contributors_all": 132,
631
  "contributors_2025": 37,
632
  "contributors_2024": 35,
633
  "contributors_2023": 28
 
638
  "category": "compiler",
639
  "github_about_section": "NumPy aware dynamic Python compiler using LLVM",
640
  "homepage_link": "https://numba.pydata.org",
641
+ "contributors_all": 446,
642
+ "contributors_2025": 40,
643
  "contributors_2024": 32,
644
  "contributors_2023": 55
645
  },
 
661
  "category": "distributed computing",
662
  "github_about_section": "Distributed Compiler based on Triton for Parallel Systems",
663
  "homepage_link": "https://triton-distributed.readthedocs.io",
664
+ "contributors_all": 35,
665
  "contributors_2025": 30,
666
  "contributors_2024": 0,
667
  "contributors_2023": 0
 
673
  "github_about_section": "Tile primitives for speedy kernels",
674
  "homepage_link": "https://hazyresearch.stanford.edu/blog/2024-10-29-tk2",
675
  "github_topic_closest_fit": "parallel-programming",
676
+ "contributors_all": 37,
677
  "contributors_2025": 29,
678
  "contributors_2024": 13,
679
  "contributors_2023": 0
 
697
  "github_about_section": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
698
  "homepage_link": "http://docs.sglang.ai/ome",
699
  "github_topic_closest_fit": "k8s",
700
+ "contributors_all": 31,
701
  "contributors_2025": 28,
702
  "contributors_2024": 0,
703
  "contributors_2023": 0
 
709
  "github_about_section": "The Triton Inference Server provides an optimized cloud and edge inferencing solution.",
710
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
711
  "github_topic_closest_fit": "inference",
712
+ "contributors_all": 149,
713
  "contributors_2025": 24,
714
  "contributors_2024": 36,
715
  "contributors_2023": 34
 
720
  "category": "compiler",
721
  "github_about_section": "ccache - a fast compiler cache",
722
  "homepage_link": "https://ccache.dev",
723
+ "contributors_all": 224,
724
  "contributors_2025": 20,
725
  "contributors_2024": 28,
726
  "contributors_2023": 22
 
732
  "github_about_section": "LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.",
733
  "homepage_link": "https://netlib.org/lapack",
734
  "github_topic_closest_fit": "linear-algebra",
735
+ "contributors_all": 187,
736
+ "contributors_2025": 23,
737
+ "contributors_2024": 25,
738
  "contributors_2023": 42
739
  },
740
  {
 
742
  "repo_link": "https://github.com/Dao-AILab/quack",
743
  "category": "kernel examples",
744
  "github_about_section": "A Quirky Assortment of CuTe Kernels",
745
+ "contributors_all": 31,
746
  "contributors_2025": 17,
747
  "contributors_2024": 0,
748
  "contributors_2023": 0
 
754
  "github_about_section": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
755
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench",
756
  "github_topic_closest_fit": "benchmark",
757
+ "contributors_all": 21,
758
  "contributors_2025": 16,
759
  "contributors_2024": 3,
760
  "contributors_2023": 0
 
765
  "category": "kernel examples",
766
  "github_about_section": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
767
  "homepage_link": "https://gpumode.com",
768
+ "contributors_all": 23,
769
  "contributors_2025": 16,
770
  "contributors_2024": 0,
771
  "contributors_2023": 0
 
788
  "category": "performance testing",
789
  "github_about_section": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
790
  "homepage_link": "https://meta-pytorch.org/tritonparse",
791
+ "contributors_all": 26,
792
  "contributors_2025": 15,
793
  "contributors_2024": 0,
794
  "contributors_2023": 0
 
798
  "repo_link": "https://github.com/huggingface/kernels",
799
  "category": "gpu kernels",
800
  "github_about_section": "Load compute kernels from the Hub",
801
+ "contributors_all": 23,
802
  "contributors_2025": 14,
803
  "contributors_2024": 2,
804
  "contributors_2023": 0
 
810
  "github_about_section": "Wan: Open and Advanced Large-Scale Video Generative Models",
811
  "homepage_link": "https://wan.video",
812
  "github_topic_closest_fit": "diffusion-models",
813
+ "contributors_all": 16,
814
  "contributors_2025": 14,
815
  "contributors_2024": 0,
816
  "contributors_2023": 0
 
820
  "repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
821
  "category": "training framework",
822
  "github_about_section": "Primus-Turbo is a high-performance acceleration library dedicated to large-scale model training on AMD GPUs. Built and optimized for the AMD ROCm platform, it covers the full training stack — including core compute operators (GEMM, Attention, GroupedGEMM), communication primitives, optimizer modules, low-precision computation (FP8), and compute–communication overlap kernels.",
823
+ "contributors_all": 14,
824
  "contributors_2025": 12,
825
  "contributors_2024": 0,
826
  "contributors_2023": 0
 
832
  "github_about_section": "Building the Virtuous Cycle for AI-driven LLM Systems",
833
  "homepage_link": "https://bench.flashinfer.ai",
834
  "github_topic_closest_fit": "benchmark",
835
+ "contributors_all": 16,
836
  "contributors_2025": 11,
837
  "contributors_2024": 0,
838
  "contributors_2023": 0
 
844
  "github_about_section": "A library for directly calling PyTorch ML models from Fortran.",
845
  "homepage_link": "https://cambridge-iccs.github.io/FTorch",
846
  "github_topic_closest_fit": "machine-learning",
847
+ "contributors_all": 22,
848
+ "contributors_2025": 12,
849
  "contributors_2024": 8,
850
  "contributors_2023": 9
851
  },
 
877
  "category": "gpu kernels",
878
  "homepage_link": "https://huggingface.co/kernels-community",
879
  "github_about_section": "Kernel sources for https://huggingface.co/kernels-community",
880
+ "contributors_all": 14,
881
  "contributors_2025": 9,
882
  "contributors_2024": 0,
883
  "contributors_2023": 0
 
888
  "category": "agent",
889
  "github_about_section": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
890
  "github_topic_closest_fit": "ai-agents",
891
+ "contributors_all": 17,
892
  "contributors_2025": 9,
893
  "contributors_2024": 0,
894
  "contributors_2023": 0
 
912
  "github_about_section": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
913
  "homepage_link": "https://developer.nvidia.com/cudnn",
914
  "github_topic_closest_fit": "parallel-programming",
915
+ "contributors_all": 14,
916
  "contributors_2025": 6,
917
  "contributors_2024": 5,
918
  "contributors_2023": 1
 
983
  "github_about_section": "Official inference library for Mistral models",
984
  "homepage_link": "https://mistral.ai",
985
  "github_topic_closest_fit": "inference",
986
+ "contributors_all": 30,
987
  "contributors_2025": 2,
988
  "contributors_2024": 17,
989
  "contributors_2023": 14
 
1040
  "category": "debugger",
1041
  "github_about_section": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
1042
  "homepage_link": "https://triton-runner.org",
1043
+ "contributors_all": 2,
1044
  "contributors_2025": 1,
1045
  "contributors_2024": 0,
1046
  "contributors_2023": 0
 
1074
  "repo_link": "https://github.com/wandb/wandb",
1075
  "category": "ml visualization",
1076
  "github_about_section": "The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models from experimentation to production.",
1077
+ "homepage_link": "https://wandb.ai",
1078
+ "contributors_all": 235,
1079
+ "contributors_2025": 46,
1080
+ "contributors_2024": 67,
1081
+ "contributors_2023": 62
1082
  },
1083
  {
1084
  "repo_name": "aws-neuron-sdk",
1085
  "repo_link": "https://github.com/aws-neuron/aws-neuron-sdk",
1086
  "category": "sdk",
1087
  "github_about_section": "Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services",
1088
+ "homepage_link": "https://aws.amazon.com/ai/machine-learning/neuron",
1089
+ "contributors_all": 142,
1090
+ "contributors_2025": 33,
1091
+ "contributors_2024": 37,
1092
+ "contributors_2023": 32
1093
  },
1094
  {
1095
  "repo_name": "onnxruntime",
1096
  "repo_link": "https://github.com/microsoft/onnxruntime",
1097
  "category": "machine learning interoperability",
1098
  "github_about_section": "ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator",
1099
+ "homepage_link": "https://onnxruntime.ai",
1100
+ "contributors_all": 876,
1101
+ "contributors_2025": 237,
1102
+ "contributors_2024": 213,
1103
+ "contributors_2023": 213
1104
  },
1105
  {
1106
  "repo_name": "ort",
1107
  "repo_link": "https://github.com/pykeio/ort",
1108
  "category": "machine learning interoperability",
1109
  "github_about_section": "Fast ML inference & training for ONNX models in Rust",
1110
+ "homepage_link": "https://ort.pyke.io",
1111
+ "contributors_all": 70,
1112
+ "contributors_2025": 25,
1113
+ "contributors_2024": 20,
1114
+ "contributors_2023": 21
1115
  },
1116
  {
1117
  "repo_name": "Triton-distributed",
1118
  "repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
1119
  "category": "distributed computing",
1120
  "github_about_section": "Distributed Compiler based on Triton for Parallel Systems",
1121
+ "homepage_link": "https://triton-distributed.readthedocs.io",
1122
+ "contributors_all": 35,
1123
+ "contributors_2025": 30,
1124
+ "contributors_2024": 0,
1125
+ "contributors_2023": 0
1126
  },
1127
  {
1128
  "repo_name": "gemlite",
1129
  "repo_link": "https://github.com/dropbox/gemlite",
1130
  "category": "gpu kernels",
1131
+ "github_about_section": "Fast low-bit matmul kernels in Triton",
1132
+ "contributors_all": 5,
1133
+ "contributors_2025": 1,
1134
+ "contributors_2024": 5,
1135
+ "contributors_2023": 0
1136
  },
1137
  {
1138
  "repo_name": "cutile-python",
1139
  "repo_link": "https://github.com/NVIDIA/cutile-python",
1140
  "category": "parallel computing",
1141
  "github_about_section": "cuTile is a programming model for writing parallel kernels for NVIDIA GPUs",
1142
+ "homepage_link": "https://docs.nvidia.com/cuda/cutile-python",
1143
+ "contributors_all": 19,
1144
+ "contributors_2025": 10,
1145
+ "contributors_2024": 0,
1146
+ "contributors_2023": 0
1147
  },
1148
  {
1149
  "repo_name": "tilus",
1150
  "repo_link": "https://github.com/NVIDIA/tilus",
1151
  "category": "parallel computing",
1152
  "github_about_section": "Tilus is a tile-level kernel programming language with explicit control over shared memory and registers.",
1153
+ "homepage_link": "https://nvidia.github.io/tilus",
1154
+ "contributors_all": 6,
1155
+ "contributors_2025": 4,
1156
+ "contributors_2024": 0,
1157
+ "contributors_2023": 0
1158
  },
1159
  {
1160
  "repo_name": "triton-windows",
1161
  "repo_link": "https://github.com/woct0rdho/triton-windows",
1162
  "category": "parallel computing dsl",
1163
+ "github_about_section": "Fork of the Triton language and compiler for Windows support and easy installation",
1164
+ "contributors_all": 537,
1165
+ "contributors_2025": 233,
1166
+ "contributors_2024": 207,
1167
+ "contributors_2023": 159
1168
  },
1169
  {
1170
  "repo_name": "flash-linear-attention",
1171
  "repo_link": "https://github.com/fla-org/flash-linear-attention",
1172
  "category": "gpu kernels",
1173
+ "github_about_section": "Efficient implementations of state-of-the-art linear attention models",
1174
+ "contributors_all": 88,
1175
+ "contributors_2025": 64,
1176
+ "contributors_2024": 22,
1177
+ "contributors_2023": 3
1178
  }
1179
  ]