MLSys bundle from NeurIPS 25 #41
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Note
Introduces a new
neurips25-mlsyssection aggregating curated NeurIPS 2025 papers and concise summaries across Architecture, Compression, Inference, Multi‑Modality, RL, and Training.neurips25-mlsys/with curated NeurIPS 2025 ML systems content.README.md: Overview, stats graphic, table of contents, category summaries, related links.architecture.md: Efficient attention/KV-cache/speculative decoding systems; sparse attention (e.g., Gated Attention), diffusion architectures; theory and benchmarks.compression.md: Quantized attention, KV-cache compression; FP4 training, FP8 fine-tuning; sparsification and compression theory.inference.md: Serving systems (scheduling, distributed, KV cache), energy benchmarking, multi‑LoRA, TPU support, reliability; speculative decoding and long‑context methods; KV cache algorithms.multi-modality.md: Multimodal serving, video systems; token pruning/merging; efficient diffusion architectures/training; multimodal adaptation; diffusion theory.rl.md: RL training infrastructure, communication‑efficient training; efficient rollout/sampling; scalable policy optimization; scaling and analysis.training.md: Distributed/communication‑efficient training; memory and long‑context training; compiler/hardware optimization; energy; stability and architectural tweaks; optimizer/data/WD scaling laws and numerical stability.Written by Cursor Bugbot for commit bd7ec06. This will update automatically on new commits. Configure here.