lambda7xx / awesome-AI-system

paper and its code for AI System
187 stars 12 forks source link

May merge these two lists #3

Closed wenhuizhang closed 4 months ago

wenhuizhang commented 1 year ago

Frameworks [VLDB '20] PyTorch Distributed: Experiences on Accelerating Data Parallel Training [NeurIPS '19] PyTorch: An Imperative Style, High-Performance Deep Learning Library [OSDI '18] Ray: A Distributed Framework for Emerging AI Applications [OSDI '16] TensorFlow: A System for Large-Scale Machine Learning

Parallelism & Distributed Systems [OSDI '22] Unity: Accelerating DNN Training Through Joint Optimization of Algebraic Transformations and Parallelization [EuroSys '22] Varuna: Scalable, Low-cost Training of Massive Deep Learning Models [SC '21'] Chimera: Efficiently Training Large-Scale Neural Networks with Bidirectional Pipelines [ICML '21] PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models [OSDI '20] A Unified Architecture for Accelerating Distributed DNN Training in Heterogeneous GPU/CPU Clusters [ATC '20] HetPipe: Enabling Large DNN Training on (Whimpy) Heterogeneous GPU Clusters through Integration of

Pipelined Model Parallelism and Data Parallelism [NeurIPS '19] GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism [SOSP '19] A Generic Communication Scheduler for Distributed DNN Training Acceleration [SOSP '19] PipeDream: Generalized Pipeline Parallelism for DNN Training [EuroSys '19] Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks [arXiv '18] Horovod: fast and easy distributed deep learning in TensorFlow [ATC '17] Poseidon: An Efficient Communication Architecture for Distributed Deep Learning on GPU Clusters [EuroSys '16] STRADS: A Distributed Framework for Scheduled Model Parallel Machine Learning [EuroSys '16] GeePS: Scalable Deep Learning on Distributed GPUs with a GPU-specialized Parameter Server [OSDI '14] Scaling Distributed Machine Learning with the Parameter Server [NIPS '12] Large Scale Distributed Deep Networks

GPU Cluster Management [OSDI '22] Looking Beyond GPUs for DNN Scheduling on Multi-Tenant Clusters [NSDI '22] MLaaS in the Wild: Workload Analysis and Scheduling in Large-Scale Heterogeneous GPU Clusters [OSDI '21] Pollux: Co-adaptive Cluster Scheduling for Goodput-Optimized Deep Learning [NSDI '21] Elastic Resource Sharing for Distributed Deep Learning [OSDI '20] Heterogeneity-Aware Cluster Scheduling Policies for Deep Learning Workloads [OSDI '20] AntMan: Dynamic Scaling on GPU Clusters for Deep Learning [NSDI '20] Themis: Fair and Efficient GPU Cluster Scheduling [EuroSys '20] Balancing Efficiency and Fairness in Heterogeneous GPU Clusters for Deep Learning [NSDI '19] Tiresias: A GPU Cluster Manager for Distributed Deep Learning [ATC '19] Analysis of Large-Scale Multi-Tenant GPU Clusters for DNN Training Workloads [OSDI '18] Gandiva: Introspective cluster scheduling for deep learning

Memory Management for Machine Learning [ASPLOS '23] DeepUM: Tensor Migration and Prefetching in Unified Memory [ATC '22] Memory Harvesting in Multi-GPU Systems with Hierarchical Unified Virtual Memory [MobiSys '22] Memory-efficient DNN Training on Mobile Devices [HPCA '22] Enabling Efficient Large-Scale Deep Learning Training with Cache Coherent Disaggregated Memory Systems [ASPLOS '20] Capuchin: Tensor-based GPU Memory Management for Deep Learning [ASPLOS '20] SwapAdvisor: Push Deep Learning Beyond the GPU Memory Limit via Smart Swapping [ISCA '19] Interplay between Hardware Prefetcher and Page Eviction Policy in CPU-GPU Unified Virtual Memory [ISCA '18] Gist: Efficient Data Encoding for Deep Neural Network Training [PPoPP '18] SuperNeurons: Dynamic GPU Memory Management for Training Deep Neural Networks [MICRO '16] vDNN: Virtualized Deep Neural Networks for Scalable, Memory-Efficient Neural Network Design

Scheduling & Resource Management [ASPLOS '23] ElasticFlow: An Elastic Serverless Training Platform for Distributed Deep Learning [arXiv '22] EasyScale: Accuracy-consistent Elastic Training for Deep Learning [MLSys '22] VirtualFlow: Decoupling Deep Learning Models from the Underlying Hardware [SIGCOMM '22] Multi-resource interleaving for deep learning training [EuroSys '22] Out-Of-Order BackProp: An Effective Scheduling Technique for Deep Learning [ATC '21] Zico: Efficient GPU Memory Sharing for Concurrent DNN Training [NeurIPS '20] Nimble: Lightweight and Parallel GPU Task Scheduling for Deep Learning [OSDI' 20] KungFu: Making Training in Distributed Machine Learning Adaptive [OSDI '20] PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications [MLSys '20] Salus: Fine-Grained GPU Sharing Primitives for Deep Learning Applications [SOSP '19] Generic Communication Scheduler for Distributed DNN Training Acceleration [EuroSys '18] Optimus: An Efficient Dynamic Resource Scheduler for Deep Learning Clusters [HPCA '18] Applied Machine Learning at Facebook: A Datacenter Infrastructure Perspective

Serving Systems (& inference acceleration) [EuroSys '23] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access [MICRO '22] DFX: A Low-latency Multi-FPGA Appliance for Accelerating Transformer-based Text Generation [ATC '22] Serving Heterogeneous Machine Learning Models on Multi-GPU Servers with Spatio-Temporal Sharing [OSDI '22] Orca: A Distributed Serving System for Transformer-Based Language Generation Tasks [OSDI '22] Achieving μs-scale Preemption for Concurrent GPU-accelerated DNN Inferences [ATC '21] INFaaS: Automated Model-less Inference Serving [OSDI '20] Serving DNNs like Clockwork: Performance Predictability from the Bottom Up [ISCA '20] MLPerf Inference Benchmark [SOSP '19] Nexus: A GPU Cluster Engine for Accelerating DNN-Based Video Analysis [ISCA '19] MnnFast: a fast and scalable system architecture for memory-augmented neural networks [EuroSys '19] μLayer: Low Latency On-Device Inference Using Cooperative Single-Layer Acceleration and Processor-

Friendly Quantization [EuroSys '19] GrandSLAm: Guaranteeing SLAs for Jobs in Microservices Execution Frameworks [OSDI '18] Pretzel: Opening the Black Box of Machine Learning Prediction Serving Systems [NSDI '17] Clipper: A Low-Latency Online Prediction Serving System

Deep Learning Compiler [PLDI '21] DeepCuts: A Deep Learning Optimization Framework for Versatile GPU Workloads [OSDI '18] TVM: An Automated End-to-End Optimizing Compiler for Deep Learning

Very Large Models [ASPLOS '23] Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression [arxiv '21] ZeRO-Infinity: Breaking the GPU Memory Wall for Extreme Scale Deep Learning [ATC '21] ZeRO-Offload: Democratizing Billion-Scale Model Training [FAST '21] Behemoth: A Flash-centric Training Accelerator for Extreme-scale DNNs

Deep Learning Recommendation Models [OSDI '22] FAERY: An FPGA-accelerated Embedding-based Retrieval System [OSDI '22] Ekko: A Large-Scale Deep Learning Recommender System with Low-Latency Model Update [EuroSys '22] Fleche: An Efficient GPU Embedding Cache for Personalized Recommendations [ASPLOS '22] RecShard: statistical feature-based memory optimization for industry-scale neural recommendation [HPCA '22] Hercules: Heterogeneity-Aware Inference Serving for At-Scale Personalized Recommendation [MLSys '21] TT-Rec: Tensor Train Compression for Deep Learning Recommendation Model Embeddings [HPCA '21] Tensor Casting: Co-Designing Algorithm-Architecture for Personalized Recommendation Training [HPCA '21] Understanding Training Efficiency of Deep Learning Recommendation Models at Scale [ISCA '20] DeepRecSys: A System for Optimizing End-To-End At-scale Neural Recommendation Inference [HPCA '20] The Architectural Implications of Facebook’s DNN-based Personalized Recommendation [MICRO '19] TensorDIMM: A Practical Near-Memory Processing Architecture for Embeddings and Tensor Operations in

Deep Learning Hardware Support for ML [ISCA '18] A Configurable Cloud-Scale DNN Processor for Real-Time AI [ISCA '17] In-Datacenter Performance Analysis of a Tensor Processing Unit

ML at Mobile & Embedded Systems [MobiCom '20] SPINN: Synergistic Progressive Inference of Neural Networks over Device and Cloud [RTSS '19] Pipelined Data-Parallel CPU/GPU Scheduling for Multi-DNN Real-Time Inference [ASPLOS '17] Neurosurgeon: Collaborative Intelligence Between the Cloud and Mobile Edge

ML Techniques for Systems [ICML '20] An Imitation Learning Approach for Cache Replacement [ICML '18] Learning Memory Access Patterns

lambda7xx commented 1 year ago

Some work of your list doesn't open soure. for example, ORCA(OSDI'23)

wenhuizhang commented 1 year ago

Some work of your list doesn't open soure. for example, ORCA(OSDI'23)

Okay, how about add some summary for each paper, so that people don't have to click into the repo to understand what the work is focusing on and what are the contributions.

lambda7xx commented 1 year ago

This repo is mainly to collect some paper code. If we add some summary for each paper, maybe we should open a new repo like this repo

lambda7xx commented 1 year ago

Your classification is better than this repo. I'll adjust this repo. Thanks

ArthurZucker commented 4 months ago

?