Search

Your search keyword '"Zhang, Minjia"' showing total 253 results

Search Constraints

Start Over You searched for: Author "Zhang, Minjia" Remove constraint Author: "Zhang, Minjia"
253 results on '"Zhang, Minjia"'

Search Results

1. Improving Retrieval-Augmented Generation in Medicine with Iterative Follow-up Questions

2. Model Tells You Where to Merge: Adaptive KV Cache Merging for LLMs on Long-Context Tasks

3. UltraEdit: Instruction-based Fine-Grained Image Editing at Scale

4. Universal Checkpointing: Efficient and Flexible Checkpointing for Large Scale Distributed Training

5. Parcae: Proactive, Liveput-Optimized DNN Training on Preemptible Instances

6. Computing in the Era of Large Generative Models: From Cloud-Native to AI-Native

7. OpenFold: retraining AlphaFold2 yields new insights into its learning mechanisms and capacity for generalization

8. Configuration Validation with Large Language Models

9. DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies

10. Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs

11. DeepSpeed Ulysses: System Optimizations for Enabling Training of Extreme Long Sequence Transformer Models

12. DeepSpeed-VisualChat: Multi-Round Multi-Image Interleave Chat via Multi-Modal Causal Attention

13. RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model

14. Cost-effective On-device Continual Learning over Memory Hierarchy with Miro

15. DeepSpeed-Chat: Easy, Fast and Affordable RLHF Training of ChatGPT-like Models at All Scales

16. FedHC: A Scalable Federated Learning Framework for Heterogeneous and Resource-Constrained Clients

17. DeepSpeed Data Efficiency: Improving Deep Learning Model Quality and Training Efficiency via Efficient Data Sampling and Routing

18. Random-LTD: Random and Layerwise Token Dropping Brings Efficient Training for Large-scale Transformers

19. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model

20. Compressing Pre-trained Transformers via Low-Bit NxM Sparsity for Natural Language Understanding

21. DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale

22. ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers

23. Extreme Compression for Pre-trained Transformers Made Simple and Efficient

24. Bamboo: Making Preemptible Instances Resilient for Affordable Training of Large DNNs

25. A Survey of Multi-Tenant Deep Learning Inference on GPU

26. Maximizing Communication Efficiency for Large-scale Training via 0/1 Adam

27. Speed-ANN: Low-Latency and High-Accuracy Nearest Neighbor Search via Intra-Query Parallelism

28. ScaLA: Accelerating Adaptation of Pre-Trained Transformer-Based Language Models via Efficient Large-Batch Adversarial Noise

29. DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI Scale

30. A Survey of Large-Scale Deep Learning Serving System Optimization: Challenges and Opportunities

31. NxMTransformer: Semi-Structured Sparsification for Natural Language Understanding via ADMM

32. Carousel Memory: Rethinking the Design of Episodic Memory for Continual Learning

33. The Stability-Efficiency Dilemma: Investigating Sequence Length Warmup for Training GPT Models

34. Understanding and Generalizing Monotonic Proximity Graphs for Approximate Nearest Neighbor Search

36. ZeRO-Offload: Democratizing Billion-Scale Model Training

37. Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping

38. SHARP: An Adaptable, Energy-Efficient Accelerator for Recurrent Neural Network

39. Sentinel: Runtime Data Management on Heterogeneous Main MemorySystems for Deep Learning

40. Zoom: SSD-based Vector Search for Optimizing Accuracy, Latency and Memory

41. Navigating with Graph Representations for Fast and Scalable Decoding of Neural Language Models

43. Vertical Scaling of Resource for OpenMP Application

44. Learning Intrinsic Sparse Structures within Long Short-Term Memory

Catalog

Books, media, physical & digital resources