Search

Your search keyword '"Phanishayee, Amar"' showing total 76 results

Search Constraints

Start Over You searched for: Author "Phanishayee, Amar" Remove constraint Author: "Phanishayee, Amar"
76 results on '"Phanishayee, Amar"'

Search Results

1. Integrated Hardware Architecture and Device Placement Search

2. Data-driven Forecasting of Deep Learning Performance on GPUs

3. Workload-Aware Hardware Accelerator Mining for Distributed Deep Learning Training

4. D\'ej\`aVu: KV-cache Streaming for Fast, Fault-tolerant Generative LLM Serving

5. Blox: A Modular Toolkit for Deep Learning Schedulers

6. A Study on the Intersection of GPU Utilization and CNN Inference

7. Harmony: Overcoming the Hurdles of GPU Memory Capacity to Train Massive DNN Models on Commodity Servers

8. Synergy: Resource Sensitive DNN Scheduling in Multi-Tenant Clusters

9. Efficient Large-Scale Language Model Training on GPU Clusters Using Megatron-LM

10. Heterogeneity-Aware Cluster Scheduling Policies for Deep Learning Workloads

11. Analyzing and Mitigating Data Stalls in DNN Training

12. Efficient Algorithms for Device Placement of DNN Graph Operators

13. Memory-Efficient Pipeline-Parallel DNN Training

14. Daydream: Accurately Estimating the Efficacy of Optimizations for DNN Training

15. Blink: Fast and Generic Collectives for Distributed ML

16. The Non-IID Data Quagmire of Decentralized Machine Learning

17. Themis: Fair and Efficient GPU Cluster Scheduling

18. Analysis of Large-Scale Multi-Tenant GPU Clusters for DNN Training Workloads

19. PipeDream: Fast and Efficient Pipeline Parallel DNN Training

20. Parameter Hub: a Rack-Scale Parameter Server for Distributed Deep Neural Network Training

21. TBD: Benchmarking and Analyzing Deep Neural Network Training

22. Parameter Box: High Performance Parameter Servers for Efficient Distributed Deep Neural Network Training

23. Compositional programming and testing of dynamic distributed systems

24. Harmony

25. Efficient large-scale language model training on GPU clusters using megatron-LM

26. Doing more with less

28. Parameter Hub

32. FAWN: A Fast Array of Wimpy Nodes.

33. ProjecToR

35. A (In)Cast of Thousands: Scaling Datacenter TCP to Kiloservers and Gigabits (CMU-PDL-09-101)

36. FAWNdamentally Power-Efficient Clusters

37. FAWN: A Fast Array of Wimpy Nodes (CMU-PDL-08-108)

38. Measurement and Analysis of TCP Throughput Collapse in Cluster-based Storage Systems (CMU-PDL-07-105)

39. Lab of things

40. HomeLab

41. Flex-KV

42. FAWN

44. FAWN

46. Ditto

50. Ditto.

Catalog

Books, media, physical & digital resources