Search

Your search keyword '"Peng, Hongwu"' showing total 102 results

Search Constraints

Start Over You searched for: Author "Peng, Hongwu" Remove constraint Author: "Peng, Hongwu"
102 results on '"Peng, Hongwu"'

Search Results

1. RTop-K: Ultra-Fast Row-Wise Top-K Algorithm and GPU Implementation for Neural Networks

2. APEER: Automatic Prompt Engineering Enhances Large Language Model Reranking

3. SSNet: A Lightweight Multi-Party Computation Scheme for Practical Privacy-Preserving Machine Learning Service in the Cloud

4. Learning from Teaching Regularization: Generalizable Correlations Should be Easy to Imitate

5. Zero-Space Cost Fault Tolerance for Transformer-based Language Models on ReRAM

6. Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Heads

7. MaxK-GNN: Extremely Fast GPU Kernel Design for Accelerating Graph Neural Networks Training

8. Advanced Large Language Model (LLM)-Driven Verilog Development: Enhancing Power, Performance, and Area Optimization in Code Synthesis

9. Evaluating Emerging AI/ML Accelerators: IPU, RDU, and NVIDIA/AMD GPUs

10. LinGCN: Structural Linearized Graph Convolutional Network for Homomorphically Encrypted Inference

11. Accel-GCN: High-Performance GPU Accelerator Design for Graph Convolution Networks

12. AutoReP: Automatic ReLU Replacement for Fast Private Network Inference

13. PASNet: Polynomial Architecture Search Framework for Two-party Computation-based Secure Neural Network Deployment

14. RRNet: Towards ReLU-Reduced Neural Network for Two-party Computation Based Private Inference

15. Dynamic Sparse Training via Balancing the Exploration-Exploitation Trade-off

16. Aerial Manipulation Using a Novel Unmanned Aerial Vehicle Cyber-Physical System

17. PolyMPCNet: Towards ReLU-free Neural Architecture Search in Two-party Computation Based Private Inference

18. Towards Sparsification of Graph Neural Networks

19. A Length Adaptive Algorithm-Hardware Co-design of Transformer on FPGA Through Sparse Attention and Dynamic Pipelining

20. An Automatic and Efficient BERT Pruning for Edge AI Systems

21. Accelerating Framework of Transformer by Hardware Design and Model Compression Co-Optimization

22. Detecting Gender Bias in Transformer-based Models: A Case Study on BERT

23. Optimizing FPGA-based Accelerator Design for Large-Scale Molecular Similarity Search

24. Binary Complex Neural Network Acceleration on FPGA

25. Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI

44. HMC-T RAN

48. Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI

Catalog

Books, media, physical & digital resources