Search

Your search keyword '"Wang, Zhangyang"' showing total 1,319 results

Search Constraints

Start Over You searched for: Author "Wang, Zhangyang" Remove constraint Author: "Wang, Zhangyang"
1,319 results on '"Wang, Zhangyang"'

Search Results

1. LLM-AutoDiff: Auto-Differentiate Any LLM Workflow

2. SPAM: Spike-Aware Adam with Momentum Reset for Stable LLM Training

3. VideoLifter: Lifting Videos to 3D with Fast Hierarchical Stereo Alignment

4. Rethinking Addressing in Language Models via Contexualized Equivariant Positional Encoding

5. Understanding and Mitigating Bottlenecks of State Space Models through the Lens of Recency and Over-smoothing

6. Enhancing Item Tokenization for Generative Recommendation through Self-Improvement

7. AutoTrust: Benchmarking Trustworthiness in Large Vision Language Models for Autonomous Driving

8. On How Iterative Magnitude Pruning Discovers Local Receptive Fields in Fully Connected Neural Networks

9. APOLLO: SGD-like Memory, AdamW-level Performance

10. A Stitch in Time Saves Nine: Small VLM is a Precise Guidance for Accelerating Large VLMs

11. Oscillation Inversion: Understand the structure of Large Flow Model through the Lens of Inversion Method

12. Know Where You're Uncertain When Planning with Multimodal Foundation Models: A Formal Framework

13. Chasing Better Deep Image Priors between Over- and Under-parameterization

14. Read-ME: Refactorizing LLMs as Router-Decoupled Mixture of Experts with System Co-Design

15. Large Spatial Model: End-to-end Unposed Images to Semantic 3D

16. Cavia: Camera-controllable Multi-view Video Diffusion with View-Integrated Attention

17. AlphaPruning: Using Heavy-Tailed Self Regularization Theory for Improved Layer-wise Pruning of Large Language Models

18. Training Dynamics of Transformers to Recognize Word Co-occurrence via Gradient Flow Analysis

19. Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild

20. On The Planning Abilities of OpenAI's o1 Models: Feasibility, Optimality, and Generalizability

21. LLM-PBE: Assessing Data Privacy in Large Language Models

22. All Against Some: Efficient Integration of Large Language Models for Message Passing in Graph Neural Networks

23. From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients

24. Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients

25. Expressive Gaussian Human Avatars from Monocular RGB Video

26. 4K4DGen: Panoramic 4D Generation at 4K Resolution

27. Learning Traffic Crashes as Language: Datasets, Benchmarks, and What-if Causal Analyses

28. Flextron: Many-in-One Flexible Large Language Model

29. LoCoCo: Dropping In Convolutions for Long Context Compression

30. Zero-Painter: Training-Free Layout Control for Text-to-Image Synthesis

31. CamCo: Camera-Controllable 3D-Consistent Image-to-Video Generation

32. VersatileGaussian: Real-Time Neural Rendering for Versatile Tasks Using Gaussian Splatting

33. FSGS: Real-Time Few-Shot View Synthesis Using Gaussian Splatting

34. DreamScene360: Unconstrained Text-to-3D Scene Generation with Panoramic Gaussian Splatting

35. Diffusion4D: Fast Spatial-temporal Consistent 4D Generation via Video Diffusion Models

36. Harnessing the power of longitudinal medical imaging for eye disease prognosis using Transformer-based sequence modeling

37. OpenBias: Open-set Bias Detection in Text-to-Image Generative Models

38. DreamScene360: Unconstrained Text-to-3D Scene Generation with Panoramic Gaussian Splatting

39. MM3DGS SLAM: Multi-modal 3D Gaussian Splatting for SLAM Using Vision, Depth, and Inertial Measurements

40. InstantSplat: Sparse-view SfM-free Gaussian Splatting in Seconds

41. Lift3D: Zero-Shot Lifting of Any 2D Vision Model to 3D

42. Generalization Error Analysis for Sparse Mixture-of-Experts: A Preliminary Study

43. Comp4D: LLM-Guided Compositional 4D Scene Generation

44. StreamingT2V: Consistent, Dynamic, and Extendable Long Video Generation from Text

45. Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression

46. Shake to Leak: Fine-tuning Diffusion Models Can Amplify the Generative Privacy Risk

47. GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection

48. Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding

49. Principled Architecture-aware Scaling of Hyperparameters

50. Take the Bull by the Horns: Hard Sample-Reweighted Continual Training Improves LLM Generalization

Catalog

Books, media, physical & digital resources