Search

Your search keyword '"Zhou, Daquan"' showing total 93 results

Search Constraints

Start Over You searched for: Author "Zhou, Daquan" Remove constraint Author: "Zhou, Daquan"
93 results on '"Zhou, Daquan"'

Search Results

1. StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation

2. PLLaVA : Parameter-free LLaVA Extension from Images to Videos for Video Dense Captioning

3. Chain of Thought Explanation for Dialogue State Tracking

4. Sora Generates Videos with Stunning Geometrical Consistency

5. Magic-Me: Identity-Specific Video Customized Diffusion

6. MagicVideo-V2: Multi-Stage High-Aesthetic Video Generation

7. A Dataset and Benchmark for Copyright Infringement Unlearning from Text-to-Image Diffusion Models

8. Factorization Vision Transformer: Modeling Long Range Dependency with Local Window Cost

9. MAgIC: Investigation of Large Language Model Powered Multi-Agent in Cognition, Adaptability, Rationality and Collaboration

10. ChatAnything: Facetime Chat with LLM-Enhanced Personas

11. EPIM: Efficient Processing-In-Memory Accelerators based on Epitome

12. Low-Resolution Self-Attention for Semantic Segmentation

13. MaskDiffusion: Boosting Text-to-Image Consistency with Conditional Mask

14. Dataset Quantization

15. BuboGPT: Enabling Visual Grounding in Multi-Modal LLMs

16. DiffFit: Unlocking Transferability of Large Diffusion Models via Simple Parameter-Efficient Fine-Tuning

17. InfoBatch: Lossless Training Speed Up by Unbiased Dynamic Data Pruning

18. DiM: Distilling Dataset into Generative Model

19. Diffusion Probabilistic Model Made Slim

20. Expanding Small-Scale Datasets with Guided Imagination

21. MagicVideo: Efficient Video Generation With Latent Diffusion Models

22. MagicMix: Semantic Mixing with Diffusion Models

23. Deep Model Reassembly

24. Scaling & Shifting Your Features: A New Baseline for Efficient Model Tuning

25. Sharpness-Aware Training for Free

26. Understanding The Robustness in Vision Transformers

27. M$^2$BEV: Multi-Camera Joint 3D Detection and Segmentation with Unified Birds-Eye View Representation

28. Shunted Self-Attention via Multi-Scale Token Aggregation

29. Refiner: Refining Self-attention for Vision Transformers

30. All Tokens Matter: Token Labeling for Training Better Vision Transformers

31. DeepViT: Towards Deeper Vision Transformer

32. AutoSpace: Neural Architecture Search with Less Human Interference

33. Coordinate Attention for Efficient Mobile Network Design

35. ConvBERT: Improving BERT with Span-based Dynamic Convolution

36. Progressive Tandem Learning for Pattern Recognition with Deep Spiking Neural Networks

37. PANet: Few-Shot Image Semantic Segmentation with Prototype Alignment

38. Neural Epitome Search for Architecture-Agnostic Network Compression

42. Rethinking Bottleneck Structure for Efficient Mobile Network Design

46. Token Selection is a Simple Booster for Vision Transformers

Catalog

Books, media, physical & digital resources