Search

Your search keyword '"Dehghani, Mostafa"' showing total 306 results

Search Constraints

Start Over You searched for: Author "Dehghani, Mostafa" Remove constraint Author: "Dehghani, Mostafa" Search Limiters Peer Reviewed Remove constraint Search Limiters: Peer Reviewed
306 results on '"Dehghani, Mostafa"'

Search Results

1. Frozen Feature Augmentation for Few-Shot Image Classification

2. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context

3. Fractal Patterns May Illuminate the Success of Next-Token Prediction

4. Gemini: A Family of Highly Capable Multimodal Models

5. Low-Rank Adaptation for Multilingual Summarization: An Empirical Study

6. How (not) to ensemble LVLMs for VQA

7. The Impact of Group Membership Bias on the Quality and Fairness of Exposure in Ranking

8. Patch n' Pack: NaViT, a Vision Transformer for any Aspect Ratio and Resolution

9. PaLI-X: On Scaling up a Multilingual Vision and Language Model

10. PaLM 2 Technical Report

11. End-to-End Spatio-Temporal Action Localisation with Video Transformers

12. Scaling Vision Transformers to 22 Billion Parameters

13. Dual PatchNorm

15. Adaptive Computation with Elastic Input Sequence

16. DSI++: Updating Transformer Memory with New Documents

17. Sparse Upcycling: Training Mixture-of-Experts from Dense Checkpoints

18. Karyotype AI for Precision Oncology

19. Scaling Instruction-Finetuned Language Models

20. Transcending Scaling Laws with 0.1% Extra Compute

21. $\Lambda$-DARTS: Mitigating Performance Collapse by Harmonizing Operation Selection among Cells

22. Intersection of Parallels as an Early Stopping Criterion

23. Scaling Laws vs Model Architectures: How does Inductive Bias Influence Scaling?

24. Confident Adaptive Language Modeling

25. Beyond Transfer Learning: Co-finetuning for Action Localisation

26. Simple Open-Vocabulary Object Detection with Vision Transformers

27. UL2: Unifying Language Learning Paradigms

28. Retrieval-Enhanced Machine Learning

29. Transformer Memory as a Differentiable Search Index

30. Ethnic differences in the lifestyle behaviors and premature coronary artery disease: a multi-center study

31. VUT: Versatile UI Transformer for Multi-Modal Multi-Task User Interface Modeling

32. PolyViT: Co-training Vision Transformers on Images, Videos and Audio

33. Discrete Representations Strengthen Vision Transformer Robustness

35. The Efficiency Misnomer

36. SCENIC: A JAX Library for Computer Vision Research and Beyond

37. Exploring the Limits of Large Scale Pre-training

38. Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers

39. The Benchmark Lottery

40. TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?

41. Gradual Domain Adaptation in the Wild:When Intermediate Distributions are Absent

42. Parameter-efficient Multi-task Fine-tuning for Transformers via Shared Hypernetworks

43. Are Pre-trained Convolutions Better than Pre-trained Transformers?

44. ViViT: A Video Vision Transformer

45. OmniNet: Omnidirectional Representations from Transformers

46. Long Range Arena: A Benchmark for Efficient Transformers

47. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

48. Efficient Transformers: A Survey

49. IDF++: Analyzing and Improving Integer Discrete Flows for Lossless Compression

50. Transferring Inductive Biases through Knowledge Distillation

Catalog

Books, media, physical & digital resources