Search

Your search keyword '"Choi, Yejin"' showing total 886 results

Search Constraints

Start Over You searched for: Author "Choi, Yejin" Remove constraint Author: "Choi, Yejin"
886 results on '"Choi, Yejin"'

Search Results

1. BLIP3-KALE: Knowledge Augmented Large-Scale Dense Captions

2. Benchmarking Vision Language Model Unlearning via Fictitious Facial Identity Dataset

3. RESTOR: Knowledge Recovery through Machine Unlearning

4. Towards Visual Text Design Transfer Across Languages

5. SafetyAnalyst: Interpretable, transparent, and steerable LLM safety moderation

6. Diverging Preferences: When do Annotators Disagree and do Models Know?

7. SimpleToM: Exposing the Gap between Explicit ToM Inference and Implicit ToM Application in LLMs

8. Model Swarms: Collaborative Search to Adapt LLM Experts via Swarm Intelligence

9. ActionAtlas: A VideoQA Benchmark for Domain-specialized Action Recognition

10. Biased AI can Influence Political Decision-Making

11. Intuitions of Compromise: Utilitarianism vs. Contractualism

12. AI as Humanity's Salieri: Quantifying Linguistic Creativity of Language Models via Systematic Attribution of Machine Text against Web Text

13. Can Language Models Reason about Individualistic Human Values and Preferences?

14. DailyDilemmas: Revealing Value Preferences of LLMs with Quandaries of Daily Life

15. CulturalBench: a Robust, Diverse and Challenging Benchmark on Measuring the (Lack of) Cultural Knowledge of LLMs

16. HAICOSYSTEM: An Ecosystem for Sandboxing Safety Risks in Human-AI Interactions

17. WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild

18. StyleRemix: Interpretable Authorship Obfuscation via Distillation and Perturbation of Style Elements

19. Symbolic Working Memory Enhances Language Models for Complex Rule Application

20. xGen-MM (BLIP-3): A Family of Open Large Multimodal Models

21. Trust or Escalate: LLM Judges with Provable Guarantees for Human Agreement

22. WildHallucinations: Evaluating Long-form Factuality in LLMs with Real-World Entity Queries

23. Data Mixture Inference: What do BPE Tokenizers Reveal about their Training Data?

24. Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild

25. CopyBench: Measuring Literal and Non-Literal Reproduction of Copyright-Protected Text in Language Model Generation

26. Perceptions to Beliefs: Exploring Precursory Inferences for Theory of Mind in Large Language Models

27. The Art of Saying No: Contextual Noncompliance in Language Models

28. Language Model Alignment in Multilingual Trolley Problems

29. Certainly Uncertain: A Benchmark and Metric for Multimodal Epistemic and Aleatoric Awareness

30. How to Train Your Fact Verifier: Knowledge Transfer with Multimodal Open Models

31. WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models

32. WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs

33. Modular Pluralism: Pluralistic Alignment via Multi-LLM Collaboration

34. MINT-1T: Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens

35. WildVision: Evaluating Vision-Language Models in the Wild with Human Preferences

36. Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback

37. Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing

38. WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild

39. DREAM: A Challenge Data Set and Models for Dialogue-Based Reading Comprehension

40. From Explicit CoT to Implicit CoT: Learning to Internalize CoT Step by Step

41. WildChat: 1M ChatGPT Interaction Logs in the Wild

42. CULTURE-GEN: Revealing Global Cultural Perception in Language Models through Natural Language Prompting

43. Foundational Challenges in Assuring Alignment and Safety of Large Language Models

44. CulturalTeaming: AI-Assisted Interactive Red-Teaming for Challenging LLMs' (Lack of) Multicultural Knowledge

45. Particip-AI: A Democratic Surveying Framework for Anticipating Future AI Use Cases, Harms and Benefits

46. RewardBench: Evaluating Reward Models for Language Modeling

47. Information-Theoretic Distillation for Reference-less Summarization

48. Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs

50. Selective 'Selective Prediction': Reducing Unnecessary Abstention in Vision-Language Reasoning

Catalog

Books, media, physical & digital resources