-
Attention Is All You Need
Paper • 1706.03762 • Published • 108 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 18 -
LLaMA: Open and Efficient Foundation Language Models
Paper • 2302.13971 • Published • 20 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 248
Collections
Discover the best community collections!
Collections including paper arxiv:1706.03762
-
High-Resolution Image Synthesis with Latent Diffusion Models
Paper • 2112.10752 • Published • 15 -
Adding Conditional Control to Text-to-Image Diffusion Models
Paper • 2302.05543 • Published • 58 -
Proximal Policy Optimization Algorithms
Paper • 1707.06347 • Published • 11 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 64
-
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 18 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 25 -
Attention Is All You Need
Paper • 1706.03762 • Published • 108 -
Lookahead Anchoring: Preserving Character Identity in Audio-Driven Human Animation
Paper • 2510.23581 • Published • 41
-
Neural Machine Translation by Jointly Learning to Align and Translate
Paper • 1409.0473 • Published • 7 -
Attention Is All You Need
Paper • 1706.03762 • Published • 108 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 25 -
Hierarchical Reasoning Model
Paper • 2506.21734 • Published • 46
-
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
Paper • 2003.08934 • Published • 2 -
Learning Transferable Visual Models From Natural Language Supervision
Paper • 2103.00020 • Published • 19 -
Emerging Properties in Self-Supervised Vision Transformers
Paper • 2104.14294 • Published • 5 -
Segment Anything
Paper • 2304.02643 • Published • 6
-
Attention Is All You Need
Paper • 1706.03762 • Published • 108 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 18 -
LLaMA: Open and Efficient Foundation Language Models
Paper • 2302.13971 • Published • 20 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 248
-
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 18 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 25 -
Attention Is All You Need
Paper • 1706.03762 • Published • 108 -
Lookahead Anchoring: Preserving Character Identity in Audio-Driven Human Animation
Paper • 2510.23581 • Published • 41
-
High-Resolution Image Synthesis with Latent Diffusion Models
Paper • 2112.10752 • Published • 15 -
Adding Conditional Control to Text-to-Image Diffusion Models
Paper • 2302.05543 • Published • 58 -
Proximal Policy Optimization Algorithms
Paper • 1707.06347 • Published • 11 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 64
-
Neural Machine Translation by Jointly Learning to Align and Translate
Paper • 1409.0473 • Published • 7 -
Attention Is All You Need
Paper • 1706.03762 • Published • 108 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 25 -
Hierarchical Reasoning Model
Paper • 2506.21734 • Published • 46
-
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
Paper • 2003.08934 • Published • 2 -
Learning Transferable Visual Models From Natural Language Supervision
Paper • 2103.00020 • Published • 19 -
Emerging Properties in Self-Supervised Vision Transformers
Paper • 2104.14294 • Published • 5 -
Segment Anything
Paper • 2304.02643 • Published • 6