Papers
-
Reflexion: Language Agents with Verbal Reinforcement Learning
-
Application-Agnostic Language Modeling for On-Device ASR
-
Language Is Not All You Need: Aligning Perception with Language Models
-
LLaMA: Open and Efficient Foundation Language Models
-
Adding Conditional Control to Text-to-Image Diffusion ModelsStanford University
-
Toolformer: Language Models Can Teach Themselves to Use Tools
-
Why we built an AI supercomputer in the cloud
-
mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image and Video
-
Neural Codec Language Models are Zero-Shot Text to Speech Synthesizers
-
Constitutional AI: Harmlessness from AI Feedback
-
Robust Speech Recognition via Large-Scale Weak Supervision
-
Stable Diffusion with Core ML on Apple Silicon
-
Fast Inference from Transformers via Speculative DecodingStanford University, University of California, Berkeley
-
Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks
-
GPTQ: Accurate Post-Training Quantization for Generative Pre-trained TransformersEidgenössische Technische Hochschule Zürich, Institute of Science and Technology Austria
-
wav2vec 2.0: A Framework for Self-Supervised Learning of Speech RepresentationsFacebook Ai Research
-
data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language
-
Synergizing Reasoning and Acting in Language Models
-
PaLM: Scaling Language Modeling with Pathways
-
Toy Models of Superposition
-
AudioLM: A Language Modeling Approach to Audio Generation
-
Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
-
Prompt Tuning for Generative Multimodal Pretrained Models
-
AlexaTM 20B: Few-Shot Learning Using a Large-Scale Multilingual Seq2seq Model
-
Discovering faster matrix multiplication algorithms with reinforcement learning
-
No Language Left Behind: Scaling Human-Centered Machine TranslationMarta R. Costa-jussà
-
Softmax Linear Units
-
Scaling Autoregressive Models for Content-Rich Text-to-Image Generation
-
OPT: Open Pre-trained Transformer Language Models
-
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-AwarenessMassachusetts Institute of Technology, Stanford University
-
mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections
-
Photorealistic Text-to-Image Diffusion Models with Deep Language UnderstandingGoogle Research
-
M6-Rec: Generative Pretrained Language Models are Open-Ended Recommender Systems
-
Least-to-Most Prompting Enables Complex Reasoning in Large Language Models
-
Hierarchical Text-Conditional Image Generation with CLIP Latents
-
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
-
Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
-
Training Compute-Optimal Large Language Models
-
Training a Tokenizer for Free with Private Federated Learning
-
Training Language Models to Follow Instructions with Human Feedback
-
FourCastNet: A Global Data-driven High-resolution Weather Model using Adaptive Fourier Neural Operators
-
Competition-Level Code Generation with AlphaCode
-
Instant Neural Graphics Primitives with a Multiresolution Hash Encoding
-
ML-Decoder: Scalable and Versatile Classification Head
-
A Mathematical Framework for Transformer Circuits
-
Training Verifiers to Solve Math Word Problems
-
On-device Panoptic Segmentation for Camera Using Transformers
-
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
-
AGENT: A Benchmark for Core Psychological Reasoning
-
Highly accurate protein structure predictionwith AlphaFold
