Papers
-
data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language
-
Synergizing Reasoning and Acting in Language Models
-
PaLM: Scaling Language Modeling with Pathways
-
Monolith: Real Time Recommendation System With Collisionless Embedding Table
-
Rethinking Personalized Ranking at Pinterest: An End-to-End Approach
-
Toy Models of Superposition
-
AudioLM: A Language Modeling Approach to Audio Generation
-
Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
-
Sampling Is All You Need on Modeling Long-Term User Behaviors for CTR Prediction
-
Prompt Tuning for Generative Multimodal Pretrained Models
-
AlexaTM 20B: Few-Shot Learning Using a Large-Scale Multilingual Seq2seq Model
-
Discovering faster matrix multiplication algorithms with reinforcement learning
-
CLIP-CLOP: CLIP-Guided Collage and Photomontage
-
No Language Left Behind: Scaling Human-Centered Machine TranslationMarta R. Costa-jussà
-
Softmax Linear Units
-
Scaling Autoregressive Models for Content-Rich Text-to-Image Generation
-
OPT: Open Pre-trained Transformer Language Models
-
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-AwarenessMassachusetts Institute of Technology, Stanford University
-
mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections
-
ItemSage: Learning Product Embeddings for Shopping Recommendations at Pinterest
-
Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding
-
MultiBiSage: A Web-Scale Recommendation System Using Multiple Bipartite Graphs at Pinterest
-
M6-Rec: Generative Pretrained Language Models are Open-Ended Recommender Systems
-
Least-to-Most Prompting Enables Complex Reasoning in Large Language Models
-
PinnerFormer: Sequence Modeling for User Representation at Pinterest
-
Towards Total Recall in Industrial Anomaly Detection
-
Hierarchical Text-Conditional Image Generation with CLIP Latents
-
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
-
Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
-
Training Compute-Optimal Large Language Models
-
CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis
-
Learning When to Translate for Streaming Speech
-
Training a Tokenizer for Free with Private Federated Learning
-
Training Language Models to Follow Instructions with Human Feedback
-
FourCastNet: A Global Data-driven High-resolution Weather Model using Adaptive Fourier Neural Operators
-
Competition-Level Code Generation with AlphaCode
-
Low-Overhead Fault-Tolerant Quantum Error Correction with the Surface-GKP Code
-
Instant Neural Graphics Primitives with a Multiresolution Hash Encoding
-
ML-Decoder: Scalable and Versatile Classification Head
-
A Mathematical Framework for Transformer Circuits
-
Training Verifiers to Solve Math Word Problems
-
Improving language models by retrieving from trillions of tokens
-
On-device Panoptic Segmentation for Camera Using Transformers
-
Merlion: A Machine Learning Library for Time Series
-
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
-
AGENT: A Benchmark for Core Psychological Reasoning
-
Highly accurate protein structure predictionwith AlphaFold
-
LoRA: Low-Rank Adaptation of Large Language Models
-
AN IMAGE IS WORTH 16X16 WORDS: TRANSFORMERS FOR IMAGE RECOGNITION AT SCALE
-
M6: A Chinese Multimodal Pretrainer
