Papers
-
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
-
AGENT: A Benchmark for Core Psychological Reasoning
-
Highly accurate protein structure predictionwith AlphaFold
-
LoRA: Low-Rank Adaptation of Large Language Models
-
AN IMAGE IS WORTH 16X16 WORDS: TRANSFORMERS FOR IMAGE RECOGNITION AT SCALE
-
M6: A Chinese Multimodal Pretrainer
-
Holographic dynamics simulations with a trapped ion quantum computer
-
An autonomous debating system (Project Debater)
-
Learning Transferable Visual Models From Natural Language Supervision
-
Learning Transferable Visual Models From Natural Language Supervision
-
Federated Evaluation and Tuning for On-Device Personalization: System Design & Applications
-
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
-
Scalable Differential Privacy with Certified Robustness in Adversarial Learning
-
PinnerSage: Multi-Modal User Embedding Framework for Recommendations at Pinterest
-
Denoising Diffusion Probabilistic Models
-
ZeRO: Memory Optimizations Toward Training Trillion Parameter Models
-
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
-
Scaling Laws for Neural Language Models
-
Dota 2 with Large Scale Deep Reinforcement Learning
-
PyTorch: An Imperative Style, High-Performance Deep Learning Library
-
Overton: A Data System for Monitoring and Improving Machine-Learned Products
-
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding
-
RoBERTa: A Robustly Optimized BERT Pretraining Approach
-
D2-Net: A Trainable CNN for Joint Description and Detection of Local Features
-
The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words, and Sentences From Natural Supervision
-
DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks
-
AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias
-
Improving Language Understanding by Generative Pre-Training (GPT-1)
-
Evaluating Discourse Phenomena in Neural Machine Translation
-
Deep Sets
-
Language Models are Unsupervised Multitask Learners
-
Mask R-CNN
-
Billion-scale similarity search with GPUs
-
High Speed All-optical extended DV-Curve-based DNA sequence alignment utilizing wavelength and polarization modulation
-
Learning with Privacy at Scale
-
Proximal Policy Optimization Algorithms
-
Learning from Simulated and Unsupervised Images through Adversarial Training
-
Attention Is All You Need
-
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
-
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
-
Bag of Tricks for Efficient Text Classification
-
Concrete Problems in AI Safety
-
Deep Residual Learning for Image Recognition
-
tttLRM: Test-Time Training for Long Context and Autoregressive 3D Reconstruction
-
Reinforced Curriculum Pre-Alignment for Domain-Adaptive VLMs
-
Performance of Large Language Models in Answering Critical Care Medicine Questions
