Papers
-
Scaling Laws for Neural Language Models
-
Dota 2 with Large Scale Deep Reinforcement Learning
-
PyTorch: An Imperative Style, High-Performance Deep Learning Library
-
Overton: A Data System for Monitoring and Improving Machine-Learned Products
-
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding
-
RoBERTa: A Robustly Optimized BERT Pretraining Approach
-
D2-Net: A Trainable CNN for Joint Description and Detection of Local Features
-
The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words, and Sentences From Natural Supervision
-
DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks
-
AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias
-
Improving Language Understanding by Generative Pre-Training (GPT-1)
-
Language Models are Unsupervised Multitask Learners
-
Mask R-CNN
-
Billion-scale similarity search with GPUs
-
Learning with Privacy at Scale
-
Proximal Policy Optimization Algorithms
-
Attention Is All You Need
-
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
-
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
-
Bag of Tricks for Efficient Text Classification
-
Concrete Problems in AI Safety
-
Deep Residual Learning for Image Recognition
-
tttLRM: Test-Time Training for Long Context and Autoregressive 3D Reconstruction
-
Performance of Large Language Models in Answering Critical Care Medicine Questions
