Papers
-
Towards Making the Most of ChatGPT for Machine Translation
-
BitNet: Scaling 1-bit Transformers for Large Language Models
-
AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation
-
Phi-2: A Small Language Model with Reasoning Capability
-
Textbooks Are All You Need
-
Orca: Progressive Learning from Complex Explanation Traces of GPT-4
-
ReWOO: Decoupling Reasoning from Observation for Efficient LLM Reasoning
-
Sparks of Artificial General Intelligence: Early experiments with GPT-4
-
Language Is Not All You Need: Aligning Perception with Language Models
-
Neural Codec Language Models are Zero-Shot Text to Speech Synthesizers
-
Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks
-
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
-
LoRA: Low-Rank Adaptation of Large Language Models
-
ZeRO: Memory Optimizations Toward Training Trillion Parameter Models
-
D2-Net: A Trainable CNN for Joint Description and Detection of Local Features
-
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
-
Deep Residual Learning for Image Recognition
