Papers
-
More, better or different? Trade-offs between group size and competence development in jury theoremsInstitute for Futures Studies, Umeå University
-
Gemma: Open Models Based on Gemini Research and Technology
-
ChipNeMo: Domain-Adapted LLMs for Chip Design
-
mPLUG-Owl : Modularization Empowers Large Language Models with Multimodality
-
GPT-4 Technical Report
-
The Claude 3 Model Family: Opus, Sonnet, and Haiku
-
DINOv2: Learning Robust Visual Features without Supervision
-
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training
-
Mixtral of Experts
-
VideoPoet: A Large Language Model for Zero-Shot Video Generation
-
Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision
-
Diffusion Models Without Attention
-
Towards Making the Most of ChatGPT for Machine Translation
-
VOYAGER: An Open-Ended Embodied Agent with Large Language Models
-
Eureka: Human-Level Reward Design via Coding Large Language Models
-
BitNet: Scaling 1-bit Transformers for Large Language Models
-
Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond
-
Solving Olympiad Geometry without Human Demonstrations (AlphaGeometry)
-
mPLUG-Owl2: Revolutionizing Multi-modal Large Language Model with Modality Collaboration
-
SteerLM: Attribute Conditioned SFT as an (User-Steerable) Alternative to RLHF
-
AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation
-
Qwen Technical Report
-
Qwen Technical Report
-
SWE-Agent: Agent-Computer Interfaces Enable Automated Software EngineeringPrinceton University
-
PaLM 2 Technical Report
-
Phi-2: A Small Language Model with Reasoning Capability
-
Efficient Memory Management for Large Language Model Serving with PagedAttentionStanford University, University of California, Berkeley
-
Textbooks Are All You Need
-
Graph of Thoughts: Solving Elaborate Problems with Large Language ModelsEidgenössische Technische Hochschule Zürich
-
3D Gaussian Splatting for Real-Time Radiance Field RenderingINRIA, Université Côte d’Azur
-
MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework
-
RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control
-
Differentially Private Heavy Hitter Detection using Federated Analytics
-
Llama 2: Open Foundation and Fine-Tuned Chat Models
-
Neuralangelo: High-Fidelity Neural Surface Reconstruction
-
Orca: Progressive Learning from Complex Explanation Traces of GPT-4Microsoft Research
-
AWQ: Activation-aware Weight Quantization for LLM Compression and AccelerationMassachusetts Institute of Technology
-
IMAGEBIND: One Embedding Space To Bind Them Al
-
ReWOO: Decoupling Reasoning from Observation for Efficient LLM Reasoning
-
Voyager: An Open-Ended Embodied Agent with Large Language Models
-
QLoRA: Efficient Finetuning of Quantized LLMsUniversity of Washington
-
Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture
-
Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models
-
Segment Anything
-
DAMO-YOLO : A Report on Real-Time Object Detection Design
-
VECO 2.0: Cross-lingual Language Model Pre-training with Multi-granularity Contrastive Learning
-
Sparks of Artificial General Intelligence: Early experiments with GPT-4
-
Generative Agents: Interactive Simulacra of Human Behavior
-
Regression Transformer enables concurrent sequence regression and generation for molecular language modelling
-
CAMEL: Communicative Agents for “Mind” Exploration of Large Language Models
