TAAFT
Free mode
100% free
Freemium
Free Trial
Deals
Create tool

LFM2-Audio-1.5

New Multimodal Gen 3
Released: October 1, 2025

Overview

LFM2-Audio-1.5 is a unified audio foundation model for real-time speech, music, and sound effects. It transcribes, speaks, and generates or edits audio from natural prompts, supports cross-lingual use, returns stable timestamps, and streams low-latency output for production apps.

Description

LFM2-Audio-1.5 brings ASR, TTS, and audio generation into a single model so teams can move from text or reference audio to finished sound without stitching tools together. It listens and speaks in the same session, preserving context across turns, and can follow detailed direction about voice, style, tempo, and atmosphere while keeping timing predictable for live agents or interactive media. On understanding tasks it produces clean transcripts with punctuation, speaker attribution when needed, and reliable timecodes that line up in an editor. On generation it renders expressive speech, creates music or sound effects that match the brief, and performs localized edits—re-voicing lines, extending ambience, inpainting gaps, or removing distractions—without breaking continuity or loudness. The model adapts to multiple languages and accents, accepts style references to maintain brand or character identity, and can perform consented voice matching under policy controls. Designed for production, it streams partial results, emits structured JSON alongside audio for automation, and fits into existing pipelines with quantization options for edge or high-throughput serving.

About Liquid AI

Liquid AI is an MIT spin-off building efficient general-purpose AI models (Liquid Foundation Models, or LFMs) that run on edge devices with less memory and power. They recently raised $250M in Series A funding to scale model development and deployment.

Website: liquid.ai
View Company Profile

Related Models

Last updated: October 8, 2025