My TIGER app is now fully working again, with fixes and full compatibility with Gradio 6 🚀
It lets you: - 🎙️ Separate multiple speakers from an audio file - 🎬 Extract each speaker directly from a video - 🎧 Split audio into dialog, music, and sound effects (DnR) - 🎥 Apply DnR separation directly on videos
All powered by lightweight TIGER models for fast and efficient speech separation.
I’ve fixed the Space and brought it back to life: - ✅ Working again after being broken for a while - ✅ Updated to Gradio 6 - ✅ Compatible with ZeroGPU - ✅ Output videos now preserve original resolution and FPS
I also added advanced controls so you can experiment more (tracking, seed, motion, sketch).
I’ve been working on a new mathematical approach to real-time video compositing and background removal, and I wanted to share a live demo.
Traditionally, real-time keyers either use 3D color-space bounding boxes (which struggle with semi-transparent hair and motion blur) or heavy Machine Learning models (which require massive GPU compute and often suffer from temporal "jitter" on the edges).
I wanted to see if I could solve this using purely deterministic math so it could run client-side in a standard browser.
The engine uses a custom mathematical framework I call CMT SRL SEFA. Instead of looking at raw color values or guessing semantics like an AI, it treats the video feed as complex-encoded sequences. It uses harmonic frequencies to map phase geometry and applies a "Stability Cost Function" to find the global minimum stability. In short: it isolates the foreground from the background by measuring signal complexity and structural contradictions.
Give it a try using your own messy plates and such. As I am not a VFX artist, I am curious to hear thoughts and what should be improved upon and made better
We should really have a release date range slider on the /models page. Tired of "trending/most downloaded" being the best way to sort and still seeing models from 2023 on the first page just because they're embedded in enterprise pipelines and get downloaded repeatedly. "Recently Created/Recently Updated" don't solve the discovery problem considering the amount of noise to sift through.
Slight caveat: Trending actually does have some recency bias, but it's not strong/precise enough.
I improved the public demo for TADA — a generative framework for speech modeling via text–acoustic dual alignment.
TADA models speech as a joint sequence of text tokens and acoustic tokens, using a transformer backbone to keep text and audio synchronized during generation.
The original demo already exposed these mechanisms, but the workflow made the pipeline hard to understand.
This updated demo makes the process clearer:
• load the model • prepare a reference voice (optionally with transcript or Whisper auto-transcription) • generate speech conditioned on that reference
It also adds multilingual support.
Presets are included for a few languages, but the model supports more:
if you like it give the demo a little star and send a shoutout to : @MaxLSB@jddqd and @GAD-cell for absolutely obliterating the pareto frontier of the french language understanding .