Discover and explore top open-source AI tools and projects—updated daily.
code-with-antonioOpen-source AI platform for advanced text-to-speech and voice cloning
Top 93.4% on SourcePulse
Resonance offers an open-source, self-hostable alternative for AI-powered text-to-speech (TTS) and voice cloning, directly competing with commercial solutions like ElevenLabs. It targets developers and power users seeking a customizable, feature-rich platform for generating speech and cloning voices, providing significant control and potential cost savings over proprietary services.
How It Works
This project is built using Next.js 16 and React 19, integrating Chatterbox TTS for its core speech generation and zero-shot voice cloning capabilities. Voice cloning requires only a 10-second audio sample, eliminating the need for fine-tuning. Inference is handled via serverless GPUs on Modal, specifically NVIDIA A10G instances, ensuring scalability. Authentication and multi-tenancy are managed by Clerk Organizations, while usage-based billing, character metering, and voice creation pricing are implemented through Polar. Audio assets and voice reference files are stored in Cloudflare R2 buckets.
Quick Start & Requirements
git clone), navigate into the directory (cd resonance), and install dependencies (npm install)..env), setting up meters and products in Polar for billing, deploying the Chatterbox TTS engine to Modal, migrating the database (npx prisma migrate deploy), and seeding built-in voices (npx prisma db seed).Highlighted Details
Maintenance & Community
The project is maintained by code-with-antonio. No specific community channels (like Discord or Slack) or details on notable contributors or sponsorships are provided in the README.
Licensing & Compatibility
The license for this repository is not specified in the provided README text. This lack of explicit licensing information presents a significant adoption blocker, particularly for commercial use or integration into closed-source projects.
Limitations & Caveats
The serverless GPU inference on Modal may incur cold start latency for the first request after a period of inactivity. The setup process is complex, requiring integration and configuration of multiple external services (Clerk, Polar, Modal, R2, PostgreSQL). Crucially, the absence of a specified open-source license requires clarification before any adoption decision can be made.
1 month ago
Inactive