
youtube_summarizer
A modern Next.js-based tool for AI-powered YouTube video summarization. This application allows you to generate concise summaries of YouTube videos using different AI models, with support for multiple languages and summary styles.
Stars: 86

README:
A modern Next.js-based tool for AI-powered YouTube video summarization. This application allows you to generate concise summaries of YouTube videos using different AI models, with support for multiple languages and summary styles.
-
Multiple AI Models: Choose your preferred AI model for summarization:
- Google Gemini 2.0 Flash (Fast and efficient)
- Groq with Llama 70B (High accuracy)
- GPT-4o-mini (Balanced performance)
-
Flexible API Key Requirements:
- Only one API key is required to start using the application
- Models become available based on the API keys you provide
- Mix and match different models as needed
-
Multilingual Support:
- Generate summaries in English and German
- Clean formatting in both languages
- Proper handling of language-specific structures
-
Flexible Summary Modes:
- Video Summary: Concise, structured overview
- Podcast Style: More narrative, detailed analysis
-
Smart History System:
- Automatic storage in SQLite database
- Quick access to previous summaries
- Unique constraint handling for video/language combinations
-
Modern UI/UX:
- Clean, responsive design with Tailwind CSS
- Automatic dark/light mode
- Progress indicators for summarization
- Beautiful markdown rendering
- Mobile-friendly interface
The main interface where users can input a YouTube URL and select their preferred language, summary type, and AI model.
Real-time progress tracking shows the current status of your summary generation, including processing stages and completion percentage.
The generated summary is displayed in a clean, well-structured format with an overview and key points from the video.
Access your previously generated summaries through the history dashboard, showing video titles and generation dates.
View complete details of past summaries, including full analysis and key points.
- Node.js 15.x or higher (for local installation)
- npm package manager (for local installation)
- Docker (optional, for containerized installation)
- API keys for the AI services
- Clone the repository:
git clone [repository-url]
cd youtube-summarizer
- Install dependencies:
npm install
# or
yarn install
- Create a
.env
file in the root directory:
# You only need to add the API keys for the models you want to use
# At least one API key is required
GEMINI_API_KEY="your-gemini-api-key"
GROQ_API_KEY="your-groq-api-key"
OPENAI_API_KEY="your-openai-api-key"
- Set up the database:
npx prisma generate
npx prisma db push
- Start the development server:
npm run dev
# or
yarn dev
- Clone the repository:
git clone [repository-url]
cd youtube-summarizer
- Build the Docker image:
docker build -t youtube-summarizer .
- Run the container:
docker run -d \
-p 3000:3000 \
-v ./prisma:/app/prisma \
-e GEMINI_API_KEY="your-key" \
-e GROQ_API_KEY="your-key" \
-e OPENAI_API_KEY="your-key" \
youtube-summarizer
Note for Docker installation:
- The
-v ./prisma:/app/prisma
flag creates a volume for the SQLite database - You only need to provide the API keys for the models you want to use
- At least one API key is required
- The application will be available at http://localhost:3000
The application will be available at http://localhost:3000
The application is designed to work with partial API key configurations:
- You only need to provide API keys for the models you want to use
- The UI will automatically show which models are available based on your API keys
- You can start with just one API key and add more later
- Models without API keys will be disabled in the interface
The application uses Prisma with SQLite for data persistence. The configuration is defined in prisma/schema.prisma
:
generator client {
provider = "prisma-client-js"
}
datasource db {
provider = "sqlite"
url = "file:./dev.db"
}
To reset the database if you encounter any issues:
# Remove the existing database
rm prisma/dev.db
# Regenerate the database
npx prisma generate
npx prisma db push
-
Google Gemini API Key (Good starting choice - free tier available):
- Visit Google AI Studio
- Create a new project if needed
- Generate an API key
- Free tier available with generous limits
-
Groq API Key:
- Go to Groq Cloud
- Sign up for an account
- Navigate to API settings
- Generate a new API key
-
OpenAI API Key:
- Visit OpenAI Platform
- Create an account or log in
- Go to API settings
- Generate a new API key
- Note: This service requires a paid subscription
- Previously built with Python and Streamlit
- Completely rebuilt using Next.js for better performance
- New architecture using the App Router for improved routing
- Enhanced state management and real-time updates
- Streaming responses for real-time progress updates
- Efficient chunk processing for long videos
- Smart caching of summaries
- Optimized database queries
- Frontend: Next.js 15+, React, TypeScript
- Styling: Tailwind CSS, shadcn/ui components
- Database: Prisma with SQLite
- AI Integration: Multiple model support
- API: Built-in API routes with streaming support
- Visit the homepage
- Paste a YouTube URL
- Select your preferred:
- Language (English/German)
- Summary mode (Video/Podcast)
- AI model
- Click "Generate Summary"
- Watch the real-time progress
- View your formatted summary
- Access previous summaries in the history section
Contributions are welcome! Please feel free to submit a Pull Request.
This project is licensed under the MIT License - see the LICENSE file for details.
If you encounter database errors like "database disk image is malformed", follow these steps:
- Stop the development server
- Delete the corrupted database:
rm prisma/dev.db
- Regenerate the database:
npx prisma generate npx prisma db push
- Restart the development server:
npm run dev
If you encounter API errors:
- Check that all environment variables are properly set in
.env
- Verify that your API keys are valid and have sufficient credits
- For history-related errors, try resetting the database as described above
-
"Invalid API Key" errors:
- Double-check your API keys in
.env
- Make sure there are no extra spaces or quotes
- Verify the keys are active in their respective platforms
- Double-check your API keys in
-
"Failed to fetch summaries" error:
- Usually indicates a database issue
- Follow the database reset steps above
- Check if your database has proper read/write permissions
-
Performance issues:
- Long videos may take more time to process
- Consider using Gemini model for faster processing
- Check your network connection
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for youtube_summarizer
Similar Open Source Tools

gemini-multimodal-playground
Gemini Multimodal Playground is a basic Python app for voice conversations with Google's Gemini 2.0 AI model. It features real-time voice input and text-to-speech responses. Users can configure settings through the GUI and interact with Gemini by speaking into the microphone. The application provides options for voice selection, system prompt customization, and enabling Google search. Troubleshooting tips are available for handling audio feedback loop issues that may occur during interactions.

web-ui
WebUI is a user-friendly tool built on Gradio that enhances website accessibility for AI agents. It supports various Large Language Models (LLMs) and allows custom browser integration for seamless interaction. The tool eliminates the need for re-login and authentication challenges, offering high-definition screen recording capabilities.

comfyui-web-viewer
The ComfyUI Web Viewer by vrch.ai is a real-time AI-generated interactive art framework that integrates realtime streaming into ComfyUI workflows. It supports keyboard control nodes, OSC control nodes, sound input nodes, and more, accessible from any device with a web browser. It enables real-time interaction with AI-generated content, ideal for interactive visual projects and enhancing ComfyUI workflows with efficient content management and display.

miner-release
Heurist Miner is a tool that allows users to contribute their GPU for AI inference tasks on the Heurist network. It supports dual mining capabilities for image generation models and Large Language Models, offers flexible setup on Windows or Linux with multiple GPUs, ensures secure rewards through a dual-wallet system, and is fully open source. Users can earn rewards by hosting AI models and supporting applications in the Heurist ecosystem.

morphic
Morphic is an AI-powered answer engine with a generative UI. It utilizes a stack of Next.js, Vercel AI SDK, OpenAI, Tavily AI, shadcn/ui, Radix UI, and Tailwind CSS. To get started, fork and clone the repo, install dependencies, fill out secrets in the .env.local file, and run the app locally using 'bun dev'. You can also deploy your own live version of Morphic with Vercel. Verified models that can be specified to writers include Groq, LLaMA3 8b, and LLaMA3 70b.

trendFinder
Trend Finder is a tool designed to help users stay updated on trending topics on social media by collecting and analyzing posts from key influencers. It sends Slack notifications when new trends or product launches are detected, saving time, keeping users informed, and enabling quick responses to emerging opportunities. The tool features AI-powered trend analysis, social media and website monitoring, instant Slack notifications, and scheduled monitoring using cron jobs. Built with Node.js and Express.js, Trend Finder integrates with Together AI, Twitter/X API, Firecrawl, and Slack Webhooks for notifications.

recommendarr
Recommendarr is a tool that generates personalized TV show and movie recommendations based on your Sonarr, Radarr, Plex, and Jellyfin libraries using AI. It offers AI-powered recommendations, media server integration, flexible AI support, watch history analysis, customization options, and dark/light mode toggle. Users can connect their media libraries and watch history services, configure AI service settings, and get personalized recommendations based on genre, language, and mood/vibe preferences. The tool works with any OpenAI-compatible API and offers various recommended models for different cost options and performance levels. It provides personalized suggestions, detailed information, filter options, watch history analysis, and one-click adding of recommended content to Sonarr/Radarr.

chunkr
Chunkr is an open-source document intelligence API that provides a production-ready service for document layout analysis, OCR, and semantic chunking. It allows users to convert PDFs, PPTs, Word docs, and images into RAG/LLM-ready chunks. The API offers features such as layout analysis, OCR with bounding boxes, structured HTML and markdown output, and VLM processing controls. Users can interact with Chunkr through a Python SDK, enabling them to upload documents, process them, and export results in various formats. The tool also supports self-hosted deployment options using Docker Compose or Kubernetes, with configurations for different AI models like OpenAI, Google AI Studio, and OpenRouter. Chunkr is dual-licensed under the GNU Affero General Public License v3.0 (AGPL-3.0) and a commercial license, providing flexibility for different usage scenarios.

horde-worker-reGen
This repository provides the latest implementation for the AI Horde Worker, allowing users to utilize their graphics card(s) to generate, post-process, or analyze images for others. It offers a platform where users can create images and earn 'kudos' in return, granting priority for their own image generations. The repository includes important details for setup, recommendations for system configurations, instructions for installation on Windows and Linux, basic usage guidelines, and information on updating the AI Horde Worker. Users can also run the worker with multiple GPUs and receive notifications for updates through Discord. Additionally, the repository contains models that are licensed under the CreativeML OpenRAIL License.

extension-gen-ai
The Looker GenAI Extension provides code examples and resources for building a Looker Extension that integrates with Vertex AI Large Language Models (LLMs). Users can leverage the power of LLMs to enhance data exploration and analysis within Looker. The extension offers generative explore functionality to ask natural language questions about data and generative insights on dashboards to analyze data by asking questions. It leverages components like BQML Remote Models, BQML Remote UDF with Vertex AI, and Custom Fine Tune Model for different integration options. Deployment involves setting up infrastructure with Terraform and deploying the Looker Extension by creating a Looker project, copying extension files, configuring BigQuery connection, connecting to Git, and testing the extension. Users can save example prompts and configure user settings for the extension. Development of the Looker Extension environment includes installing dependencies, starting the development server, and building for production.

resume-job-matcher
Resume Job Matcher is a Python script that automates the process of matching resumes to a job description using AI. It leverages the Anthropic Claude API or OpenAI's GPT API to analyze resumes and provide a match score along with personalized email responses for candidates. The tool offers comprehensive resume processing, advanced AI-powered analysis, in-depth evaluation & scoring, comprehensive analytics & reporting, enhanced candidate profiling, and robust system management. Users can customize font presets, generate PDF versions of unified resumes, adjust logging level, change scoring model, modify AI provider, and adjust AI model. The final score for each resume is calculated based on AI-generated match score and resume quality score, ensuring content relevance and presentation quality are considered. Troubleshooting tips, best practices, contribution guidelines, and required Python packages are provided.

WatermarkRemover-AI
WatermarkRemover-AI is an advanced application that utilizes AI models for precise watermark detection and seamless removal. It leverages Florence-2 for watermark identification and LaMA for inpainting. The tool offers both a command-line interface (CLI) and a PyQt6-based graphical user interface (GUI), making it accessible to users of all levels. It supports dual modes for processing images, advanced watermark detection, seamless inpainting, customizable output settings, real-time progress tracking, dark mode support, and efficient GPU acceleration using CUDA.

probe
Probe is an AI-friendly, fully local, semantic code search tool designed to power the next generation of AI coding assistants. It combines the speed of ripgrep with the code-aware parsing of tree-sitter to deliver precise results with complete code blocks, making it perfect for large codebases and AI-driven development workflows. Probe is fully local, keeping code on the user's machine without relying on external APIs. It supports multiple languages, offers various search options, and can be used in CLI mode, MCP server mode, AI chat mode, and web interface. The tool is designed to be flexible, fast, and accurate, providing developers and AI models with full context and relevant code blocks for efficient code exploration and understanding.