
Pixelle-MCP
An Open-Source Multimodal AIGC Solution based on ComfyUI + MCP + LLM https://pixelle.ai
Stars: 382

Pixelle-MCP is a multi-channel publishing tool designed to streamline the process of publishing content across various social media platforms. It allows users to create, schedule, and publish posts simultaneously on platforms such as Facebook, Twitter, and Instagram. With a user-friendly interface and advanced scheduling features, Pixelle-MCP helps users save time and effort in managing their social media presence. The tool also provides analytics and insights to track the performance of posts and optimize content strategy. Whether you are a social media manager, content creator, or digital marketer, Pixelle-MCP is a valuable tool to enhance your online presence and engage with your audience effectively.
README:
English | δΈζ
β¨ An AIGC solution based on the MCP protocol, seamlessly converting ComfyUI workflows into MCP tools with zero code, empowering LLM and ComfyUI integration.
https://github.com/user-attachments/assets/65422cef-96f9-44fe-a82b-6a124674c417
- β 2025-08-12: Integrated the LiteLLM framework, adding multi-model support for Gemini, DeepSeek, Claude, Qwen, and more
- β π Full-modal Support: Supports TISV (Text, Image, Sound/Speech, Video) full-modal conversion and generation
- β π§© ComfyUI Ecosystem: Server-side is built on ComfyUI, inheriting all capabilities from the open ComfyUI ecosystem
- β π§ Zero-code Development: Defines and implements the Workflow-as-MCP Tool solution, enabling zero-code development and dynamic addition of new MCP Tools
- β ποΈ MCP Server: Server provides functionality based on the MCP protocol, supporting integration with any MCP client (including but not limited to Cursor, Claude Desktop, etc.)
- β π MCP Client: Client is developed based on the Chainlit framework, inheriting Chainlit's UI controls and supporting integration with more MCP Servers
- β π Flexible Deployment: Supports standalone deployment of Server-side only as MCP Server, or standalone deployment of Client-side only as MCP Client, or combined deployment
- β βοΈ Unified Configuration: Uses YAML configuration scheme, one config file manages all services
- β π€ Multi-LLM Support: Supports multiple mainstream LLMs, including OpenAI, Ollama, Gemini, DeepSeek, Claude, Qwen, and more
- mcp-base: π§ Basic service, provides file storage and shared service capabilities
- mcp-client: π MCP client, a web interface built on Chainlit
- mcp-server: ποΈ MCP server, provides various AIGC tools and services
git clone https://github.com/AIDC-AI/Pixelle-MCP.git
cd Pixelle-MCP
The project uses a unified YAML configuration scheme:
# Copy the configuration example file
cp config.yml.example config.yml
# Edit configuration items as needed
π Detailed Configuration Instructions:
The configuration file contains three main sections: Basic Service, MCP Server, and MCP Client. Each section has detailed configuration item descriptions in config.yml.example
.
π Configuration Checklist:
- β
Copied
config.yml.example
toconfig.yml
- β Configured ComfyUI service address (ensure ComfyUI is running)
- β Configured at least one LLM model (OpenAI or Ollama)
- β Port numbers are not occupied by other services (9001, 9002, 9003)
This step is optional and only affects your Agent's capabilities. You can skip it if not needed for now.
The mcp-server/workflows
directory contains a set of popular workflows by default. Run the following command to copy them to your mcp-server. When the service starts, they will be automatically converted into MCP Tools for LLM use.
Note: It is strongly recommended to test the workflow in your ComfyUI canvas before copying, to ensure smooth execution later.
cp -r mcp-server/workflows/* mcp-server/data/custom_workflows/
# Start all services
docker compose up -d
Requires uv environment.
Linux/macOS users:
# Start all services (foreground)
./run.sh
# Or
# Start all services (background)
./run.sh start --daemon
Windows users:
Simply double-click the run.bat
script in the root directory
Requires uv environment.
Start Basic Service (mcp-base):
cd mcp-base
# Install dependencies (only needed on first run or after updates)
uv sync
# Start service
uv run main.py
Start Server (mcp-server):
cd mcp-server
# Install dependencies (only needed on first run or after updates)
uv sync
# Start service
uv run main.py
Start Client (mcp-client):
cd mcp-client
# Install dependencies (only needed on first run or after updates)
uv sync
# Start service (for hot-reload in dev mode: uv run chainlit run main.py -w --port 9003)
uv run main.py
After startup, the service addresses are as follows:
-
Client: π http://localhost:9003 (Chainlit Web UI, default username and password are both
dev
, can be changed inauth.py
) - Server: ποΈ http://localhost:9002/sse (MCP Server)
- Base Service: π§ http://localhost:9001/docs (File storage and basic API)
β‘ One workflow = One MCP Tool
-
π Build a workflow in ComfyUI for image Gaussian blur (Get it here), then set the
LoadImage
node's title to$image.image!
as shown below: -
π€ Export it as an API format file and rename it to
i_blur.json
. You can export it yourself or use our pre-exported version (Get it here) -
π Copy the exported API workflow file (must be API format), input it on the web page, and let the LLM add this Tool
-
β¨ After sending, the LLM will automatically convert this workflow into an MCP Tool
-
π¨ Now, refresh the page and send any image to perform Gaussian blur processing via LLM
The steps are the same as above, only the workflow part differs (Download workflow: UI format and API format)
The system supports ComfyUI workflows. Just design your workflow in the canvas and export it as API format. Use special syntax in node titles to define parameters and outputs.
In the ComfyUI canvas, double-click the node title to edit, and use the following DSL syntax to define parameters:
$<param_name>.[~]<field_name>[!][:<description>]
-
param_name
: The parameter name for the generated MCP tool function -
~
: Optional, indicates URL parameter upload processing, returns relative path -
field_name
: The corresponding input field in the node -
!
: Indicates this parameter is required -
description
: Description of the parameter
Required parameter example:
- Set LoadImage node title to:
$image.image!:Input image URL
- Meaning: Creates a required parameter named
image
, mapped to the node'simage
field
URL upload processing example:
- Set any node title to:
$image.~image!:Input image URL
- Meaning: Creates a required parameter named
image
, system will automatically download URL and upload to ComfyUI, returns relative path
π Note:
LoadImage
,VHS_LoadAudioUpload
,VHS_LoadVideo
and other nodes have built-in functionality, no need to add~
marker
Optional parameter example:
- Set EmptyLatentImage node title to:
$width.width:Image width, default 512
- Meaning: Creates an optional parameter named
width
, mapped to the node'swidth
field, default value is 512
The system automatically infers parameter types based on the current value of the node field:
- π’
int
: Integer values (e.g. 512, 1024) - π
float
: Floating-point values (e.g. 1.5, 3.14) - β
bool
: Boolean values (e.g. true, false) - π
str
: String values (default type)
The system will automatically detect the following common output nodes:
- πΌοΈ
SaveImage
- Image save node - π¬
SaveVideo
- Video save node - π
SaveAudio
- Audio save node - πΉ
VHS_SaveVideo
- VHS video save node - π΅
VHS_SaveAudio
- VHS audio save node
Usually used for multiple outputs Use
$output.var_name
in any node title to mark output:
- Set node title to:
$output.result
- The system will use this node's output as the tool's return value
You can add a node titled MCP
in the workflow to provide a tool description:
- Add a
String (Multiline)
or similar text node (must have a single string property, and the node field should be one of: value, text, string) - Set the node title to:
MCP
- Enter a detailed tool description in the value field
- π Parameter Validation: Optional parameters (without !) must have default values set in the node
- π Node Connections: Fields already connected to other nodes will not be parsed as parameters
- π·οΈ Tool Naming: Exported file name will be used as the tool name, use meaningful English names
- π Detailed Descriptions: Provide detailed parameter descriptions for better user experience
- π― Export Format: Must export as API format, do not export as UI format
Scan the QR codes below to join our communities for latest updates and technical support:
Discord Community | WeChat Group |
---|---|
![]() |
![]() |
We welcome all forms of contribution! Whether you're a developer, designer, or user, you can participate in the project in the following ways:
- π Submit bug reports on the Issues page
- π Please search for similar issues before submitting
- π Describe the reproduction steps and environment in detail
- π Submit feature requests in Issues
- π Describe the feature you want and its use case
- π― Explain how it improves user experience
- π΄ Fork this repo to your GitHub account
- πΏ Create a feature branch:
git checkout -b feature/your-feature-name
- π» Develop and add corresponding tests
- π Commit changes:
git commit -m "feat: add your feature"
- π€ Push to your repo:
git push origin feature/your-feature-name
- π Create a Pull Request to the main repo
- π Python code follows PEP 8 style guide
- π Add appropriate documentation and comments for new features
- π¦ Share your ComfyUI workflows with the community
- π οΈ Submit tested workflow files
- π Add usage instructions and examples for workflows
β€οΈ Sincere thanks to the following organizations, projects, and teams for supporting the development and implementation of this project.
This project is released under the MIT License (LICENSE, SPDX-License-identifier: MIT).
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for Pixelle-MCP
Similar Open Source Tools

Pixelle-MCP
Pixelle-MCP is a multi-channel publishing tool designed to streamline the process of publishing content across various social media platforms. It allows users to create, schedule, and publish posts simultaneously on platforms such as Facebook, Twitter, and Instagram. With a user-friendly interface and advanced scheduling features, Pixelle-MCP helps users save time and effort in managing their social media presence. The tool also provides analytics and insights to track the performance of posts and optimize content strategy. Whether you are a social media manager, content creator, or digital marketer, Pixelle-MCP is a valuable tool to enhance your online presence and engage with your audience effectively.

trubrics-sdk
Trubrics-sdk is a software development kit designed to facilitate the integration of analytics features into applications. It provides a set of tools and functionalities that enable developers to easily incorporate analytics capabilities, such as data collection, analysis, and reporting, into their software products. The SDK streamlines the process of implementing analytics solutions, allowing developers to focus on building and enhancing their applications' functionality and user experience. By leveraging trubrics-sdk, developers can quickly and efficiently integrate robust analytics features, gaining valuable insights into user behavior and application performance.

tiledesk-dashboard
Tiledesk is an open-source live chat platform with integrated chatbots written in Node.js and Express. It is designed to be a multi-channel platform for web, Android, and iOS, and it can be used to increase sales or provide post-sales customer service. Tiledesk's chatbot technology allows for automation of conversations, and it also provides APIs and webhooks for connecting external applications. Additionally, it offers a marketplace for apps and features such as CRM, ticketing, and data export.

duckduckgo-ai-chat
This repository contains a chatbot tool powered by AI technology. The chatbot is designed to interact with users in a conversational manner, providing information and assistance on various topics. Users can engage with the chatbot to ask questions, seek recommendations, or simply have a casual conversation. The AI technology behind the chatbot enables it to understand natural language inputs and provide relevant responses, making the interaction more intuitive and engaging. The tool is versatile and can be customized for different use cases, such as customer support, information retrieval, or entertainment purposes. Overall, the chatbot offers a user-friendly and interactive experience, leveraging AI to enhance communication and engagement.

chatmcp
Chatmcp is a chatbot framework for building conversational AI applications. It provides a flexible and extensible platform for creating chatbots that can interact with users in a natural language. With Chatmcp, developers can easily integrate chatbot functionality into their applications, enabling users to communicate with the system through text-based conversations. The framework supports various natural language processing techniques and allows for the customization of chatbot behavior and responses. Chatmcp simplifies the development of chatbots by providing a set of pre-built components and tools that streamline the creation process. Whether you are building a customer support chatbot, a virtual assistant, or a chat-based game, Chatmcp offers the necessary features and capabilities to bring your conversational AI ideas to life.

Memento
Memento is a lightweight and user-friendly version control tool designed for small to medium-sized projects. It provides a simple and intuitive interface for managing project versions and collaborating with team members. With Memento, users can easily track changes, revert to previous versions, and merge different branches. The tool is suitable for developers, designers, content creators, and other professionals who need a streamlined version control solution. Memento simplifies the process of managing project history and ensures that team members are always working on the latest version of the project.

Conversational-Azure-OpenAI-Accelerator
The Conversational Azure OpenAI Accelerator is a tool designed to provide rapid, no-cost custom demos tailored to customer use cases, from internal HR/IT to external contact centers. It focuses on top use cases of GenAI conversation and summarization, plus live backend data integration. The tool automates conversations across voice and text channels, providing a valuable way to save money and improve customer and employee experience. By combining Azure OpenAI + Cognitive Search, users can efficiently deploy a ChatGPT experience using web pages, knowledge base articles, and data sources. The tool enables simultaneous deployment of conversational content to chatbots, IVR, voice assistants, and more in one click, eliminating the need for in-depth IT involvement. It leverages Microsoft's advanced AI technologies, resulting in a conversational experience that can converse in human-like dialogue, respond intelligently, and capture content for omni-channel unified analytics.

verl-tool
The verl-tool is a versatile command-line utility designed to streamline various tasks related to version control and code management. It provides a simple yet powerful interface for managing branches, merging changes, resolving conflicts, and more. With verl-tool, users can easily track changes, collaborate with team members, and ensure code quality throughout the development process. Whether you are a beginner or an experienced developer, verl-tool offers a seamless experience for version control operations.

home-gallery
Home-Gallery.org is a self-hosted open-source web gallery for browsing personal photos and videos with tagging, mobile-friendly interface, and AI-powered image and face discovery. It aims to provide a fast user experience on mobile phones and help users browse and rediscover memories from their media archive. The tool allows users to serve their local data without relying on cloud services, view photos and videos from mobile phones, and manage images from multiple media source directories. Features include endless photo stream, video transcoding, reverse image lookup, face detection, GEO location reverse lookups, tagging, and more. The tool runs on NodeJS and supports various platforms like Linux, Mac, and Windows.

Pichome
PicHome is a powerful open-source cloud storage program that efficiently manages various types of files and excels in image and media file management. Its highlights include robust file sharing features and advanced AI-assisted management tools, providing users with a convenient and intelligent file management experience. The program offers diverse list modes, customizable file information display, enhanced quick file preview, advanced tagging, custom cover and preview images, multiple preview images, and multi-library management. Additionally, PicHome features strong file sharing capabilities, allowing users to share entire libraries, create personalized showcase web pages, and build complete data sharing websites. The AI-assisted management aspect includes AI file renaming, tagging, description writing, batch annotation, and file Q&A services, all aimed at improving file management efficiency. PicHome supports a wide range of file formats and can be applied in various scenarios such as e-commerce, gaming, design, development, enterprises, schools, labs, media, and entertainment institutions.

DocsGPT
DocsGPT is an open-source documentation assistant powered by GPT models. It simplifies the process of searching for information in project documentation by allowing developers to ask questions and receive accurate answers. With DocsGPT, users can say goodbye to manual searches and quickly find the information they need. The tool aims to revolutionize project documentation experiences and offers features like live previews, Discord community, guides, and contribution opportunities. It consists of a Flask app, Chrome extension, similarity search index creation script, and a frontend built with Vite and React. Users can quickly get started with DocsGPT by following the provided setup instructions and can contribute to its development by following the guidelines in the CONTRIBUTING.md file. The project follows a Code of Conduct to ensure a harassment-free community environment for all participants. DocsGPT is licensed under MIT and is built with LangChain.

J.A.R.V.I.S.
J.A.R.V.I.S.1.0 is an advanced virtual assistant tool designed to assist users in various tasks. It provides a wide range of functionalities including voice commands, task automation, information retrieval, and communication management. With its intuitive interface and powerful capabilities, J.A.R.V.I.S.1.0 aims to enhance productivity and streamline daily activities for users.

WorkflowAI
WorkflowAI is a powerful tool designed to streamline and automate various tasks within the workflow process. It provides a user-friendly interface for creating custom workflows, automating repetitive tasks, and optimizing efficiency. With WorkflowAI, users can easily design, execute, and monitor workflows, allowing for seamless integration of different tools and systems. The tool offers advanced features such as conditional logic, task dependencies, and error handling to ensure smooth workflow execution. Whether you are managing project tasks, processing data, or coordinating team activities, WorkflowAI simplifies the workflow management process and enhances productivity.

aiounifi
Aiounifi is a Python library that provides a simple interface for interacting with the Unifi Controller API. It allows users to easily manage their Unifi network devices, such as access points, switches, and gateways, through automated scripts or applications. With Aiounifi, users can retrieve device information, perform configuration changes, monitor network performance, and more, all through a convenient and efficient API wrapper. This library simplifies the process of integrating Unifi network management into custom solutions, making it ideal for network administrators, developers, and enthusiasts looking to automate and streamline their network operations.

Automodel
Automodel is a Python library for automating the process of building and evaluating machine learning models. It provides a set of tools and utilities to streamline the model development workflow, from data preprocessing to model selection and evaluation. With Automodel, users can easily experiment with different algorithms, hyperparameters, and feature engineering techniques to find the best model for their dataset. The library is designed to be user-friendly and customizable, allowing users to define their own pipelines and workflows. Automodel is suitable for data scientists, machine learning engineers, and anyone looking to quickly build and test machine learning models without the need for manual intervention.
For similar tasks

intro-llm-rag
This repository serves as a comprehensive guide for technical teams interested in developing conversational AI solutions using Retrieval-Augmented Generation (RAG) techniques. It covers theoretical knowledge and practical code implementations, making it suitable for individuals with a basic technical background. The content includes information on large language models (LLMs), transformers, prompt engineering, embeddings, vector stores, and various other key concepts related to conversational AI. The repository also provides hands-on examples for two different use cases, along with implementation details and performance analysis.

LLM-Viewer
LLM-Viewer is a tool for visualizing Language and Learning Models (LLMs) and analyzing performance on different hardware platforms. It enables network-wise analysis, considering factors such as peak memory consumption and total inference time cost. With LLM-Viewer, users can gain valuable insights into LLM inference and performance optimization. The tool can be used in a web browser or as a command line interface (CLI) for easy configuration and visualization. The ongoing project aims to enhance features like showing tensor shapes, expanding hardware platform compatibility, and supporting more LLMs with manual model graph configuration.

llm-colosseum
llm-colosseum is a tool designed to evaluate Language Model Models (LLMs) in real-time by making them fight each other in Street Fighter III. The tool assesses LLMs based on speed, strategic thinking, adaptability, out-of-the-box thinking, and resilience. It provides a benchmark for LLMs to understand their environment and take context-based actions. Users can analyze the performance of different LLMs through ELO rankings and win rate matrices. The tool allows users to run experiments, test different LLM models, and customize prompts for LLM interactions. It offers installation instructions, test mode options, logging configurations, and the ability to run the tool with local models. Users can also contribute their own LLM models for evaluation and ranking.

eureka-ml-insights
The Eureka ML Insights Framework is a repository containing code designed to help researchers and practitioners run reproducible evaluations of generative models efficiently. Users can define custom pipelines for data processing, inference, and evaluation, as well as utilize pre-defined evaluation pipelines for key benchmarks. The framework provides a structured approach to conducting experiments and analyzing model performance across various tasks and modalities.

Pixelle-MCP
Pixelle-MCP is a multi-channel publishing tool designed to streamline the process of publishing content across various social media platforms. It allows users to create, schedule, and publish posts simultaneously on platforms such as Facebook, Twitter, and Instagram. With a user-friendly interface and advanced scheduling features, Pixelle-MCP helps users save time and effort in managing their social media presence. The tool also provides analytics and insights to track the performance of posts and optimize content strategy. Whether you are a social media manager, content creator, or digital marketer, Pixelle-MCP is a valuable tool to enhance your online presence and engage with your audience effectively.

trae-agent
Trae-agent is a Python library for building and training reinforcement learning agents. It provides a simple and flexible framework for implementing various reinforcement learning algorithms and experimenting with different environments. With Trae-agent, users can easily create custom agents, define reward functions, and train them on a variety of tasks. The library also includes utilities for visualizing agent performance and analyzing training results, making it a valuable tool for both beginners and experienced researchers in the field of reinforcement learning.

basehub
JavaScript / TypeScript SDK for BaseHub, the first AI-native content hub. **Features:** * β¨ Infers types from your BaseHub repository... _meaning IDE autocompletion works great._ * ποΈ No dependency on graphql... _meaning your bundle is more lightweight._ * π Works everywhere `fetch` is supported... _meaning you can use it anywhere._

AI-Prompt-Genius
AI Prompt Genius is a Chrome extension that allows you to curate a custom library of AI prompts. It is built using React web app and Tailwind CSS with DaisyUI components. The extension enables users to create and manage AI prompts for various purposes. It provides a user-friendly interface for organizing and accessing AI prompts efficiently. AI Prompt Genius is designed to enhance productivity and creativity by offering a personalized collection of prompts tailored to individual needs. Users can easily install the extension from the Chrome Web Store and start using it to generate AI prompts for different tasks.
For similar jobs

LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.

daily-poetry-image
Daily Chinese ancient poetry and AI-generated images powered by Bing DALL-E-3. GitHub Action triggers the process automatically. Poetry is provided by Today's Poem API. The website is built with Astro.

exif-photo-blog
EXIF Photo Blog is a full-stack photo blog application built with Next.js, Vercel, and Postgres. It features built-in authentication, photo upload with EXIF extraction, photo organization by tag, infinite scroll, light/dark mode, automatic OG image generation, a CMD-K menu with photo search, experimental support for AI-generated descriptions, and support for Fujifilm simulations. The application is easy to deploy to Vercel with just a few clicks and can be customized with a variety of environment variables.

SillyTavern
SillyTavern is a user interface you can install on your computer (and Android phones) that allows you to interact with text generation AIs and chat/roleplay with characters you or the community create. SillyTavern is a fork of TavernAI 1.2.8 which is under more active development and has added many major features. At this point, they can be thought of as completely independent programs.

Twitter-Insight-LLM
This project enables you to fetch liked tweets from Twitter (using Selenium), save it to JSON and Excel files, and perform initial data analysis and image captions. This is part of the initial steps for a larger personal project involving Large Language Models (LLMs).

AISuperDomain
Aila Desktop Application is a powerful tool that integrates multiple leading AI models into a single desktop application. It allows users to interact with various AI models simultaneously, providing diverse responses and insights to their inquiries. With its user-friendly interface and customizable features, Aila empowers users to engage with AI seamlessly and efficiently. Whether you're a researcher, student, or professional, Aila can enhance your AI interactions and streamline your workflow.

ChatGPT-On-CS
This project is an intelligent dialogue customer service tool based on a large model, which supports access to platforms such as WeChat, Qianniu, Bilibili, Douyin Enterprise, Douyin, Doudian, Weibo chat, Xiaohongshu professional account operation, Xiaohongshu, Zhihu, etc. You can choose GPT3.5/GPT4.0/ Lazy Treasure Box (more platforms will be supported in the future), which can process text, voice and pictures, and access external resources such as operating systems and the Internet through plug-ins, and support enterprise AI applications customized based on their own knowledge base.

obs-localvocal
LocalVocal is a live-streaming AI assistant plugin for OBS that allows you to transcribe audio speech into text and perform various language processing functions on the text using AI / LLMs (Large Language Models). It's privacy-first, with all data staying on your machine, and requires no GPU, cloud costs, network, or downtime.