wanwu

wanwu

China Unicom's Yuanjing Wanwu Agent Platform is an enterprise-grade, multi-tenant AI agent development platform. It helps users build applications such as intelligent agents, workflows, and rag, and also supports model management. The platform features a developer-friendly license, and we welcome all developers to build upon the platform.

Stars: 1398

Visit
 screenshot

Wanwu AI Agent Platform is an enterprise-grade one-stop commercially friendly AI agent development platform designed for business scenarios. It provides enterprises with a safe, efficient, and compliant one-stop AI solution. The platform integrates cutting-edge technologies such as large language models and business process automation to build an AI engineering platform covering model full life-cycle management, MCP, web search, AI agent rapid development, enterprise knowledge base construction, and complex workflow orchestration. It supports modular architecture design, flexible functional expansion, and secondary development, reducing the application threshold of AI technology while ensuring security and privacy protection of enterprise data. It accelerates digital transformation, cost reduction, efficiency improvement, and business innovation for enterprises of all sizes.

README:

Feature ListTypical Application ScenariosQuick StartUsing WanwuQ & AContact Us

License Go Version Release Notes

English | 简体中文 | 繁體中文

Wanwu AI Agent Platform is an enterprise-grade one-stop commercially friendly AI agent development platform designed for business scenarios. It is committed to providing enterprises with a safe, efficient, and compliant one-stop AI solution. With the core philosophy of "technology openness and ecological co-construction", we integrate cutting-edge technologies such as large language models and business process automation to build an AI engineering platform with a complete functional system covering model full life-cycle management, MCP, web search, AI agent rapid development, enterprise knowledge base construction, and complex workflow orchestration. The platform adopts a modular architecture design, supports flexible functional expansion and secondary development, and greatly reduces the application threshold of AI technology while ensuring the security and privacy protection of enterprise data. Whether it is for small and medium-sized enterprises to quickly build intelligent applications or for large enterprises to achieve intelligent transformation of complex business scenarios, the Wanwu AI Agent Platform can provide strong technical support to help enterprises accelerate the process of digital transformation, achieve cost reduction and efficiency improvement, and business innovation.



🔥 Adopt a permissive and friendly Apache 2.0 License, supporting developers to freely expand and develop secondary

Enterprise-level engineering: Provides a complete toolchain from model management to application landing, solving the "last mile" problem of LLM technology landing

Open-source ecological: Adopt a permissive and friendly Apache 2.0 License, supporting developers to freely expand and develop

Full-stack technology support: Equipped with a professional team to provide architecture consulting, performance optimization and full-cycle empowerment for ecological partners

Multi-tenant architecture: Provides a multi-tenant account system to meet the core needs of users in cost control, data security isolation, business elasticity expansion, industry customization, rapid online and ecological collaboration

XinChuang adaptation: Already adapted to domestic XinChuang databases TiDB and OceanBase


🚩 Core Function Modules

1. Model Management (Model Hub) ▸ Supports the unified access and lifecycle management of hundreds of proprietary/open-source large models (including GPT, Claude, Llama, etc.)

▸ Deeply adapts to OpenAI API standards and Unicom Yuanjing ecological models, realizing seamless switching of heterogeneous models

▸ Provides multi-inference backend support (vLLM, TGI, etc.) and self-hosted solutions to meet the computing power needs of enterprises of different scales

2. MCP

Standardized interfaces: Enable AI models to seamlessly connect to various external tools (such as GitHub, Slack, databases, etc.) without the need to develop adapters for each data source separately

Built-in rich and selected recommendations: Integrates 100+ industry MCP interfaces, making it easy for users to call up quickly and easily

3. Web Search

Real-time information acquisition: Possesses powerful web search capabilities, capable of obtaining the latest information from the Internet in real-time. In question and answer scenarios, when a user's question requires the latest news, data, and other information, the platform can quickly search and return accurate results, enhancing the timeliness and accuracy of the answers

Multi-source data integration: Integrates various Internet data sources, including news websites, academic databases, industry reports, etc. Through the integration and analysis of multi-source data, it provides users with more comprehensive and in-depth information. For example, in market research scenarios, relevant data can be obtained from multiple data sources at the same time for comprehensive analysis and evaluation

Intelligent search strategy: Adopt intelligent search algorithms, automatically optimize search strategies based on user questions to improve search efficiency and accuracy. Support keyword search, semantic search and other search methods to meet the needs of different users. At the same time, intelligently sort and filter search results, prioritize the display of the most relevant and valuable information

4. Visual Workflow (Workflow Studio)

▸ Quickly build complex AI business processes through low-code drag-and-drop canvas

▸ Built-in conditional branching, API, large model, knowledge base, code, MCP and other nodes, support end-to-end process debugging and performance analysis

5. Enterprise-level Knowledge Base, RAG Pipeline

▸ Provides the whole process knowledge management capabilities of knowledge base creation → document parsing → vectorization → retrieval → fine sorting, supports multiple formats such as pdf/docx/txt/xlsx/csv/pptx documents, and also supports the capture and access of web resources

▸ Integrates multi-modal retrieval, cascading segmentation and adaptive segmentation, significantly improves the accuracy of Q&A

6. AI Agent Development Framework (Agent Framework)

▸ Can be based on the function call (Function Calling) agent construction paradigm, supports tool expansion, private knowledge base association and multi-round dialogue

▸ Support online debugging

7. Backend as a Service (BaaS)

▸ Provides RESTful API, supports deep integration with existing enterprise systems (OA/CRM/ERP, etc.)

▸ Provides fine-grained permission control to ensure stable operation in production environments


📢 Function Comparison

Function Wanwu Dify.AI Fastgpt Ragflow Coze open source version
Model import ❌(Built-in models) ❌(Built-in models)
RAG engine
MCP ✅(Need to install tools to use)
Direct OCR import
Search enhancement ✅(Need to install tools to use) ✅(Need to install tools to use)
Agent
Workflow
Local deployment
license friendly ❌(Commercially restricted) ❌(Commercially restricted) Not fully open source
Multi-tenant ❌(Commercially restricted) ❌(Commercially restricted) ✅(Users are not interconnected)

As of August 1, 2025.


🎯 Typical Application Scenarios

  • Intelligent Customer Service: Realize high-accuracy business consultation and ticket processing based on RAG + Agent
  • Knowledge Management: Build an exclusive enterprise knowledge base, support semantic search and intelligent summary generation
  • Process Automation: Realize AI-assisted decision-making for business processes such as contract review and reimbursement approval through the workflow engine The platform has been successfully applied in multiple industries such as finance, industry, and government, helping enterprises transform the theoretical value of LLM technology into actual business benefits. We sincerely invite developers to join the open source community and jointly promote the democratization of AI technology.

🚀 Quick Start

  • The workflow module of the Wanwu AI Agent Platform uses the following project, you can go to its warehouse to view the details.
  • Docker Installation (Recommended)
  1. Before the first run

    1.1 Copy the environment variable file

    cp .env.bak .env

    1.2 Modify the WANWU_ARCH and WANWU_EXTERNAL_IP variables in the .env file according to the system

    # amd64 / arm64
    WANWU_ARCH=amd64
    
    # external ip port (Note: if the browser accesses Wanwu deployed on a non-localhost server, you need to change localhost to the external IP, for example, 192.168.xx.xx)
    WANWU_EXTERNAL_IP=localhost
    

    1.3 Create a Docker running network

    docker network create wanwu-net
    
  2. Start the service (the image will be automatically pulled from Docker Hub during the first run)

    # For amd64 system:
    docker compose --env-file .env --env-file .env.image.amd64 up -d
    # For arm64 system:
    docker compose --env-file .env --env-file .env.image.arm64 up -d
  3. Log in to the system: http://localhost:8081

    Default user: admin
    Default password: Wanwu123456
    
  4. Stop the service

    # For amd64 system:
    docker compose --env-file .env --env-file .env.image.amd64 down
    # For arm64 system:
    docker compose --env-file .env --env-file .env.image.arm64 down
  • Source Code Start (Development)
  1. Based on the above Docker installation steps, start the system service completely
  2. Take the backend bff-service service as an example 2.1 Stop bff-service
    make -f Makefile.develop stop-bff
    
    2.2 Compile the bff-service executable file
    # For amd64 system:
    make build-bff-amd64
    # For arm64 system:
    make build-bff-arm64
    
    2.3 Start bff-service
    make -f Makefile.develop run-bff
    

⬆️ Version Upgrade

  1. Based on the above Docker installation steps, completely stop the system service
  2. Update to the latest version of the code 2.1 In the wanwu repository directory, update the code
    # Switch to the main branch
    git checkout main
    # Pull the latest code
    git pull
    2.2 Recopy the environment variable file (if there are changes to the environment variables, please modify them again)
    # Backup the current .env file
    cp .env .env.old
    # Copy the .env file
    cp .env.bak .env
  3. Based on the above Docker installation steps, completely start the system service

📑 Using Wanwu

To help you quickly get started with this project, we strongly recommend that you first check out the Documentation Operation Manual. We provide users with interactive and structured operation guides, where you can directly view operation instructions, interface documents, etc., greatly reducing the threshold for learning and use. The detailed function list is as follows:

Feature Detailed Description
Model Management Supports users to import LLM, Embedding, and Rerank models from various model providers, including Unicom Yuanjing, OpenAI-API-compatible, Ollama, Tongyi Qianwen, and Volcano Engine. Model Import Methods - Detailed Version
Knowledge Base In terms of document parsing capabilities: supports uploading of 12 file types and URL parsing; document parsing methods include OCR and high-precision model parsing (titles/tables/formulas); document segmentation settings support both general segmentation and parent-child segmentation. In terms of optimization capabilities: supports metadata management and metadata filtering queries, supports adding, deleting, and modifying segmented content, supports setting keyword tags for segments to improve recall performance, supports segment enable/disable operations, and supports hit testing. In terms of retrieval capabilities: supports multiple retrieval modes including vector search, full-text search, and hybrid search. In terms of Q&A capabilities: supports automatic citation of sources and generating answers with both text and images.<br
Resource Library Supports importing your own MCP services or custom tools for use in workflows and agents.
Safety Guardrails Users can create sensitive word lists to control the safety of the model's output.
Text Q&A A dedicated knowledge advisor based on a private knowledge base. It supports features like knowledge base management, Q&A, knowledge summarization, personalized parameter configuration, safety guardrails, and retrieval configuration to improve the efficiency of knowledge management and learning. Supports publishing text Q&A applications publicly or privately, and can be published as an API.
Workflow Extends the capabilities of agents. Composed of nodes, it provides a visual workflow editor. Users can orchestrate multiple different workflow nodes to implement complex and stable business processes. Supports publishing workflow applications publicly or privately, can be published as an API, and supports import/export.
Agent Create agents based on user scenarios and business requirements. Supports model selection, prompt setting, web search, knowledge base selection, MCP, workflows, and custom tools. Supports publishing agent applications publicly or privately, and can be published as an API and a Web URL.
App Marketplace Allows users to experience published applications, including Text Q&A, Workflows, and Agents.
MCP Hub Features 100+ pre-selected industry-specific MCP servers, ready for immediate use.
Settings The platform supports multi-tenancy, allowing users to manage organizations, roles, users, and perform basic platform configuration.

📰 TO DO LIST

  • [ ] Multi-modal model access
  • [ ] Support custom MCP Server, which means that workflows, agents, or APIs that conform to the OpenAPI specification can be added to the MCP Server for release
  • [ ] Knowledge base sharing
  • [ ] Agent and model evaluation
  • [ ] Agent monitoring statistics
  • [ ] Model experience
  • [ ] Prompt engineering

💬 Q & A

  • [Q] Error when starting Elastic (elastic-wanwu) on Linux system: Memory limited without swap. [A] Stop the service, run sudo sysctl -w vm.max_map_count=262144, and then restart the service.

  • [Q] After the system services start normally, the mysql-wanwu-setup and elastic-wanwu-setup containers exit with status code Exited (0). [A] This is normal. These two containers are used to complete some initialization tasks and will automatically exit after execution.

  • [Q] Regarding model import [A] Taking the import of Unicom Yuanjing LLM as an example (the process is similar for importing OpenAI-API-compatible models, Embedding, or Rerank types):

    1. The Open API interface for Unicom Yuanjing MaaS Cloud LLM is, for example: https://maas.ai-yuanjing.com/openapi/compatible-mode/v1/chat/completions
    2. The API Key applied for by the user on Unicom Yuanjing MaaS Cloud looks like: sk-abc********************xyz
    3. Confirm that the API and Key can correctly request the LLM. Taking a request to yuanjing-70b-chat as an example:
        curl --location 'https://maas.ai-yuanjing.com/openapi/compatible-mode/v1/chat/completions' \
        --header 'Content-Type: application/json' \
        --header 'Accept: application/json' \
        --header 'Authorization: Bearer sk-abc********************xyz' \
        --data '{
                "model": "yuanjing-70b-chat",
                "messages": [{
                        "role": "user",
                        "content": "你好"
                }]
        }'
    4. Import the model:
    4.1 [Model Name] must be the model that can be correctly requested in the curl command above; for example, yuanjing-70b-chat.
    4.2 [API Key] must be the key that can be correctly requested in the curl command above; for example, sk-abc********************xyz (note: do not include the 'Bearer' prefix).
    4.3 [Inference URL] must be the URL that can be correctly requested in the curl command above; for example, https://maas.ai-yuanjing.com/openapi/compatible-mode/v1 (note: do not include the /chat/completions suffix).
    5. Importing an Embedding model is the same as importing an LLM as described above. Note that the inference URL should not include the /embeddings suffix.
    6. Importing a Rerank model is the same as importing an LLM as described above. Note that the inference URL should not include the /rerank suffix.
    

🔗 Acknowledgments


⚖️ License

The Yuanjing Wanwu AI Agent Platform is released under the Apache License 2.0.


📩 Contact Us

QQ Group1(Full):490071123 QQ Group2:1026898615
image image

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for wanwu

Similar Open Source Tools

For similar tasks

For similar jobs