Gaia-X Next-Generation Enterprise AI Application Platform Based on AI New Paradigm
🌟 Project Overview
Gaia-X is the first chatbot application platform designed for enterprise scenarios based on the new AI paradigm, addressing the core pain points of existing AI products with innovative technical architecture:
🚀 Enterprise Management | 🤖 MCP Protocol Support | 👥 Multi-Agent Collaboration
💻 Natural Language RPA | ✅ Human Confirmation Mechanism | 🎨 Intelligent Canvas Rendering
Try Now | Admin Center | Documentation Center
🎯 Core Pain Point Solutions
Pain Point Area |
Gaia-X Innovative Solutions |
Lack of Enterprise Management |
Complete user/permission/billing system and LLM API hosting |
No MCP Protocol Support |
The first enterprise-grade MCP support |
Risk of Sensitive Operations |
ReAct tool calls with human confirmation, dynamic rendering |
Difficulty in Natural Language RPA |
Large model-driven RPA automation |
Weak Multi-Agent Collaboration |
Intelligent Agent retrieval and multi-agent collaboration for complex tasks |
📦 Project Architecture
1.1 Overall Project Architecture

1.2 MCP Call Process
During the Agent call process, the management center is not responsible for calling MCP tools. Instead, after the client's node program receives the response results, it will uniformly determine if there are function calls. If so, it will call the corresponding MCP tools to execute. The execution chain is roughly as follows:
graph LR
U --> |1|C[Gaia-X Client]
C-->|2|API[Admin Center]
API-->|3|LLM[LLM API]
LLM-->|4|API[Admin Center]
API-->|5|C[Gaia-X Client]
C-->|6|B{Function Call?}
B -- No -->U[User]
B -- Yes --> MCP[MCP Servers]
MCP -->|8a. Call| H[Tool 1]
MCP -->|8b. Call| I[Tool 2]
H-->|8.1|C[Gaia-X Client]
I-->|8.1|C[Gaia-X Client]
C-->|8.2|API[Admin Center]
API-->|8.3|LLM[LLM API]
LLM-->|8.4|API[Admin Center]
API-->|8.5|C[Gaia-X Client]
C-->|8.6|U[User]
subgraph Possible Nested/Iterative Calls
H
I
end
1.3 Client Authentication Logic at Startup
As this is an enterprise-level project, the client needs to complete login authentication before entering the Chatbot interface. Using OAuth2.0 as an example (other logic is similar), the authentication process is as follows:
sequenceDiagram
participant C as Client Program
participant A as Admin Center
participant O as OAuth2.0 Server
participant U as User
C->>A: Request authentication page (non-OAuth2.0 login page)
A->>A: Check current login status
alt Not logged in
A->>O: Redirect and open OAuth2.0 login page
O->>U: Display login interface
U->>O: Submit credentials for login
O->>A: Return authentication result (e.g., Token)
A->>A: Generate JWT after verifying Token and update login status
else Already logged in
A->>U: Display authorization button
end
U->>A: Click authorize
A->>C: Redirect back to client and pass authorization information (gaia://oauth-callback?code=xyz)
C->>C: Client completes login process
Process Explanation :
- Client requests authentication page from admin center;
- Admin backend determines user login status;
- If not logged in, redirects to OAuth2.0 login page;
- User submits credentials on OAuth2.0 login page, returns authentication result after successful login;
- After OAuth2.0 login, redirects to admin backend with authentication information, generates JWT and updates login status;
- User clicks authorize button, admin backend redirects back to client, completing the login.
2. Core Features (Some still under development)
2.1 MCP (Model Context Protocol) Support
2.2 Multi-Agent Intelligent Coordination
2.3 New Paradigm Infinite Canvas
2.4 Intelligent Computer Operations
2.5 Text Selection Analysis
2.6 Enterprise Management Center
2.7 Third-party Agent Integration
3. Technology Stack
3.1 Client
The client adopts a plugin-based design, and will later be architected with a micro-kernel plugin architecture similar to VS Code, enabling arbitrary extensions.
- Framework : Electron + React
- LLM UI : Ant Design X
- Text Selection Monitoring Technology : C++ (Windows), Objective-C (MacOS)
3.2 Admin Backend
The client only serves as an interactive UI and MCP management execution, without actual model calls. Model calls and tool calls are implemented by the backend. However, the backend is not responsible for the specific execution of MCP tools.
- Large Model Interaction : Eino Framework + Self-developed Multi-Agent
- Admin UI : Ant Design Pro (refactoring the frontend page code of GVA for technology stack uniformity)
- API Service : golang + Gin (based on the GVA framework)
3.3 MCP Server
- Runs in independent sandbox environment
- Supports implementation in both Python and TypeScript
4. Model Selection Reference
- Computer Operation Models : Claude Sonnet 3.5+, CogAgent, UI-TARS, OpenAI computer-use
- ReAct Recommended Models : Claude Sonnet 3.5+, GPT 4o (DeepSeek v3 model performs poorly, not recommended for critical tasks)
5. Acknowledgements
The admin API is built based on Gin-Vue-Admin, reducing much of our basic management backend development work.
Eino by ByteDance is one of the few golang-based large model application development frameworks, solving basic model integration and Agent issues.
We used Ant Design Pro to refactor the frontend pages of Gin-Vue-Admin, unifying the UI frameworks for both the admin backend and client, reducing community development difficulty.
We've made many secondary developments based on Eino, including integration of common model providers and Multi-Agent architecture. After secondary development, we published it as Eino-X.
To solve the problem of running MCP Servers on ordinary user clients, we packaged an embedded binary MCP runtime environment (supporting MacOS and Windows) under the project mcp-runtime.
6. License
This project is licensed under the Apache 2.0 License. Feel free to customize or contribute!
7. Community Support
We welcome feedback and suggestions through Issues. We look forward to your participation in building the future enterprise AI application ecosystem together.
🎉 Thank you for your attention and support to the project!