UBOS Asset Marketplace: Gemini Multimodal Chat with MCP Server
In the rapidly evolving landscape of AI, multimodal interaction stands out as a pivotal advancement, bridging the gap between human communication and machine understanding. The UBOS Asset Marketplace now features an innovative MCP Server that allows users to deploy a Gemini multimodal chat website in just 10 seconds. This tool simplifies the integration of advanced AI capabilities, making it accessible to developers and businesses of all sizes.
What is an MCP Server?
MCP (Model Context Protocol) is an open protocol designed to standardize how applications provide context to Large Language Models (LLMs). In essence, an MCP server acts as a bridge, enabling AI models to access and interact with external data sources and tools. This capability is crucial for creating more informed, context-aware AI applications. The Gemini Multimodal Chat, hosted on an MCP server, leverages this protocol to enhance its ability to understand and respond to diverse inputs.
Use Cases
- Rapid Prototyping of AI Chatbots: Developers can quickly deploy a fully functional Gemini-powered chatbot for prototyping and testing new features without the complexities of server management.
- Customer Support Enhancement: Businesses can integrate the Gemini chat into their customer support systems, providing instant, AI-driven assistance that understands and responds to customer queries with relevant context.
- Educational Tools: Educators can use the platform to create interactive learning environments where students can engage with AI tutors capable of understanding and responding to multimodal inputs.
- Content Creation: Content creators can leverage the AI to generate creative content, such as image captions, video scripts, and interactive narratives, with minimal setup.
- Accessibility Solutions: The multimodal capabilities can be used to develop accessibility tools that aid individuals with disabilities, such as converting speech to text with image context or providing visual descriptions of audio content.
- AI-Powered Research: Researchers can use the tool to quickly set up AI-driven research platforms that can analyze and respond to complex data sets, integrating text, images, and other media types.
Key Features
- 10-Second Deployment: The primary allure is the incredibly fast deployment time. By leveraging serverless architecture, users can launch a fully functional Gemini multimodal chat website almost instantly.
- Serverless Architecture: Utilizing Deno and Cloudflare Workers for deployment, the server operates without the need for managing traditional server infrastructure. This leads to cost savings and simplified maintenance.
- Multimodal Input: The platform supports a variety of input methods, including text, images, audio, and video, allowing users to interact with the AI in the most natural way possible.
- OpenAI Format Compatibility: The Gemini Chat API is converted to the more universal OpenAI format, ensuring compatibility with a wide range of AI clients and programming tools.
- Domestic Direct Connection: By providing a direct connection within domestic networks, the platform bypasses common network restrictions, ensuring smooth and uninterrupted access.
- Mobile-Friendly: The platform is fully optimized for mobile devices, ensuring a seamless user experience across all screen sizes.
- API Proxy: Functions as an API proxy, streamlining the integration of the Gemini Chat API into existing applications and workflows.
Deployment Options
The MCP Server offers two primary deployment methods:
1. Deno Deployment (Recommended)
Deno deployment is recommended for its simplicity and efficiency. Here’s how to deploy using Deno:
- Obtain a Gemini API Key: Sign up at aistudio.google.com to get your free Gemini API Key.
- Fork the Project: Fork the gemini-playground repository on GitHub.
- Set Up Deno Deploy: Log in or register at dash.deno.com.
- Create a New Project: Create a new project and fill in the project details, paying close attention to the project name as it determines the automatically assigned domain.
- Configure Entrypoint: Set the entry point to
src/deno_index.tsand leave other fields blank. - Deploy: Click the Deploy Project button to launch the application. Once deployed, the domain will be available for use.
2. Cloudflare Worker Deployment
Cloudflare Worker deployment provides another robust option for hosting the Gemini multimodal chat. Here’s how to deploy using Cloudflare Workers:
- Obtain a Gemini API Key: Sign up at aistudio.google.com to get your free Gemini API Key.
- Deploy to Cloudflare Workers: Use the provided deployment button.
- Log in to Cloudflare: Log in to your Cloudflare account.
- Enter Account ID and API Token: Fill in your Cloudflare Account ID and API Token.
- Fork and Enable GitHub Actions: Fork the project and enable GitHub Actions.
- Deploy: Access the Cloudflare dashboard to view the deployed worker. For domestic use, binding a custom domain is recommended.
Local Debugging
For local debugging, you’ll need to install Deno on your machine.
Windows: Use the following command in PowerShell:
powershell irm https://deno.land/install.ps1 | iex
Mac/Linux: Use the following command in your terminal:
bash curl -fsSL https://deno.land/install.sh | sh
To start the project locally:
bash cd 项目目录 deno run --allow-net --allow-read src/deno_index.ts
How to Use
Multimodal Conversation
Enter your API Key and click the Connect button to start the conversation. The platform provides buttons for:
- Enabling the microphone
- Enabling the camera
- Sharing the screen
API Proxy
The API is proxied to the OpenAI format, allowing the use of OpenAI-formatted API calls without restrictions.
Available Models List:
bash
curl --location ‘http://your.domain.com/v1/models’
–header ‘Authorization: Bearer YOUR-GEMINI-API-KEY’
Conversation:
bash
curl --location ‘https://your.domain.com/v1/chat/completions’
–header ‘Authorization: Bearer YOUR-GEMINI-API-KEY’
–header ‘Content-Type: application/json’
–data ‘{
“messages”: [
{
“role”: “system”,
“content”: “You are a test assistant.”
},
{
“role”: “user”,
“content”: “Testing. Just say hi and nothing else.”
}
],
“model”: “gemini-2.0-flash-exp”
}’
Integration with ChatBOX and Cursor
The MCP Server seamlessly integrates with tools like ChatBOX and Cursor, enhancing its utility for AI programming and general chat applications.
Why UBOS?
UBOS is a full-stack AI Agent Development Platform designed to bring AI Agents to every business department. Our platform helps you:
- Orchestrate AI Agents
- Connect AI Agents with your enterprise data
- Build custom AI Agents with your LLM model
- Create Multi-Agent Systems
The UBOS Asset Marketplace is a curated collection of tools and components designed to accelerate AI development. By providing pre-built solutions like the Gemini Multimodal Chat MCP Server, UBOS empowers developers to focus on innovation rather than infrastructure. The platform’s commitment to open standards, like MCP, ensures interoperability and reduces vendor lock-in, fostering a collaborative and dynamic AI ecosystem.
Conclusion
The Gemini Multimodal Chat MCP Server on the UBOS Asset Marketplace represents a significant step forward in simplifying AI deployment. By offering a serverless, rapidly deployable solution that supports multimodal inputs and OpenAI compatibility, UBOS is lowering the barrier to entry for AI innovation. Whether you are a developer prototyping new AI features, a business enhancing customer support, or an educator creating interactive learning environments, this tool provides the capabilities you need to succeed. Embrace the future of AI with UBOS and unlock the potential of multimodal interaction today.
Gemini 2.0 PlayGround
Project Details
- playWithAIGC/gemini-playground
- MIT License
- Last Updated: 1/9/2025
Recomended MCP Servers
A Model Context Protocol (MCP) server for Apache Dolphinscheduler. This provides access to your Apache Dolphinshcheduler RESTful API...
MCP Server to connect Notion with Claude
Figma MCP Server with full API functionality
HeyBeauty MCP Server for Virtual TryOn
Dify 1.0 Plugin MCP HTTP with SSE or Streamable HTTP transport Tools
An MCP server that provides Unix command documentation directly within LLMs conversations.
MCP server that facilitates an SSH connection to a deployed Rails app so you can run Rails REPL...
An MCP server to create secure code sandbox environment for executing code within Docker containers. This MCP server...





