open-responses

open-responses

OpenResponses API empowers developers to leverage the incredible capabilities of various LLM providers through a familiar interface - the OpenAI Responses API structure.

Stars: 54

Visit
 screenshot

README:

OpenResponses API

OpenResponses - All-in-One AI Toolkit

Unlock enterprise-grade AI capabilities through a single, powerful API โ€” simplify development, accelerate deployment, and maintain complete data control

Overview

OpenResponses revolutionizes how developers build AI applications by providing a comprehensive, production-ready toolkit with essential enterprise featuresโ€”all through an elegantly simplified API interface. Stop cobbling together disparate tools and start building what matters.

Key Problems Solved

Development Challenges

  • Feature Gap: Most open-source AI models lack critical enterprise capabilities required for production environments
  • Integration Complexity: Implementing supplementary features like retrieval augmentation and monitoring requires significant development overhead
  • Resource Diversion: Engineering teams spend excessive time on infrastructure rather than core application logic

Operational Concerns

  • Data Privacy: Organizations with sensitive data face compliance barriers when using cloud-hosted AI services
  • Operational Control: Many applications require full control over the AI processing pipeline

Core Capabilities

Feature Description Benefit
Automated Tracing Comprehensive request and response monitoring Track performance and usage without additional code
Integrated RAG Contextual information retrieval Enhance responses with relevant external data automatically
Pre-built Tool Integrations Web search, GitHub access, and more Deploy advanced capabilities instantly
Self-Hosted Architecture Full control of deployment infrastructure Maintain complete data sovereignty
OpenAI-Compatible Interface Drop-in replacement for existing OpenAI implementations Minimal code changes for migration

Why Engineering Teams Should Choose OpenResponses

  • Developer Productivity: Focus engineering efforts on application features rather than infrastructure
  • Production Readiness: Enterprise capabilities and batteries included out-of-the-box
  • Compliance Confidence: Deploy with data privacy requirements fully addressed
  • Simplified Architecture: Consolidate AI infrastructure through widely used OpenAI API Specifications

Getting Started

Visit our Quick Start Guide to deploy OpenResponses in your environment in minutes.

The API implements the following OpenAI-compatible endpoints:

Endpoint Description
POST /v1/responses Create a new model response
GET /v1/responses/{responseId} Retrieve a specific response
DELETE /v1/responses/{responseId} Delete a response
GET /v1/responses/{responseId}/input_items List input items for a response

๐Ÿš€ Quick Start

Ready to transform your AI development workflow? Follow our step-by-step guide here: Quick Start Guide

Example API Calls

Replace the placeholder API keys with your own values.

OpenAI Example

curl --location 'http://localhost:8080/v1/responses' \
--header 'Content-Type: application/json' \
--header 'Authorization: Bearer OPENAI_API_KEY' \
--header 'x-model-provider: openai' \
--data '{
    "model": "gpt-4o",
    "stream": false,
    "input": [
        {
            "role": "user",
            "content": "Write a poem on OpenResponses"
        }
    ]
}'

Groq Example (with Streaming)

curl --location 'http://localhost:8080/v1/responses' \
--header 'Content-Type: application/json' \
--header 'Authorization: Bearer GROQ_API_KEY' \
--data '{
    "model": "llama-3.2-3b-preview",
    "stream": true,
    "input": [
        {
            "role": "user",
            "content": "Write a poem on OpenResponses"
        }
    ]
}'

Claude Example

curl --location 'http://localhost:8080/v1/responses' \
--header 'Content-Type: application/json' \
--header 'Authorization: Bearer ANTHROPIC_API_KEY' \
--header 'x-model-provider: claude' \
--data '{
    "model": "claude-3-5-sonnet-20241022",
    "stream": false,
    "input": [
        {
            "role": "user",
            "content": "Write a poem on OpenResponses"
        }
    ]
}'

Using Tools in API Calls

curl --location 'http://localhost:8080/v1/responses' \
--header 'Content-Type: application/json' \
--header 'Authorization: Bearer YOUR_API_KEY' \
--data '{
    "model": "your-model",
    "stream": false,
    "tools": [
        {
            "type": "brave_web_search"
        }
    ],
    "input": [
        {
            "role": "user",
            "content": "What are the latest developments in AI?"
        }
    ]
}'

๐Ÿ”ฎ Coming Soon

We're continuously evolving OpenResponses with powerful new features to elevate your AI applications even further. Stay tuned!

โ“ Frequently Asked Questions

Can I use my existing provider API keys?

Yes! OpenResponses acts as a pass-through to the provider APIs using your own keys.

Is there any performance penalty?

Our benchmarks show minimal overhead compared to direct API calls.

How do I handle errors?

OpenResponses standardizes error responses across providers:

{
  "type": "rate_limit_exceeded",
  "message": "Rate limit exceeded. Please try again in 30 seconds.",
  "param": null,
  "code": "rate_limit"
}

โš™๏ธ Configuration

The application supports the following environment variables:

Variable Description Default
MCP_SERVER_CONFIG_FILE_PATH Path to MCP server configuration -
MASAIC_MAX_TOOL_CALLS Maximum number of allowed tool calls 10
MASAIC_MAX_STREAMING_TIMEOUT Maximum streaming timeout in ms 60000

๐Ÿ“š Documentation

For more detailed information about using OpenResponses API via Completion API, check out our documentation:

๐Ÿ’ป Local Development

Follow these instructions to set up the project locally for development:

Prerequisites

  • Java JDK 21+
  • Gradle (optional, as project includes Gradle Wrapper)
  • Docker (optional, for containerized setup)

Steps

  1. Clone the repository
git clone https://github.com/masaic-ai-platform/open-responses.git
cd open-responses
  1. Build the project

Use the Gradle Wrapper included in the project:

./gradlew build
  1. Configure Environment Variables

Create or update the application.properties file with necessary configuration under src/main/resources:

server.port: 8080

Set any additional configuration required by your project.

  1. Run the server

To start the server in development mode:

./gradlew bootRun

Docker Setup (Optional)

Build and run the application using Docker:

./gradlew build
docker build -t openresponses .
docker run -p 8080:8080 -d openresponses

Testing

Run the tests with:

./gradlew test

โš ๏ธ Production Use

Alpha Release Disclaimer: This project is currently in alpha stage. The API and features are subject to breaking changes as we continue to evolve and improve the platform. While we strive to maintain stability, please be aware that updates may require modifications to your integration code.

๐Ÿค Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

"Alone we can do so little; together we can do so much." โ€” Helen Keller

๐Ÿ“„ License

This project is licensed under the Apache License 2.0 - see the LICENSE file for details.


Made with โค๏ธ by the Masaic AI Team

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for open-responses

Similar Open Source Tools

For similar tasks

No tools available

For similar jobs

No tools available