Model as a Service -
Central AI model management
with German proxy

Weißes partimus Logo.

GDPR-compliant access to top models

Your advantages with the MaaS Proxy

Safety, comfort and strategic
Freedom in model selection

With the MaaS Proxy from partimus, you can create a central, secure and efficient basis for the productive use of modern AI models in your organization. The solution not only simplifies the connection of different applications to different language models, but also creates clarity in terms of costs, governance and data protection. Standardized interfaces, full control over the data flow and the ability to integrate your own models create a flexible and future-proof infrastructure – optimized for business use.

Your advantages with partimus

🏛️ German proxy service

GDPR-compliant, uniform access via German data centers.

🇩🇪 Local hosting available

Open source models (e.g. DeepSeek R1, GLM 4.5) can be operated entirely in Germany.

🎯 Central budget management with spending limits

Daily, weekly and monthly limits per team or developer. Transparent euro prices with no hidden costs.

🔧 One API for all models

OpenAI compatible interface: Switch between GPT-4o, Grok 4 or Qwen3 coders without code changes.

⚡ Immediate availability

No waiting lists or additional registration – get started in minutes with prototypes up to production.

🛡️ Enterprise-grade security

ISO 27001-certified infrastructure, end-to-end encryption and central access controls.

Your AI - your model portfolio

Wide range of models - flexible and future-proof

The MaaS Proxy supports a variety of powerful language models – from open source LLMs such as LLaMA, Qwen, DeepSeek or Mistral to models from OpenAI, Google or Anthropic. You decide which models you want to use, prioritize or block – depending on your use case and the requirements of your organization. You can integrate your own models as well as external APIs so that you can flexibly adapt and expand your AI infrastructure at any time. New releases can be managed and deployed centrally – without the need for manual adjustments in the specialist applications.

Premium models (via proxy)

  • GPT-5 (OpenAI) - next-generation flagship, 1 M token context
  • GPT-4o (OpenAI) - Multimodal, 128 K Context
  • Gemini 2.5 Pro (Google) - Thinking Model, 1 M Token Context
  • Grok 4 (xAI) - Real-time Reasoning, 256 K Context

Code-specialized models (via proxy)

  • Qwen3-Coder (Alibaba) - 480 B Parameter MoE for agent programming

Efficient models (proxy & optional local hosting)

  • DeepSeek R1 (DeepSeek) - Cost-effective reasoning model
  • GLM 4.5 (Zhipu AI)) - 355 B Parameter Hybrid Reasoning
  • GPT OSS 120B (OpenAI)) - Open-Source Reasoning
  • Kimi K2 (Moonshot AI)) - 1 T Parameter Agentic Intelligence

Use AI centrally, securely & efficiently now

Test the partimus MaaS Proxy without obligation

Benefit from central model management, GDPR-compliant access to leading LLMs and maximum control over your AI usage – with no technical barriers to entry. With the partimus MaaS Proxy, you can integrate language models such as OpenAI, Mistral or LLaMA into your systems via a standardized interface – securely hosted in Germany.

The technical implementation

Uniform API gateway for all models

The partimus MaaS Proxy is designed so that you can quickly and easily add AI functionality to your existing applications. Thanks to standardized REST API, OpenAI-compatible endpoints and full multi-provider support, the proxy can be integrated into existing tools, backends or user interfaces – without adapting the model API. Self-hosted LLMs can also be seamlessly integrated.

AI Proxy in Germany

All requests run via a central gateway service in German data centers with the highest security and data protection standards. This gives you full control over data flow, security and infrastructure - GDPR-compliant and high-performance.

Multi-provider connection

Use a wide range of AI providers such as OpenAI, Google, xAI, Alibaba and others with a single access point. The proxy takes care of routing and administration - without any adjustments to the respective API.

Optional local hosting

Use open source models such as LLaMA, Mistral or Qwen directly in your own infrastructure. The proxy enables seamless integration of local models via the same interface as with cloud providers.

Central cost management

Keep your spending under control - with automatic budget allocation per team, model or application. Comprehensive analytics and reports help you to analyze and optimize usage transparently.

Code example

import openai

client = openai.OpenAI(
    api_key="<Ihr-partimus-api-key>",
    base_url="https://maas.partimus.com/v1"
)

response = client.chat.completions.create(
    model="gpt-4o",  # oder grok-4, qwen3-coder, gemini-2.5-pro, etc.
    messages=[{"role": "user", "content": "Hallo Welt!"}]
)

print(response.choices[0].message.content)
Transparent billing - usage-based or flat rate

Flexible pricing models for every scenario

Whether you are a startup, research team or corporation – our pricing adapts to your needs. Choose between usage-based billing per token, flat-rate monthly plans for specific providers or individually agreed enterprise models. You only pay for what you actually use – with full cost control and a clear overview.

*All prices are monthly/non-recurring fees plus VAT. Our offer is aimed exclusively at traders.

Model prices

partimus gmbh openai logo
OpenAI
Model Cost per 1M input tokens Cost per 1M output tokens
openai/gpt-5 1,40€ 10,50€
openai/gpt-5-mini 0,30€ 2,10€
openai/gpt-5-nano 0,10€ 0,50€
openai/gpt-oss-20b 0,20€ 0,60€
openai/gpt-oss-120b 0,20€ 0,80€
partimus gmbh moonshot ai logo
MoonShotAI
Model Cost per 1M input tokens Cost per 1M output tokens
moonshotai/kimi-k2-instruct 1,10€ 3,20€
partimus gmbh meta logo
Meta
Model Cost per 1M input tokens Cost per 1M output tokens
meta/llama-4-maverick-17b 0,30€ 0,70€
meta/llama3-70b 0,70€ 0,90€
meta/llama3-8b 0,10€ 0,10€
partimus gmbh logo google
Google
Model Cost per 1M input tokens Cost per 1M output tokens
gemini/gemini-2.5-pro 1,40€ 10,50€
gemini/gemini-2.5-flash 0,40€ 2,70€
partimus gmbh logo
XAI
Model Cost per 1M input tokens Cost per 1M output tokens
xai/grok-4 3,20€ 15,80€
xai/grok-code-fast 0,30€ 1,60€
partimus gmbh qwen logo
QWEN
Model Cost per 1M input tokens Cost per 1M output tokens
qwen/qwen3-coder-480b-a35b-instruct 0,50€ 1,70€
qwen/qwen3-235b-a22b-thinking 0,20€ 0,70€
partimus gmbh logo zai
ZAI
Model Cost per 1M input tokens Cost per 1M output tokens
zai/glm-4.5 0,60€ 2,10€
zai/glm-4.5-air 0,30€ 1,20€
partimus gmbh logo deepseek
DeepSeek
Model Cost per 1M input tokens Cost per 1M output tokens
deepseek/deepseek-v3.1 0,40€ 1,10€
deepseek/deepseek-r1-0528 0,60€ 2,30€
partimus gmbh logo anthropic
Anthropic
Model Cost per 1M input tokens Cost per 1M output tokens
anthropic/claude-opus-4.1 15,80€ 78,60€
anthropic/claude-sonnet-4 6,30€ 15,80€
anthropic/claude-haiku-3.5 0,90€ 4,20€
Well-founded & practical

The partimus AI introductory workshop

Would you like to use modern AI tools safely, efficiently and practically in your day-to-day work? In our workshop, you will not only receive an in-depth overview of current language models and the use of central proxy access, but also get to know tried-and-tested tools such as RooCode and Chatbox. Together, we will implement individual use cases and show you how to integrate AI into your processes smoothly and in compliance with data protection regulations.

In interactive exercises, you will actively work with the tools, create your own chatbots and automated workflows and benefit from concrete best practices for cost and budget control. After just one day, you will be ready to use AI responsibly and productively in your company.

Book your workshop now and get your team ready for the use of AI.

Contents of the workshop

Introduction to modern AI models

Basics of current language models, possible applications and trends.

Practical tool introduction (RooCode & Chatbox)

Presentation of proven tools for collaborative coding and workflow automation.

Interactive exercise phases with specific use cases

Implement your own use cases and experience tools in practice.

Step-by-step instructions for typical application scenarios

Clearly understandable guidelines for implementation in everyday working life.

Error prevention & best practices in the use of AI

Practical tips for safe and efficient use.

Data protection & compliance in everyday life

Legally compliant handling of AI in your company.

Cost and budget control in AI tools

Strategies for cost optimization and resource control.

Your advantages at a glance

Secure your place now!
Venue
Online
Price
1.800€ / participant
GDPR-compliant, auditable and under your control

Maximum safety & full control

The MaaS Proxy was specially developed to meet the highest data protection requirements. All requests are processed via our infrastructure in German data centers – without detours via third countries. Comprehensive logging, audit and control mechanisms allow you to keep track of data flows and access at all times. On request, sensitive models can also be hosted completely locally – for full sovereignty over your AI applications.

German proxy server

with German data centers and logging/monitoring.

Optional hosting in Germany

for maximum data sovereignty.

ISO 27001 certification

for our information security management system.

Central access control

with strict authorization and role management.

partimus gmbh maas proxy 3
Central model use - secure, efficient and flexible

Wide range of applications for your teams

With the MaaS Proxy, you can provide powerful AI models specifically where they create real added value. Whether development, support or marketing – your entire team benefits from centrally controlled, data protection-compliant access to leading models.

Software development

Create clean code automatically, get support with debugging or generate technical documentation - directly from your development environment. Models such as Qwen3 Coder or GPT OSS 120B can be seamlessly integrated via the proxy.

Customer Support

Use advanced AI models such as GPT-4o or Grok 4 to create dynamic, multilingual chatbots. Answers to current questions are generated in real time from linked knowledge sources.

Content marketing

Produce creative, brand-compliant content in seconds - from blog posts to social media and product texts. Gemini 2.5 Pro supports your editorial processes with convincing quality and consistent style.

Get to know partimus

Get started with partimus

Take the opportunity to optimize your IT infrastructure and drive your business forward – contact us today to find out more about the tailored benefits partimus can offer you.

FAQ - Frequently asked questions

Questions and answers about the MaaS proxy

The MaaS Proxy provides you with a central platform for managing all AI access – regardless of the provider. You benefit from standardized authentication, central monitoring, cost control, GDPR compliance and the option to also securely integrate local models.

Both are possible. The proxy supports both common provider models and the local operation of your own open source models – e.g. on an inference solution from partimus or on your own infrastructure.

The entire proxy service is operated in German data centers and is ISO 27001-certified. You retain full control over data flows, usage rights and storage locations – including logging and role-based access.

Yes, the flexible cost control, simple integration into existing systems and central model management in particular make the proxy suitable for smaller organizational units. This allows you to scale without having to set up a complex infrastructure.

You receive a monthly consolidated invoice for all requests – regardless of whether they were sent to OpenAI, Google, Mistral or a local model. Budgets and limits can be flexibly defined per team, project or user.

Are you interested in our MaaS Proxy?
We will be happy to advise you!
Anmeldung zum KI-Einführungs­workshop
Ihre Anfrage wurde abgeschickt!
Registration for the AI introductory workshop
Your request has been sent!