Model as a Service -
Central AI model management
with German proxy
GDPR-compliant access to top models
The use of powerful AI models such as GPT-4o, Gemini 2.5 or Grok 4 presents companies with new challenges: Data protection, budget control, technical access and model diversity. With the MaaS Proxy from partimus, you can access leading models via a central API – secured by German infrastructure, flexible budget management and optional hosting for open source solutions. Develop, scale and integrate AI applications with maximum security and control – without waiting lists, hidden costs or compliance risks.
- GDPR-compliant hosted in Germany
- Central budget control & API access
- Ready to go immediately - no waiting time
Safety, comfort and strategic
Freedom in model selection
With the MaaS Proxy from partimus, you can create a central, secure and efficient basis for the productive use of modern AI models in your organization. The solution not only simplifies the connection of different applications to different language models, but also creates clarity in terms of costs, governance and data protection. Standardized interfaces, full control over the data flow and the ability to integrate your own models create a flexible and future-proof infrastructure – optimized for business use.
🏛️ German proxy service
GDPR-compliant, uniform access via German data centers.
🇩🇪 Local hosting available
Open source models (e.g. DeepSeek R1, GLM 4.5) can be operated entirely in Germany.
🎯 Central budget management with spending limits
Daily, weekly and monthly limits per team or developer. Transparent euro prices with no hidden costs.
🔧 One API for all models
OpenAI compatible interface: Switch between GPT-4o, Grok 4 or Qwen3 coders without code changes.
⚡ Immediate availability
No waiting lists or additional registration – get started in minutes with prototypes up to production.
🛡️ Enterprise-grade security
ISO 27001-certified infrastructure, end-to-end encryption and central access controls.
Wide range of models - flexible and future-proof
The MaaS Proxy supports a variety of powerful language models – from open source LLMs such as LLaMA, Qwen, DeepSeek or Mistral to models from OpenAI, Google or Anthropic. You decide which models you want to use, prioritize or block – depending on your use case and the requirements of your organization. You can integrate your own models as well as external APIs so that you can flexibly adapt and expand your AI infrastructure at any time. New releases can be managed and deployed centrally – without the need for manual adjustments in the specialist applications.
- GPT-5 (OpenAI) - next-generation flagship, 1 M token context
- GPT-4o (OpenAI) - Multimodal, 128 K Context
- Gemini 2.5 Pro (Google) - Thinking Model, 1 M Token Context
- Grok 4 (xAI) - Real-time Reasoning, 256 K Context
- Qwen3-Coder (Alibaba) - 480 B Parameter MoE for agent programming
- DeepSeek R1 (DeepSeek) - Cost-effective reasoning model
- GLM 4.5 (Zhipu AI)) - 355 B Parameter Hybrid Reasoning
- GPT OSS 120B (OpenAI)) - Open-Source Reasoning
- Kimi K2 (Moonshot AI)) - 1 T Parameter Agentic Intelligence
Test the partimus MaaS Proxy without obligation
Benefit from central model management, GDPR-compliant access to leading LLMs and maximum control over your AI usage – with no technical barriers to entry. With the partimus MaaS Proxy, you can integrate language models such as OpenAI, Mistral or LLaMA into your systems via a standardized interface – securely hosted in Germany.
Uniform API gateway for all models
The partimus MaaS Proxy is designed so that you can quickly and easily add AI functionality to your existing applications. Thanks to standardized REST API, OpenAI-compatible endpoints and full multi-provider support, the proxy can be integrated into existing tools, backends or user interfaces – without adapting the model API. Self-hosted LLMs can also be seamlessly integrated.
All requests run via a central gateway service in German data centers with the highest security and data protection standards. This gives you full control over data flow, security and infrastructure - GDPR-compliant and high-performance.
Use a wide range of AI providers such as OpenAI, Google, xAI, Alibaba and others with a single access point. The proxy takes care of routing and administration - without any adjustments to the respective API.
Use open source models such as LLaMA, Mistral or Qwen directly in your own infrastructure. The proxy enables seamless integration of local models via the same interface as with cloud providers.
Keep your spending under control - with automatic budget allocation per team, model or application. Comprehensive analytics and reports help you to analyze and optimize usage transparently.
Code example
import openai
client = openai.OpenAI(
api_key="<Ihr-partimus-api-key>",
base_url="https://maas.partimus.com/v1"
)
response = client.chat.completions.create(
model="gpt-4o", # oder grok-4, qwen3-coder, gemini-2.5-pro, etc.
messages=[{"role": "user", "content": "Hallo Welt!"}]
)
print(response.choices[0].message.content)
Flexible pricing models for every scenario
Whether you are a startup, research team or corporation – our pricing adapts to your needs. Choose between usage-based billing per token, flat-rate monthly plans for specific providers or individually agreed enterprise models. You only pay for what you actually use – with full cost control and a clear overview.
- Developer accounts
- AI credit
- Proxy access
- Analytics
- Team management
- Support
- Custom SLAs
- Local hosting
- Dedicated Infrastructure
- Price
Recommended!
25
200€
Advanced with reports
Priority Support
199 €/month*
Unlimited
Individual
Advanced with reports
24/7 Premium Support
Individual
*All prices are monthly/non-recurring fees plus VAT. Our offer is aimed exclusively at traders.
Model prices

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
openai/gpt-5 | 1,40€ | 10,50€ |
openai/gpt-5-mini | 0,30€ | 2,10€ |
openai/gpt-5-nano | 0,10€ | 0,50€ |
openai/gpt-oss-20b | 0,20€ | 0,60€ |
openai/gpt-oss-120b | 0,20€ | 0,80€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
moonshotai/kimi-k2-instruct | 1,10€ | 3,20€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
meta/llama-4-maverick-17b | 0,30€ | 0,70€ |
meta/llama3-70b | 0,70€ | 0,90€ |
meta/llama3-8b | 0,10€ | 0,10€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
gemini/gemini-2.5-pro | 1,40€ | 10,50€ |
gemini/gemini-2.5-flash | 0,40€ | 2,70€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
xai/grok-4 | 3,20€ | 15,80€ |
xai/grok-code-fast | 0,30€ | 1,60€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
qwen/qwen3-coder-480b-a35b-instruct | 0,50€ | 1,70€ |
qwen/qwen3-235b-a22b-thinking | 0,20€ | 0,70€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
zai/glm-4.5 | 0,60€ | 2,10€ |
zai/glm-4.5-air | 0,30€ | 1,20€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
deepseek/deepseek-v3.1 | 0,40€ | 1,10€ |
deepseek/deepseek-r1-0528 | 0,60€ | 2,30€ |

Model | Cost per 1M input tokens | Cost per 1M output tokens |
---|---|---|
anthropic/claude-opus-4.1 | 15,80€ | 78,60€ |
anthropic/claude-sonnet-4 | 6,30€ | 15,80€ |
anthropic/claude-haiku-3.5 | 0,90€ | 4,20€ |
The partimus AI introductory workshop
Would you like to use modern AI tools safely, efficiently and practically in your day-to-day work? In our workshop, you will not only receive an in-depth overview of current language models and the use of central proxy access, but also get to know tried-and-tested tools such as RooCode and Chatbox. Together, we will implement individual use cases and show you how to integrate AI into your processes smoothly and in compliance with data protection regulations.
In interactive exercises, you will actively work with the tools, create your own chatbots and automated workflows and benefit from concrete best practices for cost and budget control. After just one day, you will be ready to use AI responsibly and productively in your company.
Book your workshop now and get your team ready for the use of AI.
Contents of the workshop
Basics of current language models, possible applications and trends.
Presentation of proven tools for collaborative coding and workflow automation.
Implement your own use cases and experience tools in practice.
Clearly understandable guidelines for implementation in everyday working life.
Practical tips for safe and efficient use.
Legally compliant handling of AI in your company.
Strategies for cost optimization and resource control.
Your advantages at a glance
- Practical online webinar (1 day) for developers and users
- Introduction to modern AI models and enterprise applications
- Step-by-step guide to using the AI proxy and budget management
- Training in best practices for production, security and compliance
- Active tool demos: try out using RooCode and Chatbox live
- Independent exercises & individual questions directly in the workshop
- Handout with all workflows & checklists for direct use in your team
- Certificate of participation for all participants
Maximum safety & full control
The MaaS Proxy was specially developed to meet the highest data protection requirements. All requests are processed via our infrastructure in German data centers – without detours via third countries. Comprehensive logging, audit and control mechanisms allow you to keep track of data flows and access at all times. On request, sensitive models can also be hosted completely locally – for full sovereignty over your AI applications.
with German data centers and logging/monitoring.
for maximum data sovereignty.
for our information security management system.
with strict authorization and role management.

Wide range of applications for your teams
With the MaaS Proxy, you can provide powerful AI models specifically where they create real added value. Whether development, support or marketing – your entire team benefits from centrally controlled, data protection-compliant access to leading models.
Create clean code automatically, get support with debugging or generate technical documentation - directly from your development environment. Models such as Qwen3 Coder or GPT OSS 120B can be seamlessly integrated via the proxy.
Use advanced AI models such as GPT-4o or Grok 4 to create dynamic, multilingual chatbots. Answers to current questions are generated in real time from linked knowledge sources.
Produce creative, brand-compliant content in seconds - from blog posts to social media and product texts. Gemini 2.5 Pro supports your editorial processes with convincing quality and consistent style.
Analyze large amounts of data quickly and efficiently with models such as DeepSeek R1 or GLM 4.5 - without the need for an expensive GPU infrastructure. This gives you well-founded insights for better decisions in your day-to-day business.
Get started with partimus
Take the opportunity to optimize your IT infrastructure and drive your business forward – contact us today to find out more about the tailored benefits partimus can offer you.
Questions and answers about the MaaS proxy
Are you wondering how the MaaS proxy fits into your infrastructure, which models you can integrate or what the data protection situation is? Our FAQ will give you a quick overview of the most important topics. If you have individual questions, we will be happy to answer them personally by e-mail or telephone.
What advantages does the MaaS Proxy offer compared to direct API use with OpenAI, Google & Co.
The MaaS Proxy provides you with a central platform for managing all AI access – regardless of the provider. You benefit from standardized authentication, central monitoring, cost control, GDPR compliance and the option to also securely integrate local models.
Can we integrate our own open source models or rely exclusively on cloud providers?
Both are possible. The proxy supports both common provider models and the local operation of your own open source models – e.g. on an inference solution from partimus or on your own infrastructure.
How does the proxy help to meet data protection and compliance requirements?
The entire proxy service is operated in German data centers and is ISO 27001-certified. You retain full control over data flows, usage rights and storage locations – including logging and role-based access.
Does the use of the proxy also make sense for smaller teams or individual departments?
Yes, the flexible cost control, simple integration into existing systems and central model management in particular make the proxy suitable for smaller organizational units. This allows you to scale without having to set up a complex infrastructure.
How does billing work when using multiple models and providers?
You receive a monthly consolidated invoice for all requests – regardless of whether they were sent to OpenAI, Google, Mistral or a local model. Budgets and limits can be flexibly defined per team, project or user.
We will be happy to advise you!