Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

An Introduction to GWDG Services on Generative AI for MPIA Employees

Generative Artificial Intelligence (GenAI) and Large Language Models (LLMs) have rapidly become essential tools in modern academic research. From streamlining literature reviews to assisting in code debugging and data analysis, these tools offer significant potential to enhance productivity. However, the computational demands of running state-of-the-art models often exceed the capabilities of standard workstations, and commercial cloud solutions can raise concerns regarding data privacy and cost.

For researchers at the Max Planck Institute for Astronomy (MPIA), the Gesellschaft für wissenschaftliche Datenverarbeitung mbH Göttingen (GWDG) provides a robust, self-hosted solution that balances performance, ethics, and accessibility.

The Role of GenAI in Academic Research

LLMs function as sophisticated pattern-matching engines, capable of understanding context, generating text, and writing code based on vast training datasets. In an astrophysical context, these capacities can be leveraged for:

While these models are powerful, they require significant GPU resources to run efficiently. Budget aside, Running them locally on a laptop is often infeasible for larger models, while relying on public commercial APIs may expose sensitive research data to third-party training policies.

Advantages of Self-Hosted Institutional Infrastructure

Utilizing a centralized, self-hosted service like GWDG’s offers distinct advantages over local execution or public commercial clouds:

  1. Cost & Accessibility: High-performance GPU hardware is expensive. GWDG provides MPIA employees with free access to powerful compute resources, removing the financial barrier of procuring individual hardware or using cloud credits.

  2. Data Privacy & Ethics: Unlike some commercial providers, institutional services can be configured to prioritize data sovereignty. GWDG’s self-hosted approach ensures that prompts and data are processed within the institution’s controlled environment, mitigating the risk of sensitive research data being used to train public models.

  3. Sustainability: Centralized computing centers are generally more energy-efficient per computation than individual desktop GPUs. By consolidating workloads in a facility optimized for power usage effectiveness (PUE), researchers can reduce the carbon footprint of their computational workflows.

GWDG AI Services

GWDG offers several interfaces for interacting with generative AI models: Chat AI and CoCo AI for text and code, and Image AI for visual content generation.

Chat AI

Chat AI is a web-based interface similar to commercial chatbots. It provides access to various open-source models, allowing users to engage in conversational queries for text generation, coding assistance, and general problem-solving. It is designed for ease of use, requiring no setup beyond authentication via your institutional account.

CoCo AI (Command & Control AI)

CoCo AI is designed for more advanced integration into research workflows. It allows users to interact with AI models directly from the command line or via scripts. This is particularly useful for automating tasks, such as batch processing text analysis or integrating AI assistance into shell scripts and Python pipelines used for data reduction.

Image AI

Image AI is a web-based service for generating and editing visual content using open-weight diffusion models. It supports two primary workflows:

As with the other GWDG AI services, both models are self-hosted — prompts and images are never stored on GWDG systems and are not used for model training. The simple web interface requires only an AcademicCloud login.

Available Models

Chat AI provides access to a broad selection of large language models. All self-hosted models run entirely on GWDG hardware with no data storage — prompts and message contents are never retained. The catalogue is updated regularly as newer models are released; the full, up-to-date list is maintained at the Available Models page. We list below some noteworthy models as of this writing, split into two categories: self-hosted open-weight models and externallyhosted models.

Data privacy diagram

overview of the GWDG service.

Self-Hosted Open-Weight Models (hosted by GWDG)

Example of hosted models available in Chat AI:

ModelProviderParametersContext WindowStrengths
Apertus 70B InstructSwiss AI70B65kFully open-source; multilingual
DeepSeek R1 Distill Llama 70BDeepSeek70B32kGood overall performance; faster than R1
Devstral 2 123BMistral AI123B256kCoding and agentic software-engineering tasks
Gemma 3 27B InstructGoogle27B128kVision support; great general-purpose performance
GLM-4.7Z.ai200kGreat overall performance
InternVL 3.5 30B A3BOpenGVLab30B (3B active)40kVision; lightweight and fast
MedGemma 27B InstructGoogle27B128kVision; specialised medical knowledge
Llama 3.3 70B InstructMeta70B128kSolid all-rounder; creative writing & reasoning
Mistral Large 3 675BMistral AI675B256kStrong general-purpose & vision; multilingual
GPT OSS 120BOpenAI (open-weight)120B128kGreat overall performance; fast inference
Qwen 3 30B A3B InstructAlibaba Cloud30B (3B active)256kGood performance; fast
Qwen 3.5 122B A10BAlibaba Cloud122B (10B active)256kVision; great overall performance
Qwen 3.5 397B A17BAlibaba Cloud397B (17B active)256kVision; largest model in the lineup
Qwen 3 Coder 30B A3BAlibaba Cloud30B (3B active)256kSpecialised for coding tasks
Qwen 3 Omni 30B A3BAlibaba Cloud30B (3B active)256kMultimodal (text, audio, video)
Teuken 7B InstructOpenGPT-X7B128kOptimised for European languages

This list is not exhaustive. For more models, please refer to the official documentation.

External Models (hosted by OpenAI / Microsoft Azure)

OpenAI models are relayed through Microsoft Azure under GDPR terms. Microsoft may store messages for up to 30 days. OpenAI may do more on its side. For maximum data privacy, prefer the self-hosted models above.

ModelContext WindowNotes
Claude Opus 4.61MState-of-the-art reasoning, coding, complex analysis
Claude Sonnet 4.61MBalanced performance; fast responses
GPT-5.4 / 5.4 Mini / 5.4 Nano272kLatest flagship; professional knowledge work, coding, agentic workflows
GPT-5.3 Chat128kMultimodal chat; adaptive chain-of-thought
GPT-5 / 5 Chat / 5 Mini / 5 Nano400kRange of speed-vs-capability trade-offs
o3 / o3-mini200kComplex reasoning (older knowledge cutoff)
GPT-4.1 / 4.1 Mini1MLarge context window; coding & instruction following

Tip for MPIA users: The ChatGPT (OpenAI) models are available free of charge to Max Planck Society employees. All self-hosted models are free for everyone with a GWDG account.

Getting Started

To access these services, MPIA employees should visit the GWDG Chat AI portal and log in using their standard GWDG credentials. We encourage researchers to experiment with these tools to identify how they can best support their specific scientific workflows.