Skip to content

Custom GPT

Across Industries

For companies like BMW, Allianz, and Sennebogen, Motius has developed an approach to building AI assistants and agents that can work both on-premise and in the cloud, with data ranging from classified internal documents to public data ingested through web search.

A consolidated data layer ensures governance and reusability, while a flexible platform enables the development of use case-specific AI agents, including third-party solutions. Users can interact via chat app, custom web interface, or IT system integrations.

AI Tools

The platform approach provides enterprise grade scalability and security while still providing flexibility, in concrete technical implementation. The solution will empower Keller & Kalmbach with AI-driven insights and automation to enhance workflows and decision-making.

Typical Challenges for other Customers
  • There is a lot of unstructured knowledge in the company, such as documents, Wikis, emails, and chat messages
  • Existing systems have limited search capabilities. Usually there is no global search
  • Incomplete datasets spread over multiple systems, such as ERP, MES, CRM, and file storage
  • Knowledge distributed across multiple locations, in different languages

Demo

Gandalv is an internal product we built to showcase our approach, and help us during the requirements engineering phase of our projects.

Key Features

While the demo shows a specific use case, the platform is designed to be flexible and extensible:

  • On-Premise or Cloud


    Our approach works on-premise or using a cloud-based AI model in Azure or AWS. We use open-source components to stay flexible and avoid vendor lock-in, while using the state of the art in AI and LLMs.

  • AI-Powered Assistant


    The AI assistant is capable of understanding and processing natural language queries, providing intelligent responses and insights based on the data it has access to. It seamlessly works with different input languages in Keller & Kalmbach's data.

  • Data Integration


    Using MCP and RAG we ensure that the AI model has access to the most relevant data, without spending months preparing and cleaning it.

  • Speed & Quick Iterations


    With powerful open-source components and state of te art models, Motius only needs a few weeks to set up the - infrastructure and first use cases for Keller & Kalmbach. From there, we can quickly iterate on data sources and new use cases.

Approach

Knowledge Structuring and Extraction

A team from Motius first analyzes existing content and develops a strategy for providing a structured knowledge base. The strategy may include the following technologies and approaches:

Technology Data Source Description
RAG Wikis, internal documentation, operating manuals AI-powered search and extraction of relevant information from various sources. Very efficient for searching, but requires dedicated infrastructure
MCP Confluence, CAD systems, ERP, MES Standardized interface for querying information. Ad hoc, requires very little infrastructure, but is slower than RAG

Rapid Prototyping and Integration

Once the data sources have been reviewed and connected, a prototype is developed usually using on-premise LLMs and open source components.

Employees at Keller & Kalmbach can start working with the system and provide feedback after just 1-2 sprints (each 2 weeks).

Note

If possible, we deploy this system to a cloud environment, such as Azure or AWS, to ensure scalability and performance.

Frequently, internal security policies require that the system runs on-premise, in which case we use open source LLMs and components.

Data Enrichment and Automation

After the first tests, we add more data sources and build agents that can automate tasks and provide insights for different business processes, such as:

  • Accessing knowledge across many systems and regions, asking follow-up questions, and getting relevant documents as sources in the response
  • Automating repetitive tasks, such as generating reports, summarizing documents, or extracting key information
  • Tool calling with MCP, which allows the AI assistant to act on behalf of the user, but using their credentials and permissions

Architecture

graph TB
    subgraph SS ["Source Systems"]
        ERP[(ERP)]
        CRM[(CRM)]
        EXT[(External Data)]
        FILES[Files]
        OTHER[(...)]
    end

    subgraph DP ["Data Platform"]
        subgraph STORAGE ["Storage"]
            RAW[(Raw)]
            CLEANSED[(Cleansed)]
            ENRICHED[(Enriched)]
            CURATED[(Curated)]
        end

        subgraph SPECIALIZED ["Specialized Storage"]
            VECTOR[(Vector DB)]
            GRAPH[(Graph DB)]
            DWH[(Data Warehouse)]
        end

        subgraph COMPUTE ["Compute"]
            INGEST[Ingest]
            PARSE[Parse/Chunk/Filter]
            EMBED[LLM Model Embedding]
        end
    end

    subgraph AL ["Application Layer"]
        subgraph CUSTOMER ["Keller & Kalmbach GPT Service"]
            SIMILARITY[Similarity Matching]
            LLM_GEN[LLM Model Generative]
            LLM_EMB[LLM Model Embedding]
            INFO_QUERY[Information Querying]
            CHAT_HIST[(Chat History)]
        end

        DASHBOARDS[Reporting Dashboards]
        ERP_APP[ERP]
        CRM_APP[CRM]
        CUSTOMER_APP[Keller & Kalmbach GPT App]
    end

    %% Data Flow
    ERP --> RAW
    CRM --> RAW
    OTHER --> RAW
    EXT --> RAW
    FILES --> RAW

    RAW --> CLEANSED
    CLEANSED --> ENRICHED
    ENRICHED --> CURATED

    CURATED --> VECTOR
    CURATED --> GRAPH
    CURATED --> DWH

    %% Compute Flow
    RAW --> INGEST
    INGEST --> PARSE
    PARSE --> EMBED
    EMBED --> VECTOR

    %% Application Layer Connections
    VECTOR --> SIMILARITY
    SIMILARITY --> LLM_EMB
    LLM_EMB --> INFO_QUERY
    INFO_QUERY --> CHAT_HIST
    LLM_GEN --> INFO_QUERY

    DWH --> DASHBOARDS
    DASHBOARDS --> CUSTOMER
    ERP_APP --> CUSTOMER
    CRM_APP --> CUSTOMER
    CUSTOMER_APP --> CUSTOMER

    %% Styling
    classDef sourceSystem stroke:#0288d1,stroke-width: 2px
    classDef dataStorage stroke:#7b1fa2,stroke-width: 2px
    classDef compute stroke:#ef6c00,stroke-width: 2px
    classDef application stroke:#388e3c,stroke-width: 2px
    classDef specialized stroke:#c2185b,stroke-width: 2px

    class ERP,CRM,OTHER,EXT,FILES sourceSystem
    class RAW,CLEANSED,ENRICHED,CURATED dataStorage
    class INGEST,PARSE,EMBED compute
    class SIMILARITY,LLM_GEN,LLM_EMB,INFO_QUERY,DASHBOARDS,ERP_APP,CRM_APP,CUSTOMER_APP application
    class VECTOR,GRAPH,DWH,CHAT_HIST specialized
Press "Alt" / "Option" to enable Pan & Zoom

We deploy a mix of open source components and open-weight LLM models:

  • LibreChat is an extensible chat interface with built-in support for all common LLMs, including open source models like Meta's Llama or Google DeepMind's Gemma
  • PostgreSQL with pgvector for vector storage and as a simple data warehouse
  • MCP for tool calling and integration with existing systems

Azure's OpenAI services provide a powerful platform for building AI assistants and agents:

  • Azure OpenAI Service for scalable LLM hosting and integration
  • Azure AI Foundry portal and the ingestion API for data integration and management
  • MCP for tool calling and integration with existing systems

AWS Bedrock is the most complete and flexible platform for building AI assistants and agents, out of all the public cloud providers:

  • Amazon Bedrock for LLM hosting and orchestration, as well as RAG
  • MCP for tool calling and integration with existing systems

Application at Keller & Kalmbach

Keller & Kalmbach Logo

With this approach, Keller & Kalmbach can start building AI assistants for specific use cases within weeks, while building on an architecture that allows for future growth:

  • Cloud and AI-agnostic architecture means Keller & Kalmbach can choose the best providers with no lock-in
  • Immediate benefits of AI assistants for specific use cases, such as customer support, internal knowledge management, and process automation
  • Go further with AI agents that can automate tasks, and even let your engineering teams build their own agents