The global and independent platform for the SAP community.

Small but fine AI

RPT-1 is SAP's technical attempt to transfer Transformer technology from language processing to SAP's core competence - relational business data - in order to eliminate the weaknesses of conventional LLMs.
E3 Magazine
March 25, 2026
avatar
This text has been automatically translated from German to English.

The decisive advantage of SAP RPT-1 over the popular LLMs from OpenAI or Anthropic lies in the fundamental way in which data is processed. Conventional language models are brilliant rhetoricians, but often fail miserably at simple arithmetic or the precise interpretation of series of numbers. When an LLM sees the number „1000“, it is often broken down into individual tokens, such as „1“, „0“, „0“ and „0“, and treated as a sequence of characters rather than a mathematical value, which can lead to the infamous hallucinations during calculations.

RPT-1 and relational values

RPT-1, on the other hand, recognizes „1000“ as a single relational value, which massively increases precision when analyzing business data such as that found in SAP's MARA tables or sales databases. Where a text-based model hallucinates or misunderstands the context of a complex table structure, RPT-1 acts like a specialized „GPT for tables“ that natively understands the semantic logic of rows and columns without having to take a detour via language. In the heated debate about artificial intelligence, which is usually dominated by the linguistically adept all-rounders from Silicon Valley such as ChatGPT or Claude, SAP is trying to make up ground with a technical niche solution that goes by the cryptic abbreviation RPT-1. While the world gazes in amazement at large language models (LLMs) that write poetry and generate code, SAP is returning to its true DNA: the mundane but business-critical database tables.

RPT-1, which stands for „Relational Pre-trained Transformer“, is an attempt to transfer the Transformer technology that founded the success of OpenAI and Anthropic from the unstructured world of language to the highly structured world of relational enterprise data. It is SAP's first real „foundation model“ that was developed specifically for tables and not for text, and thus marks a strategically necessary counter-design to the language-based models of the US competition. Another major advantage lies in its efficiency and the elimination of time-consuming training phases for specific tasks. In the traditional world of machine learning (ML), companies had to train and maintain a separate, dedicated model for each problem - be it predicting customer churn, fraud detection or demand planning - which tied up enormous resources in the data science teams.

In-Context Learning

RPT-1 breaks with this paradigm through so-called in-context learning. The model does not have to memorize the customer-specific data or undergo lengthy training; instead, it reads the data during inference time, makes the prediction and „forgets“ the data immediately, similar to an exam where the textbook is allowed to lie open on the table. This not only drastically lowers the hurdle for deployment, but also significantly reduces resource consumption compared to the token-hungry LLMs of hyperscalers, as no huge amounts of text need to be processed.
The possible applications of this model are aimed right at the heart of ERP processes. SAP is positioning RPT-1 for scenarios in which speed and precision are more important than eloquence. A small version of the model, the so-called Speedster, is optimized for millisecond latency times and is designed to block fraudulent transactions in real time, for example. Larger versions of the model address complex tasks such as predicting supply chain bottlenecks or payment risks. Tasks that conventional ML pipelines often take hours to calculate, while RPT-1 delivers results almost instantly. The model also offers new approaches to IT operations management for the technical monitoring of the SAP landscape itself - for example, to predict whether batch jobs or interfaces will fail at the end of the month.

RPT-1 and LLMs

Based on the available sources (https://community.sap.com/), the technical difference between SAP RPT-1 (Relational Pre-trained Transformer) and the LLM approaches of OpenAI (such as GPT) can be fundamentally determined by the type of data processing and the architecture, see box. While OpenAI models are language models for unstructured text data, RPT-1 is a specialized foundation model for structured, tabular business data.

However, despite all the technical finesse, a critical classification is essential. In the SAP community, RPT-1 is sometimes viewed skeptically as part of an „AI chaos theory“. Critics note that RPT-1 is in fact a „narrow-gauge“ LLM that may simply be selling old wine in new bottles. There is a suspicion that mechanisms such as the „Predictive Analysis Library“ (PAL) of the Hana database are still being used under the hood.
work, supplemented by modern marketing terms so as not to lose touch with the AI hype.

Abap tables „only“

Furthermore, as innovative as the approach to tables may be, RPT-1 is limited in its range of applications: It works exclusively on tables, not text or images, and is therefore not a replacement, but at most a supplement to the powerful models of OpenAI. While SAP is trying to regain sovereignty over „business semantics“ with RPT-1, it remains to be seen whether this specialized approach will be enough to stand up to the sheer superiority and speed of innovation of the generative AI giants, which are increasingly learning to handle structured data better. RPT-1 is therefore SAP's bet that, in the ERP environment, the bare figures will ultimately count more than fine words. (pmf)

Because SAP RPT-1 is optimized to „understand“ Abap tables, the requirements and the AI architecture are not really comparable with classic LLMs. (Source: SAP)

SAP RPT-1 versus ChatGPT

Data modality: table vs. text. With OpenAI (LLM), the models are trained to understand and generate natural language. They work with unstructured text. SAP RPT-1 was developed specifically for tabular data. It acts as a „GPT for tables“. It understands the structure of rows and columns (as in SAP tables MARA or sales data) natively instead of interpreting them as text.

Tokenization and numerical understanding (the „1000“ problem): A technical difference lies in the treatment of numerical values. When OpenAI (LLM) sees the number „1000“, it is often tokenized and broken down into separate characters. The model then tries to perform math based on these text characters, which often leads to hallucinations or calculation errors. RPT-1 recognizes „1000“ as a single relational value. It understands the mathematical meaning and context of the number within the table structure, allowing for more accurate predictions on business data.

Learning methods: In-context learning vs. fine-tuning. In order to solve specific business tasks, LLMs from OpenAI often have to be fine-tuned with new data or require complex RAG (Retrieval Augmented Generation) architectures to obtain context. SAP RPT-1 uses in-context learning. The model does not have to be used for every task.
(e.g. churn prediction or fraud detection) can be retrained. It reads the data at runtime (inference time), makes the prediction and then „forgets“ the data again.

Resources and efficiency: OpenAI (LLM) is resource-hungry, consumes many tokens and requires massive computing power (GPUs) to process context. SAP RPT-1 is optimized for the tabular structure, so it consumes significantly fewer resources (tokens) and performs fewer computational operations (FLOPs). It is more efficient for tasks such as filling missing values in a database or predicting time series.

Write a comment

Working on the SAP basis is crucial for successful S/4 conversion. 

This gives the Competence Center strategic importance for existing SAP customers. Regardless of the S/4 Hana operating model, topics such as Automation, Monitoring, Security, Application Lifecycle Management and Data Management the basis for S/4 operations.

For the fourth time, E3 magazine is organizing a summit for the SAP community in Salzburg to provide comprehensive information on all aspects of S/4 Hana groundwork.

Venue

FourSide Hotel Salzburg,
Trademark Collection by Wyndham
Am Messezentrum 2, 5020 Salzburg, Austria
+43-662-4355460

Event date

Wednesday, June 10, and
Thursday, June 11, 2026

AI experience workshop only on June 11, 2026 (limited places)
Bonus: Access to all lectures on June 11, 2026

Regular ticket

Lectures, evening event and, depending on availability, the AI workshop on June 11, 2026
Places at the AI experience workshop are limited and registration is required.

Subscribers to the E3 Magazine Ticket

reduced with promocode CCAbo26

Students*

reduced with promocode CCStud26.
Please send proof of studies by e-mail to office@b4bmedia.net.
*The first 10 tickets are free of charge for students. Try your luck! 🍀
EUR 305 excl. VAT.
EUR 590 excl. VAT
EUR 390 excl. VAT
EUR 290 excl. VAT

Venue

Hotel Hilton Heidelberg
Kurfürstenanlage 1
D-69115 Heidelberg

Event date

Wednesday, April 22 and
Thursday, April 23, 2026

Tickets

AI onlyExperience workshop on April 23, 2026 
Bonus: Access to all lectures on April 23, 2026
Regular ticket
April 22, 2026: Lectures and evening event
April 23, 2026: Lectures and AI workshop
EUR 305 excl. VAT
EUR 590 excl. VAT
Subscribers to the E3 magazine
reduced with promocode STAbo26
EUR 390 excl. VAT
Students*
reduced with promocode STStud26.
Please send proof of studies by e-mail to office@b4bmedia.net.
EUR 290 excl. VAT
*The first 10 tickets are free of charge for students. Try your luck! 🍀
The event is organized by the E3 magazine of the publishing house B4Bmedia.net AG. The presentations will be accompanied by an exhibition of selected SAP partners. The ticket price includes attendance at all presentations of the Steampunk and BTP Summit 2026, a visit to the exhibition area, participation in the evening event and catering during the official program. The lecture program and the list of exhibitors and sponsors (SAP partners) will be published on this website in due course.