Microsoft Certified Azure AI Engineer Associate

Generative Business AI creates new valuable content for SAP Business Processes on cloud platforms like SAP Business Technology Platform SAP BTP, AWS or Azure.

Empower SAP On BTP Cloud With Latest GenAI LLMs 🚀

Generative AI harnesses Foundation Models (FM) and Large Language Models (LLM) to transform single or multi modal inputs to new output content.

SAP Azure BTP Generative AI

Foundation models are trained for general purposes and can be integrated into versatile business scenarios, in contrast to LLMs which are often specialized for specific Business AI use-cases. These Generative AI models can be integrated into SAP Business Processes with the SAP Generative AI Hub which is built on top of SAP AI Core.

SAP Generative Business AI

Generative Business AI models perform downstream tasks in SAP business scenarios. AI foundation models process multi-modal prompts as input to generate generate new business content with various formats like text, audio, speech, code or image.

Based on Generative AI Transformer Architectures, foundation models are trained with transfer and self-supervised learning. Self-supervised training createa labels based on data structures which can be applied to versatile downstream tasks.

Customization and orchestration empowers GenAI models with SAP Business AI capabilities like domain specific context, prompt management and enhanced security features.

Embeddings Vector Tokenization

Input features and output predictions of Genenerative AI models are represented as vectors. As part of Generative AI data preparation, tokenization splits natural language input into small text pieces and converts these pieces into vectors.

SAP Azure Cloud Generative & Business AI NLP Vector Embeddings

Byte-Pair Encoding (BPE) is widely used to implement Generative AI tokenizers like for instance tiktoken for OpenAI models.

Vectorization converts text tokens into Embeddings which are numerical vector representations of features like texts or images optimized for machine learning processing.

SAP Azure Cloud Generative & Business AI NLP Vector Cosine Vector Similarity

Vector representations and their directions are used for similarity calculations with Euclidian distance, dot product or cosine formulas. Use cases in NLP analysis are comparisons for text mining, sentiment analysis, document clustering or similarity search.

Large Language Models (LLM)

Foundation models are general-purpose Large Language Models trained on large text datasets to perform different downstream tasks. The reusability of pre-trained foundation models reduces costly training effort and enables use-cases with small training datasets or machine learning teams.

NLP Business AI scenarios implemented with multi-modal Large Language Models (LLMs) can process language input and generate text, code or image output. Typical Generative AI NLP use-cases are classification, sentiment determination, summarization, comparison or text generation for application areas like autonomous AI assistants or different kind of document processing.

GenAI Foundation Model Comparison

Business AI Cloud Architects can compare Generative AI LLMs on cloud platforms like AWS Bedrock, Azure AI Foundry or SAP GenAI Hub to find AI capabilities which fit best to their specific Business AI scenarios.

Some important selection criteria of GenAI foundation models are parameter count, model size, benchmarks, costs of licenses, context window size or latency.

Some important GenAI model benchmarks, which are evaluated with prepared datasets and questions, are listed below.

Name Description
MMLU general mostly knowledge-driven evaluation of various tasks
MMLU-PRO enhanced, more reasoning-focused multi-choice options
GPQA evaluates expert-level reasoning
HumanEval measures the functional correctness of generated code and programs
MATH / MATH-500 / GSM-8K measures arithmetic reasoning from basic arithmetic to advanced mathematics
BFCL / NFCL evaluate the ability to generate arguments to call functions of external tools
MTEB evaluates the performance of embedding models

SAP Generative AI Hub

The SAP Generative AI Hub is based on SAP AI Core to integrate GenAI deep learning models into SAP Business applications. LLM foundation models are offered with toolsets to engineer LLM prompts and integrate GenAI into business processes.

SAP manages Open Source from MistralAI or Meta and GenAI models hosted on other cloud platforms like AWS Bedrock or Azure OpenAI as SAP AI Core scenarios.

SAP Generative AI Hub offers foundation-models and orchestration scenarios with predefined capabilities to implement GenAI Business AI scenarios. Foundation-models scenarios like azure-openai, aws-bedrock, gcp-vertexai or aicore-opensource group pre-built AI models with versions which can be parameterized in configurations. GenAI configurations can be deployed as instances of serving templates and executables to proxy models hosted on cloud platforms like AWS, Azure or Google.

SAP BTP AI Core Concepts Generative AI Hub LLM

SAP Generative AI Hub Embedding models:

Model Context Dim Costs MTEB
AWS Bedrock
Titan Embed Text 8k 256-1024 0.0004 66
Azure OpenAI
Embedding 3 Small 8191 tokens 1536 0.00006 62.3
Embedding 3 Large 8191 tokens 3072 0.0003 64.6

GenAI Transformer Models

State of the art (SOTA) GenAI models mostly implement transformer architectures with neural networks to remember long-range dependencies and to process whole sentences with positional embeddings and self attention.

SAP Cloud Generative & Business AI Transformer Encoder Decoder

Transformer architectures are separated into encoder and decoder parts both with attention and feed forward components.

Encoders transform language tokens into coordinates within multidimensional vector spaces of semantic language models. The distance of tokens within these embedding models represent their semantic relationship. Embeddings are used for NLP analysis tasks like summarization, key phrase extraction, sentiment analysis with confidence scores or translation.

Decoder are able to generate new text sequences and enable decoder only Conversational or Generative AI solutions.

Attention layer weights of encoder or decoder components control the choice of prediction results. Encoder attention layer weights try to quantify the meaning of words within text sequences. Decoder attention layers predict the most probable output token in a sequence.

Interactive Assistants

Interaction with Generative AI assistants like Microsoft Copilot or SAP Joule improve digital user experience based on data-driven decisions supported by foundation models or LLMs. Microsoft Copilot combines ChatGPT LLMs with data provided by Microsoft Graph and offers the option to build custom copilots for various business-specific tasks.

Open Source GenAI

The Open Source community offers Large Language Models (LLMs), inference servers and GenAI tools to implement intelligent AI solutions.

Advantages of open-source GenAI are transparency, cost reduction, local solutions to fulfill advanced data security requirements or flexible model customization opens like fine-tuning. Examples of open source GenAI model families are LLama, Mistral or Falcon.

Bring Your Own Model (BYOM)

GenAI platforms like SAP Generative AI Hub or AWS Bedrock support training and serving Bring Your Own Models (BYOM) as containerized web applications. GenAI model inference server definitions use serving templates with parameters like number of replicas. Generative AI models can be called via inference request endpoints to return predictions for consuming customer services.