Exploring AI Prompt Engineering: Mathematical Foundations and RAG Methodologies
Last Updated on June 3, 2024 by Editorial Team
Author(s): Shenggang Li
Originally published on Towards AI.
Photo by ZHENYU LUO on Unsplash
We first represent a Large Language Model (LLM) with the formula:
It looks simple, However, if we want to understand its application and the AI techniques on LLMs, we need to interpret π, π, and π as specific items within the LLM framework.
LLM Weights (Ο)
The parameter set π is neural network weights (model coefficients) and biases, updated during model training. π can affect LLMs responses, but remain fixed unless fine-tuning is performed. The parameters can be expressed as a vector:
Where π is the number of parameters, which can be in the billions for modern LLMs. However, these parameters are not the focus of this paper, as we aim at AI prompt engineering and RAG methodologies.
Model Input (X):
The input π includes several items working together to generate a prediction π:
Β· X_query: A specific question or request from users.
Β· X_prompt: The initial prompt set by the developer.
Β· X_RAG Prompts: Additional prompts (documents), retrieved from data source D based on the X_query.
Β· X_parameters: Parameters like temperature, max tokens, and stream options
Where X_parameters are not model training parameters (e.g., learning rate, number of neural network layers), but they play a similar role in affecting the LLM behavior during inference If… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI