AI Concept
https://learn.microsoft.com/en-us/azure/azure-functions/functions-bindings-http-webhook-trigger?tabs=python-v2%2Cisolated-process%2Cnodejs-v4%2Cfunctionsv2&pivots=programming-language-csharp
https://learn.microsoft.com/en-us/azure/azure-functions/functions-networking-options?tabs=azure-portal
https://komodor.com/blog/aiops-for-kubernetes-or-kaiops/
https://learn.microsoft.com/en-us/azure/architecture/ai-ml/guide/ai-agent-design-patterns
https://learn.microsoft.com/en-us/azure/architecture/ai-ml/architecture/baseline-azure-ai-foundry-chat
Technologies :
- K8SGPT
- LangChain (+ LangFuse pour monitoring?)
- Ollama (+ il y a des trucs qui permettent d'améliorer les perfs des modèles)
- MCPO // OpenWebUI
- Machine Virtuelle chez Exo bien boostée
Glossary
Large Language Model (LLM)
A Large Language Model (LLM) is the engine behind an AI application such as ChatGPT. In this case, the engine powering ChatGPT is GPT-4 (or GPT-4o, previously), which is the LLM used by the application.
Azure AI Foundry is a service that allows you to choose which Large Language Model (LLM) you want to use.
Model Context Protocol (MCP)
A Model Context Protocol (MCP) is a protocol that standardizes communication between Large Language Models (LLMs) and external systems , such as ITSM tools (like ServiceNow), Kubernetes clusters, and more.
You can use an MCP client , for example, Continue.dev in your IDE (like VS Code) and then configure MCP servers, such as your Kubernetes cluster, to enable your LLM to interact with these systems.
Retrieval-Augmented Generation (RAG)
Technology Stack
LangChain
LangChain is an application framework that helps structure your prompts using PromptTemplate. For example, with an alerting system, when the AI is queried, you can create a template that guides it to follow a consistent debugging structure in its responses :
from langchain import PromptTemplate
prompt = PromptTemplate(
input_variables=["alert", "logs", "metrics"],
template="""
Tu es un expert Kubernetes.
Un incident a été détecté :
{alert}
Voici les logs du pod :
{logs}
Voici ses métriques :
{metrics}
Analyse les causes probables et propose des actions correctives précises.
"""
)
K8sGPT
K8sGPT is an open-source tool that scans Kubernetes clusters, detects issues, and uses a Large Language Model (LLM) such as Azure OpenAI to explain problems and suggest solutions in natural language.
Ollama
Ollama is an open-source tool that lets you download and run large language models (LLMs) like Llama 3 or Mistral locally, allowing you to use AI without relying on the cloud.

