« AI Concept » : différence entre les versions

De Marijan Stajic | Wiki
Aller à la navigation Aller à la recherche
mAucun résumé des modifications
Aucun résumé des modifications
Ligne 13 : Ligne 13 :
* Ollama (+ il y a des trucs qui permettent d'améliorer les perfs des modèles)
* Ollama (+ il y a des trucs qui permettent d'améliorer les perfs des modèles)
* MCPO // OpenWebUI
* MCPO // OpenWebUI
* Machine Virtuelle chez Exo bien boostée
* Machine Virtuelle chez Exo bien booostée


= Glossary =
= Glossary =

Version du 9 novembre 2025 à 22:12

N1 Kubernetes.png

https://learn.microsoft.com/en-us/azure/azure-functions/functions-bindings-http-webhook-trigger?tabs=python-v2%2Cisolated-process%2Cnodejs-v4%2Cfunctionsv2&pivots=programming-language-csharp
https://learn.microsoft.com/en-us/azure/azure-functions/functions-networking-options?tabs=azure-portal
https://komodor.com/blog/aiops-for-kubernetes-or-kaiops/
https://learn.microsoft.com/en-us/azure/architecture/ai-ml/guide/ai-agent-design-patterns
https://learn.microsoft.com/en-us/azure/architecture/ai-ml/architecture/baseline-azure-ai-foundry-chat

N1 Kubernetes Without Azure.drawio.png

Technologies :

  • K8SGPT
  • LangChain (+ LangFuse pour monitoring?)
  • Ollama (+ il y a des trucs qui permettent d'améliorer les perfs des modèles)
  • MCPO // OpenWebUI
  • Machine Virtuelle chez Exo bien booostée

Glossary

Large Language Model (LLM)

A Large Language Model (LLM) is the engine behind an AI application such as ChatGPT. In this case, the engine powering ChatGPT is GPT-4 (or GPT-4o, previously), which is the LLM used by the application.

Azure AI Foundry is a service that allows you to choose which Large Language Model (LLM) you want to use.

Model Context Protocol (MCP)

A Model Context Protocol (MCP) is a protocol that standardizes communication between Large Language Models (LLMs) and external systems , such as ITSM tools (like ServiceNow), Kubernetes clusters, and more.

You can use an MCP client , for example, Continue.dev in your IDE (like VS Code) and then configure MCP servers, such as your Kubernetes cluster, to enable your LLM to interact with these systems.

Retrieval-Augmented Generation (RAG)

Technology Stack

LangChain

LangChain is an application framework that helps structure your prompts using PromptTemplate. For example, with an alerting system, when the AI is queried, you can create a template that guides it to follow a consistent debugging structure in its responses :

from langchain import PromptTemplate

prompt = PromptTemplate(
    input_variables=["alert", "logs", "metrics"],
    template="""
Tu es un expert Kubernetes.
Un incident a été détecté :
{alert}

Voici les logs du pod :
{logs}

Voici ses métriques :
{metrics}

Analyse les causes probables et propose des actions correctives précises.
"""
)

K8sGPT

K8sGPT is an open-source tool that scans Kubernetes clusters, detects issues, and uses a Large Language Model (LLM) such as Azure OpenAI to explain problems and suggest solutions in natural language.

Ollama

Ollama is an open-source tool that lets you download and run large language models (LLMs) like Llama 3 or Mistral locally, allowing you to use AI without relying on the cloud.