AI Concept

De Marijan Stajic | Wiki
Aller à la navigation Aller à la recherche
N1 Kubernetes.png

https://learn.microsoft.com/en-us/azure/azure-functions/functions-bindings-http-webhook-trigger?tabs=python-v2%2Cisolated-process%2Cnodejs-v4%2Cfunctionsv2&pivots=programming-language-csharp
https://learn.microsoft.com/en-us/azure/azure-functions/functions-networking-options?tabs=azure-portal
https://komodor.com/blog/aiops-for-kubernetes-or-kaiops/
https://learn.microsoft.com/en-us/azure/architecture/ai-ml/guide/ai-agent-design-patterns
https://learn.microsoft.com/en-us/azure/architecture/ai-ml/architecture/baseline-azure-ai-foundry-chat

N1 Kubernetes Without Azure.drawio.png

Technologies :

  • K8SGPT
  • LangChain (+ LangFuse pour monitoring?)
  • Ollama (+ il y a des trucs qui permettent d'améliorer les perfs des modèles)
  • MCPO // OpenWebUI
  • Machine Virtuelle chez Exo bien boostée

Glossary

Large Language Model (LLM)

A Large Language Model (LLM) is the engine behind an AI application such as ChatGPT. In this case, the engine powering ChatGPT is GPT-4 (or GPT-4o, previously), which is the LLM used by the application.

Azure AI Foundry is a service that allows you to choose which Large Language Model (LLM) you want to use.

Model Context Protocol (MCP)

A Model Context Protocol (MCP) is a protocol that standardizes communication between Large Language Models (LLMs) and external systems , such as ITSM tools (like ServiceNow), Kubernetes clusters, and more.

You can use an MCP client , for example, Continue.dev in your IDE (like VS Code) and then configure MCP servers, such as your Kubernetes cluster, to enable your LLM to interact with these systems.

Retrieval-Augmented Generation (RAG)

Technology Stack

LangChain

LangChain is an application framework that helps structure your prompts using PromptTemplate. For example, with an alerting system, when the AI is queried, you can create a template that guides it to follow a consistent debugging structure in its responses :

from langchain import PromptTemplate

prompt = PromptTemplate(
    input_variables=["alert", "logs", "metrics"],
    template="""
Tu es un expert Kubernetes.
Un incident a été détecté :
{alert}

Voici les logs du pod :
{logs}

Voici ses métriques :
{metrics}

Analyse les causes probables et propose des actions correctives précises.
"""
)

K8sGPT

K8sGPT is an open-source tool that scans Kubernetes clusters, detects issues, and uses a Large Language Model (LLM) such as Azure OpenAI to explain problems and suggest solutions in natural language.

Ollama

Ollama is an open-source tool that lets you download and run large language models (LLMs) like Llama 3 or Mistral locally, allowing you to use AI without relying on the cloud.