Skip to content

K8sGPT AI Backends

A Backend (also called Provider) is a service that provides access to the AI language model. There are many different backends available for K8sGPT. Each backend has its own strengths and weaknesses, so it is important to choose the one that is right for your needs.

Currently, we have a total of 8 backends available:

OpenAI

OpenAI is the default backend for K8sGPT. We recommend using OpenAI first if you are new to K8sGPT and if you have an account on OpenAI. OpenAI comes with the access to powerful language models such as GPT-3.5-Turbo, GPT-4. If you are looking for a powerful and easy-to-use language modeling service, OpenAI is a great option.

  • To use OpenAI you'll need an OpenAI token for authentication purposes. To generate a token use: bash k8sgpt generate
  • To set the token in K8sGPT, use the following command: bash k8sgpt auth add
  • Run the following command to analyze issues within your cluster using OpenAI: bash k8sgpt analyze --explain

Cohere

Cohere allows building conversational apps. It uses Retrieval Augmented Generation (RAG) toolkit that improves LLM's answer accuracy.

  • To use Cohere, visit Cohere dashboard.
  • To configure backend in K8sGPT, use the following command: bash k8sgpt auth add --backend cohere --model command-nightly
  • Run the following command to analyze issues within your cluster using Cohere: bash k8sgpt analyze --explain --backend cohere

Amazon Bedrock

Amazon Bedrock allows building and scaling generative AI applications.

  • To use Bedrock, make sure you have access to Bedrock API and models e.g. in AWS Console you should see something like this:

Bedrock

  • You will need to set the follow local environmental variables: ```

    • AWS_ACCESS_KEY
    • AWS_SECRET_ACCESS_KEY
    • AWS_DEFAULT_REGION ```
  • To configure backend in K8sGPT use auth command: bash k8sgpt auth add --backend amazonbedrock --model anthropic.claude-v2

  • Run the following command to analyze issues within your cluster using Amazon Bedrock: bash k8sgpt analyze --explain --backend amazonbedrock

Amazon SageMaker

The Amazon SageMaker backend allows you to leverage a self-deployed and managed Language Models (LLM) on Amazon SageMaker.

Example how to deploy Amazon SageMaker with cdk is available in llm-sagemaker-jumpstart-cdk repo.

  • To use SageMaker, make sure you have the AWS CLI configured on your machine.
  • You need to have an Amazon SageMaker instance set up.
  • Run the following command to add SageMaker: bash k8sgpt auth add --backend amazonsagemaker --providerRegion eu-west-1 --endpointname endpoint-xxxxxxxxxx
  • Now you are ready to analyze with the Amazon SageMaker backend: bash k8sgpt analyze --explain --backend amazonsagemaker

Azure OpenAI

Azure OpenAI Provider provides REST API access to OpenAI's powerful language models. It gives the users an advanced language AI with powerful models with the security and enterprise promise of Azure.

  • The Azure OpenAI Provider requires a deployment as a prerequisite. You can visit their documentation to create your own. To authenticate with k8sgpt, you would require an Azure OpenAI endpoint of your tenant https://your Azure OpenAI Endpoint,the API key to access your deployment, the deployment name of your model and the model name itself.

  • Run the following command to authenticate with Azure OpenAI: bash k8sgpt auth add --backend azureopenai --baseurl https://<your Azure OpenAI endpoint> --engine <deployment_name> --model <model_name>

  • Now you are ready to analyze with the Azure OpenAI backend: bash k8sgpt analyze --explain --backend azureopenai

Google Gemini

Google Gemini allows generative AI capabilities with multimodal approach (it is capable to understand not only text, but also code, audio, image and video). With Gemini models, a new API was introduced, and this is what is now built-in K8sGPT. This API also works against the Google Cloud Vertex AI service. See also Google AI Studio to get started.

NOTE: Gemini API might be still rolling to some regions. See the available regions for details.

  • To use Google Gemini API in K8sGPT, obtain the API key.
  • To configure Google backend in K8sGPT with gemini-pro model (see all models here) use auth command: bash k8sgpt auth add --backend google --model gemini-pro --password "<Your API KEY>"
  • Run the following command to analyze issues within your cluster with the Google provider: bash k8sgpt analyze --explain --backend google

HuggingFace

Hugging Face is a versatile backend for K8sGPT, offering access to a wide range of pre-trained language models. It provides easy-to-use interfaces for both training and inference tasks. Refer to the Hugging Face documentation for further insights into model usage and capabilities.

  • To use Hugging Face API in K8sGPT, obtain the API key.
  • Configure the HuggingFace backend in K8sGPT by specifying the desired model (see all models here) using auth command: bash k8sgpt auth add --backend huggingface --model <model name>

    NOTE: Since the default gpt-3.5-turbo model is not available in Hugging Face, a valid backend model is required.

  • Once configured, you can analyze issues within your cluster using the Hugging Face provider with the following command: bash k8sgpt analyze --explain --backend huggingface

LocalAI

LocalAI is a local model, which is an OpenAI compatible API. It uses llama.cpp and ggml to run inference on consumer-grade hardware. Models supported by LocalAI for instance are Vicuna, Alpaca, LLaMA, Cerebras, GPT4ALL, GPT4ALL-J and koala.

  • To run local inference, you need to download the models first, for instance you can find ggml compatible models in huggingface.com(for example vicuna, alpaca and koala).
  • To start the API server, follow the instruction in LocalAI.
  • Authenticate K8sGPT with LocalAI: bash k8sgpt auth new --backend localai --model <model_name> --baseurl http://localhost:8080/v1
  • Analyze with a LocalAI backend: bash k8sgpt analyze --explain --backend localai

FakeAI

FakeAI or the NoOpAiProvider might be useful in situations where you need to test a new feature or simulate the behaviour of an AI based-system without actually invoking it. It can help you with local development, testing and troubleshooting. The NoOpAiProvider does not actually perform any AI-based operations but simulates them by echoing the input given as a problem.

Follow the steps outlined below to learn how to utilize the NoOpAiProvider:

  • Authorize k8sgpt with noopai or noop as the Backend Provider: k8sgpt auth -b noopai
  • For the auth token, you can leave it blank as the NoOpAiProvider is configured to work fine with or without any token.

  • Use the analyze and explain command to check for errors in your kubernetes cluster and the NoOpAiProvider should return the error as the solution itself: k8sgpt analyze --explain --backend noopai