# GKE Chatbot | K8s Summit 2024 GCP Workshop
Preparation
- Run Gemma 2 Powered Chatbot on GKE Autopilot
- 主要使用這個頁面的 command, 僅使用 qwiklab 的資源
|
|
任務 2. 部署 Ollama
|
|
任務 3. 使用 Ollama CLI 與 Gemma 2 互動
|
|
任務 4. 部署 Open WebUI,打造更友善的 Chatbot 互動介面
openwebui.yaml
|
|
|
|
|
|
openwebui.yaml
|
|
|
|
|
|
Finetuning requires a base model and a dataset. For this post, the dell-research-harvard/AmericanStories
dataset will be used to fine-tune the Llama 2 7b base model. GCS will be used for storing the base model. GKE with GCSFuse is used to transparently save the fine-tuned model to GCS. This provides a cost efficient way to store and serve the model and only pay for the storage used by the model.
This is the sample code for the Distributed load testing using Google Kubernetes Engine tutorial.
This code is Apache 2.0 licensed and more information can be found in LICENSE
. For information on licenses for third party software and libraries, refer to the docker-image/licenses
directory.
The Guestbook sample demonstrates how to deploy a Kubernetes application with a front end service and a back end service using the Cloud Code IDE extension.
For details on how to use this sample as a template in Cloud Code, read the documentation for Cloud Code for VS Code or IntelliJ.
{{m.message}}