Unverified Commit 3f71ba40 authored by Mélony QIN's avatar Mélony QIN Committed by GitHub
Browse files

Correct the kubernetes terminology (#3843)



* add details on kubernetes deployment and separate the testing process

* Update examples/kubernetes/README.md

thanks for suggesting this change, I agree with you and let's make this project better together !
Co-authored-by: default avatarJonZeolla <Zeolla@gmail.com>

---------
Co-authored-by: default avatarQIN Mélony <MQN1@dsone.3ds.com>
Co-authored-by: default avatarJonZeolla <Zeolla@gmail.com>
parent 88a67127
...@@ -7,12 +7,24 @@ ...@@ -7,12 +7,24 @@
## Steps ## Steps
1. Create the Ollama namespace, daemon set, and service 1. Create the Ollama namespace, deployment, and service
```bash ```bash
kubectl apply -f cpu.yaml kubectl apply -f cpu.yaml
``` ```
## (Optional) Hardware Acceleration
Hardware acceleration in Kubernetes requires NVIDIA's [`k8s-device-plugin`](https://github.com/NVIDIA/k8s-device-plugin) which is deployed in Kubernetes in form of daemonset. Follow the link for more details.
Once configured, create a GPU enabled Ollama deployment.
```bash
kubectl apply -f gpu.yaml
```
## Test
1. Port forward the Ollama service to connect and use it locally 1. Port forward the Ollama service to connect and use it locally
```bash ```bash
...@@ -23,14 +35,4 @@ ...@@ -23,14 +35,4 @@
```bash ```bash
ollama run orca-mini:3b ollama run orca-mini:3b
``` ```
\ No newline at end of file
## (Optional) Hardware Acceleration
Hardware acceleration in Kubernetes requires NVIDIA's [`k8s-device-plugin`](https://github.com/NVIDIA/k8s-device-plugin). Follow the link for more details.
Once configured, create a GPU enabled Ollama deployment.
```bash
kubectl apply -f gpu.yaml
```
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment