Skip to content

Commit f0ff81a

Browse files
authored
Typo and linting (#945)
1 parent 1f28372 commit f0ff81a

File tree

1 file changed

+10
-26
lines changed
  • docs-source/spring/content/infrastructure/gpu

1 file changed

+10
-26
lines changed

docs-source/spring/content/infrastructure/gpu/_index.md

Lines changed: 10 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,9 @@ description: "GPUs for AI workload"
44
keywords: "gpu ollama llm spring springboot microservices oracle"
55
---
66

7-
Oracle Backend for Spring Boot and Microservices provides an option during installation to provision a set
8-
of Kubernetes nodes with NVIDIA A10 GPUs that are suitable for running AI workloads. If you choose that option during
9-
installation, you may also specify how many nodes are provisioned. The GPU nodes will be in a separate
10-
Node Pool to the normal CPU nodes, which allows you to scale it independently of the CPU nodes.
11-
They are also labeled so that you can target appropriate workloads to them using node selectors
12-
and/or affinity rules.
7+
Oracle Backend for Spring Boot and Microservices provides an option during installation to provision a set of Kubernetes nodes with NVIDIA A10 GPUs that are suitable for running AI workloads. If you choose that option during installation, you may also specify how many nodes are provisioned. The GPU nodes will be in a separate Node Pool to the normal CPU nodes, which allows you to scale it independently of the CPU nodes. They are also labeled so that you can target appropriate workloads to them using node selectors and/or affinity rules.
138

14-
To view a list of nodes in your cluster with a GPU, you can use this command:
9+
To view a list of nodes in your cluster with a GPU, you can use this command:
1510

1611
```bash
1712
$ kubectl get nodes -l 'node.kubernetes.io/instance-type=VM.GPU.A10.1'
@@ -21,12 +16,9 @@ NAME STATUS ROLES AGE VERSION
2116

2217
## Running a Large Language Model on your GPU nodes
2318

24-
One very common use for GPU nodes is to run a self-hosted Large Language Model (LLM)
25-
such as `llama3` for inferencing or `nomic-embed-text` for embedding.
19+
One very common use for GPU nodes is to run a self-hosted Large Language Model (LLM) such as `llama3` for inferencing or `nomic-embed-text` for embedding.
2620

27-
Companies often want to self-host an LLM to avoid sending private or sensitive data
28-
outside of their organization to a third-party provider, or to have more control over
29-
the costs of running the LLM and assocatied infrastructure.
21+
Companies often want to self-host an LLM to avoid sending private or sensitive data outside of their organization to a third-party provider, or to have more control over the costs of running the LLM and associated infrastructure.
3022

3123
One excellent way to self-host LLMs is to use [Ollama](https://ollama.com/).
3224

@@ -44,9 +36,7 @@ To install Ollama on your GPU nodes, you can use the following commands:
4436
helm repo update
4537
```
4638

47-
1. Create a `values.yaml` file to configure how Ollama should be installed, including
48-
which node(s) to run it on. Here is an example that will run Ollama on a GPU node
49-
and will pull the `llama3` model.
39+
1. Create a `values.yaml` file to configure how Ollama should be installed, including which node(s) to run it on. Here is an example that will run Ollama on a GPU node and will pull the `llama3` model.
5040

5141
```yaml
5242
ollama:
@@ -60,8 +50,7 @@ To install Ollama on your GPU nodes, you can use the following commands:
6050
node.kubernetes.io/instance-type: VM.GPU.A10.1
6151
```
6252

63-
For more information on how to configure Ollama using the helm chart, refer to
64-
[its documentation](https://artifacthub.io/packages/helm/ollama-helm/ollama).
53+
For more information on how to configure Ollama using the helm chart, refer to [its documentation](https://artifacthub.io/packages/helm/ollama-helm/ollama).
6554

6655
1. Create a namespace to deploy Ollama in:
6756

@@ -77,8 +66,7 @@ To install Ollama on your GPU nodes, you can use the following commands:
7766

7867
### Interacting with Ollama
7968

80-
You can interact with Ollama using the provided command line tool, called `ollama`.
81-
For example, to list the available models, use the `ollama ls` command:
69+
You can interact with Ollama using the provided command line tool, called `ollama`. For example, to list the available models, use the `ollama ls` command:
8270

8371
```bash
8472
kubectl -n ollama exec svc/ollama -- ollama ls
@@ -88,17 +76,13 @@ llama3:latest 365c0bd3c000 4.7 GB 2 minutes ago
8876

8977
To ask the LLM a question, you can use the `ollama run` command:
9078

91-
```
79+
```bash
9280
$ kubectl -n ollama exec svc/ollama -- ollama run llama3 "what is spring boot?"
93-
Spring Boot is an open-source Java-based framework that simplifies the development
94-
of web applications and microservices. It's a subset of the larger Spring ecosystem,
95-
which provides a comprehensive platform for building enterprise-level applications.
81+
Spring Boot is an open-source Java-based framework that simplifies the development of web applications and microservices. It's a subset of the larger Spring ecosystem, which provides a comprehensive platform for building enterprise-level applications.
9682
9783
...
9884
```
9985
10086
### Using LLMs hosted by Ollama in your Spring application
10187
102-
Our self-paced hands-on example **CloudBank AI** includes an example of how
103-
to [build a simple chatbot](https://oracle.github.io/microservices-datadriven/cloudbank/springai/simple-chat)
104-
using Spring AI and Ollama.
88+
Our self-paced hands-on example **CloudBank AI** includes an example of how to [build a simple chatbot](https://oracle.github.io/microservices-datadriven/cloudbank/springai/simple-chat) using Spring AI and Ollama.

0 commit comments

Comments
 (0)