mirror of
https://github.com/amithkoujalgi/ollama4j.git
synced 2025-05-15 11:57:12 +02:00
clean up
This commit is contained in:
parent
e9ec19a62d
commit
08118b88fc
13
README.md
13
README.md
@ -81,9 +81,9 @@ You can then use the Ollama Java APIs by importing `ollama4j`:
|
|||||||
import io.github.amithkoujalgi.ollama4j.core.OllamaAPI;
|
import io.github.amithkoujalgi.ollama4j.core.OllamaAPI;
|
||||||
```
|
```
|
||||||
|
|
||||||
### Try out the APIs
|
### Try out the APIs with Ollama server
|
||||||
|
|
||||||
For simplest way to get started, I prefer to use the Ollama docker setup.
|
For simplest way to get started, I prefer to use the Ollama server in a docker setup.
|
||||||
|
|
||||||
Start the Ollama docker container:
|
Start the Ollama docker container:
|
||||||
|
|
||||||
@ -97,6 +97,9 @@ With GPUs
|
|||||||
docker run -it --gpus=all -v ~/ollama:/root/.ollama -p 11434:11434 ollama/ollama
|
docker run -it --gpus=all -v ~/ollama:/root/.ollama -p 11434:11434 ollama/ollama
|
||||||
```
|
```
|
||||||
|
|
||||||
|
Read the full list of APIs
|
||||||
|
here: https://amithkoujalgi.github.io/ollama4j/io/github/amithkoujalgi/ollama4j/core/OllamaAPI.html
|
||||||
|
|
||||||
Instantiate `OllamaAPI`
|
Instantiate `OllamaAPI`
|
||||||
|
|
||||||
```java
|
```java
|
||||||
@ -118,7 +121,7 @@ public class Main {
|
|||||||
public static void main(String[] args) {
|
public static void main(String[] args) {
|
||||||
String host = "http://localhost:11434/";
|
String host = "http://localhost:11434/";
|
||||||
OllamaAPI ollamaAPI = new OllamaAPI(host);
|
OllamaAPI ollamaAPI = new OllamaAPI(host);
|
||||||
ollamaAPI.pullModel(OllamaModel.LLAMA2);
|
ollamaAPI.pullModel(OllamaModelType.LLAMA2);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
@ -216,7 +219,7 @@ public class Main {
|
|||||||
public static void main(String[] args) {
|
public static void main(String[] args) {
|
||||||
String host = "http://localhost:11434/";
|
String host = "http://localhost:11434/";
|
||||||
OllamaAPI ollamaAPI = new OllamaAPI(host);
|
OllamaAPI ollamaAPI = new OllamaAPI(host);
|
||||||
String response = ollamaAPI.ask(OllamaModel.LLAMA2, "Who are you?");
|
String response = ollamaAPI.ask(OllamaModelType.LLAMA2, "Who are you?");
|
||||||
System.out.println(response);
|
System.out.println(response);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -229,7 +232,7 @@ public class Main {
|
|||||||
public static void main(String[] args) {
|
public static void main(String[] args) {
|
||||||
String host = "http://localhost:11434/";
|
String host = "http://localhost:11434/";
|
||||||
OllamaAPI ollamaAPI = new OllamaAPI(host);
|
OllamaAPI ollamaAPI = new OllamaAPI(host);
|
||||||
OllamaAsyncResultCallback ollamaAsyncResultCallback = ollamaAPI.askAsync(OllamaModel.LLAMA2, "Who are you?");
|
OllamaAsyncResultCallback ollamaAsyncResultCallback = ollamaAPI.askAsync(OllamaModelType.LLAMA2, "Who are you?");
|
||||||
while (true) {
|
while (true) {
|
||||||
if (ollamaAsyncResultCallback.isComplete()) {
|
if (ollamaAsyncResultCallback.isComplete()) {
|
||||||
System.out.println(ollamaAsyncResultCallback.getResponse());
|
System.out.println(ollamaAsyncResultCallback.getResponse());
|
||||||
|
Loading…
x
Reference in New Issue
Block a user