mirror of
				https://github.com/amithkoujalgi/ollama4j.git
				synced 2025-10-28 23:20:42 +01:00 
			
		
		
		
	Compare commits
	
		
			31 Commits
		
	
	
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
|   | 43f43c9f81 | ||
|   | 65f00defcf | ||
|   | d716b81342 | ||
|   | 272ba445f6 | ||
|   | d9816d8869 | ||
|   | 874736eb16 | ||
|   | 9c16ccbf81 | ||
|   | 40a3aa31dc | ||
|   | 90669b611b | ||
|   | f10c7ac725 | ||
|   | 38dca3cd0d | ||
|   | 44bb35b168 | ||
|   | 9832caf503 | ||
|   | 0c4e8e306e | ||
|   | 075416eb9c | ||
|   | 4260fbbc32 | ||
|   | 0bec697a86 | ||
|   | 4ca6eef8fd | ||
|   | a635dd9be2 | ||
|   | 14982011d9 | ||
|   | 65d852fdc9 | ||
|   | d483c23c81 | ||
|   | 273b1e47ca | ||
|   | 5c5cdba4cd | ||
|   | 24674ea483 | ||
|   | f9063484f3 | ||
|   | 00a3e51a93 | ||
|   | bc20468f28 | ||
|   | c7ac50a805 | ||
|   | f8cd7bc013 | ||
|   | 3469bf314b | 
							
								
								
									
										98
									
								
								docs/docs/apis-ask/chat.md
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										98
									
								
								docs/docs/apis-ask/chat.md
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,98 @@ | ||||
| --- | ||||
| sidebar_position: 7 | ||||
| --- | ||||
|  | ||||
| # Chat | ||||
|  | ||||
| This API lets you create a conversation with LLMs. Using this API enables you to ask questions to the model including  | ||||
| information using the history of already asked questions and the respective answers. | ||||
|  | ||||
| ## Create a new conversation and use chat history to augment follow up questions | ||||
|  | ||||
| ```java | ||||
| public class Main { | ||||
|  | ||||
|     public static void main(String[] args) { | ||||
|  | ||||
|         String host = "http://localhost:11434/"; | ||||
|  | ||||
|         OllamaAPI ollamaAPI = new OllamaAPI(host); | ||||
|         OllamaChatRequestBuilder builder = OllamaChatRequestBuilder.getInstance(OllamaModelType.LLAMA2); | ||||
|  | ||||
|         // create first user question | ||||
|         OllamaChatRequestModel requestModel = builder.withMessage(OllamaChatMessageRole.USER,"What is the capital of France?") | ||||
|              .build(); | ||||
|  | ||||
|         // start conversation with model | ||||
|         OllamaChatResult chatResult = ollamaAPI.chat(requestModel); | ||||
|  | ||||
|         System.out.println("First answer: " + chatResult.getResponse()); | ||||
|  | ||||
|         // create next userQuestion | ||||
|         requestModel = builder.withMessages(chatResult.getChatHistory()).withMessage(OllamaChatMessageRole.USER,"And what is the second largest city?").build(); | ||||
|  | ||||
|         // "continue" conversation with model | ||||
|         chatResult = ollamaAPI.chat(requestModel); | ||||
|  | ||||
|         System.out.println("Second answer: " + chatResult.getResponse()); | ||||
|  | ||||
|         System.out.println("Chat History: " + chatResult.getChatHistory()); | ||||
|     } | ||||
| } | ||||
|  | ||||
| ``` | ||||
| You will get a response similar to: | ||||
|  | ||||
| > First answer: Should be Paris! | ||||
| >  | ||||
| > Second answer: Marseille. | ||||
| >  | ||||
| > Chat History: | ||||
|  | ||||
| ```json | ||||
| [ { | ||||
|     "role" : "user", | ||||
|     "content" : "What is the capital of France?", | ||||
|     "images" : [ ] | ||||
|   }, { | ||||
|     "role" : "assistant", | ||||
|     "content" : "Should be Paris!", | ||||
|     "images" : [ ] | ||||
|   }, { | ||||
|     "role" : "user", | ||||
|     "content" : "And what is the second largest city?", | ||||
|     "images" : [ ] | ||||
|   }, { | ||||
|     "role" : "assistant", | ||||
|     "content" : "Marseille.", | ||||
|     "images" : [ ] | ||||
|   } ] | ||||
| ``` | ||||
|  | ||||
| ## Create a new conversation with individual system prompt | ||||
| ```java | ||||
| public class Main { | ||||
|  | ||||
|     public static void main(String[] args) { | ||||
|  | ||||
|         String host = "http://localhost:11434/"; | ||||
|  | ||||
|         OllamaAPI ollamaAPI = new OllamaAPI(host); | ||||
|         OllamaChatRequestBuilder builder = OllamaChatRequestBuilder.getInstance(OllamaModelType.LLAMA2); | ||||
|  | ||||
|         // create request with system-prompt (overriding the model defaults) and user question | ||||
|         OllamaChatRequestModel requestModel = builder.withMessage(OllamaChatMessageRole.SYSTEM, "You are a silent bot that only says 'NI'. Do not say anything else under any circumstances!") | ||||
|              .withMessage(OllamaChatMessageRole.USER,"What is the capital of France? And what's France's connection with Mona Lisa?") | ||||
|              .build(); | ||||
|  | ||||
|         // start conversation with model | ||||
|         OllamaChatResult chatResult = ollamaAPI.chat(requestModel); | ||||
|  | ||||
|         System.out.println(chatResult.getResponse()); | ||||
|     } | ||||
| } | ||||
|  | ||||
| ``` | ||||
| You will get a response similar to: | ||||
|  | ||||
| > NI. | ||||
							
								
								
									
										4
									
								
								pom.xml
									
									
									
									
									
								
							
							
						
						
									
										4
									
								
								pom.xml
									
									
									
									
									
								
							| @@ -4,7 +4,7 @@ | ||||
|  | ||||
|     <groupId>io.github.amithkoujalgi</groupId> | ||||
|     <artifactId>ollama4j</artifactId> | ||||
|     <version>1.0.49</version> | ||||
|     <version>1.0.51</version> | ||||
|  | ||||
|     <name>Ollama4j</name> | ||||
|     <description>Java library for interacting with Ollama API.</description> | ||||
| @@ -39,7 +39,7 @@ | ||||
|         <connection>scm:git:git@github.com:amithkoujalgi/ollama4j.git</connection> | ||||
|         <developerConnection>scm:git:https://github.com/amithkoujalgi/ollama4j.git</developerConnection> | ||||
|         <url>https://github.com/amithkoujalgi/ollama4j</url> | ||||
|         <tag>v1.0.49</tag> | ||||
|         <tag>v1.0.51</tag> | ||||
|     </scm> | ||||
|  | ||||
|     <build> | ||||
|   | ||||
| @@ -2,10 +2,16 @@ package io.github.amithkoujalgi.ollama4j.core; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.exceptions.OllamaBaseException; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.*; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatMessage; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatRequestBuilder; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatRequestModel; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatResult; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.request.CustomModelFileContentsRequest; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.request.CustomModelFilePathRequest; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.request.ModelEmbeddingsRequest; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.request.ModelRequest; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.request.OllamaChatEndpointCaller; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.request.OllamaGenerateEndpointCaller; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Options; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Utils; | ||||
| import java.io.BufferedReader; | ||||
| @@ -343,7 +349,7 @@ public class OllamaAPI { | ||||
|       throws OllamaBaseException, IOException, InterruptedException { | ||||
|     OllamaRequestModel ollamaRequestModel = new OllamaRequestModel(model, prompt); | ||||
|     ollamaRequestModel.setOptions(options.getOptionsMap()); | ||||
|     return generateSync(ollamaRequestModel); | ||||
|     return generateSyncForOllamaRequestModel(ollamaRequestModel); | ||||
|   } | ||||
|  | ||||
|   /** | ||||
| @@ -387,7 +393,7 @@ public class OllamaAPI { | ||||
|     } | ||||
|     OllamaRequestModel ollamaRequestModel = new OllamaRequestModel(model, prompt, images); | ||||
|     ollamaRequestModel.setOptions(options.getOptionsMap()); | ||||
|     return generateSync(ollamaRequestModel); | ||||
|     return generateSyncForOllamaRequestModel(ollamaRequestModel); | ||||
|   } | ||||
|  | ||||
|   /** | ||||
| @@ -411,9 +417,50 @@ public class OllamaAPI { | ||||
|     } | ||||
|     OllamaRequestModel ollamaRequestModel = new OllamaRequestModel(model, prompt, images); | ||||
|     ollamaRequestModel.setOptions(options.getOptionsMap()); | ||||
|     return generateSync(ollamaRequestModel); | ||||
|     return generateSyncForOllamaRequestModel(ollamaRequestModel); | ||||
|   } | ||||
|  | ||||
|  | ||||
|    | ||||
|   /** | ||||
|    * Ask a question to a model based on a given message stack (i.e. a chat history). Creates a synchronous call to the api  | ||||
|    * 'api/chat'.  | ||||
|    *  | ||||
|    * @param model the ollama model to ask the question to | ||||
|    * @param messages chat history / message stack to send to the model | ||||
|    * @return {@link OllamaChatResult} containing the api response and the message history including the newly aqcuired assistant response. | ||||
|      * @throws OllamaBaseException any response code than 200 has been returned | ||||
|      * @throws IOException in case the responseStream can not be read | ||||
|      * @throws InterruptedException in case the server is not reachable or network issues happen | ||||
|    */ | ||||
|   public OllamaChatResult chat(String model, List<OllamaChatMessage> messages)  throws OllamaBaseException, IOException, InterruptedException{ | ||||
|     OllamaChatRequestBuilder builder = OllamaChatRequestBuilder.getInstance(model); | ||||
|     return chat(builder.withMessages(messages).build()); | ||||
|   } | ||||
|  | ||||
|   /** | ||||
|    * Ask a question to a model using an {@link OllamaChatRequestModel}. This can be constructed using an {@link OllamaChatRequestBuilder}.  | ||||
|    *  | ||||
|    * Hint: the OllamaChatRequestModel#getStream() property is not implemented. | ||||
|    *  | ||||
|    * @param request request object to be sent to the server | ||||
|    * @return  | ||||
|   * @throws OllamaBaseException any response code than 200 has been returned | ||||
|   * @throws IOException in case the responseStream can not be read | ||||
|   * @throws InterruptedException in case the server is not reachable or network issues happen | ||||
|    */ | ||||
|   public OllamaChatResult chat(OllamaChatRequestModel request)  throws OllamaBaseException, IOException, InterruptedException{ | ||||
|     OllamaChatEndpointCaller requestCaller = new OllamaChatEndpointCaller(host, basicAuth, requestTimeoutSeconds, verbose); | ||||
|     //TODO: implement async way | ||||
|     if(request.isStream()){ | ||||
|       throw new UnsupportedOperationException("Streamed chat responses are not implemented yet"); | ||||
|     } | ||||
|     OllamaResult result = requestCaller.generateSync(request); | ||||
|     return new OllamaChatResult(result.getResponse(), result.getResponseTime(), result.getHttpStatusCode(), request.getMessages()); | ||||
|   } | ||||
|  | ||||
|   // technical private methods // | ||||
|  | ||||
|   private static String encodeFileToBase64(File file) throws IOException { | ||||
|     return Base64.getEncoder().encodeToString(Files.readAllBytes(file.toPath())); | ||||
|   } | ||||
| @@ -436,58 +483,10 @@ public class OllamaAPI { | ||||
|     } | ||||
|   } | ||||
|  | ||||
|   private OllamaResult generateSync(OllamaRequestModel ollamaRequestModel) | ||||
|   private OllamaResult generateSyncForOllamaRequestModel(OllamaRequestModel ollamaRequestModel) | ||||
|       throws OllamaBaseException, IOException, InterruptedException { | ||||
|     long startTime = System.currentTimeMillis(); | ||||
|     HttpClient httpClient = HttpClient.newHttpClient(); | ||||
|     URI uri = URI.create(this.host + "/api/generate"); | ||||
|     HttpRequest.Builder requestBuilder = | ||||
|         getRequestBuilderDefault(uri) | ||||
|             .POST( | ||||
|                 HttpRequest.BodyPublishers.ofString( | ||||
|                     Utils.getObjectMapper().writeValueAsString(ollamaRequestModel))); | ||||
|     HttpRequest request = requestBuilder.build(); | ||||
|     if (verbose) logger.info("Asking model: " + ollamaRequestModel); | ||||
|     HttpResponse<InputStream> response = | ||||
|         httpClient.send(request, HttpResponse.BodyHandlers.ofInputStream()); | ||||
|     int statusCode = response.statusCode(); | ||||
|     InputStream responseBodyStream = response.body(); | ||||
|     StringBuilder responseBuffer = new StringBuilder(); | ||||
|     try (BufferedReader reader = | ||||
|         new BufferedReader(new InputStreamReader(responseBodyStream, StandardCharsets.UTF_8))) { | ||||
|       String line; | ||||
|       while ((line = reader.readLine()) != null) { | ||||
|         if (statusCode == 404) { | ||||
|           logger.warn("Status code: 404 (Not Found)"); | ||||
|           OllamaErrorResponseModel ollamaResponseModel = | ||||
|               Utils.getObjectMapper().readValue(line, OllamaErrorResponseModel.class); | ||||
|           responseBuffer.append(ollamaResponseModel.getError()); | ||||
|         } else if (statusCode == 401) { | ||||
|           logger.warn("Status code: 401 (Unauthorized)"); | ||||
|           OllamaErrorResponseModel ollamaResponseModel = | ||||
|               Utils.getObjectMapper() | ||||
|                   .readValue("{\"error\":\"Unauthorized\"}", OllamaErrorResponseModel.class); | ||||
|           responseBuffer.append(ollamaResponseModel.getError()); | ||||
|         } else { | ||||
|           OllamaResponseModel ollamaResponseModel = | ||||
|               Utils.getObjectMapper().readValue(line, OllamaResponseModel.class); | ||||
|           if (!ollamaResponseModel.isDone()) { | ||||
|             responseBuffer.append(ollamaResponseModel.getResponse()); | ||||
|           } | ||||
|         } | ||||
|       } | ||||
|     } | ||||
|  | ||||
|     if (statusCode != 200) { | ||||
|       logger.error("Status code " + statusCode); | ||||
|       throw new OllamaBaseException(responseBuffer.toString()); | ||||
|     } else { | ||||
|       long endTime = System.currentTimeMillis(); | ||||
|       OllamaResult ollamaResult = | ||||
|           new OllamaResult(responseBuffer.toString().trim(), endTime - startTime, statusCode); | ||||
|       if (verbose) logger.info("Model response: " + ollamaResult); | ||||
|       return ollamaResult; | ||||
|     } | ||||
|         OllamaGenerateEndpointCaller requestCaller = new OllamaGenerateEndpointCaller(host, basicAuth, requestTimeoutSeconds, verbose); | ||||
|         return requestCaller.generateSync(ollamaRequestModel); | ||||
|   } | ||||
|  | ||||
|   /** | ||||
|   | ||||
| @@ -1,8 +1,6 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models; | ||||
|  | ||||
| import com.fasterxml.jackson.annotation.JsonIgnoreProperties; | ||||
| import com.fasterxml.jackson.annotation.JsonProperty; | ||||
| import java.util.List; | ||||
| import lombok.Data; | ||||
|  | ||||
| @Data | ||||
|   | ||||
| @@ -3,12 +3,15 @@ package io.github.amithkoujalgi.ollama4j.core.models; | ||||
| import static io.github.amithkoujalgi.ollama4j.core.utils.Utils.getObjectMapper; | ||||
|  | ||||
| import com.fasterxml.jackson.core.JsonProcessingException; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.OllamaRequestBody; | ||||
|  | ||||
| import java.util.List; | ||||
| import java.util.Map; | ||||
| import lombok.Data; | ||||
|  | ||||
| @Data | ||||
| public class OllamaRequestModel { | ||||
| public class OllamaRequestModel implements OllamaRequestBody{ | ||||
|  | ||||
|   private String model; | ||||
|   private String prompt; | ||||
|   | ||||
| @@ -0,0 +1,41 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.chat; | ||||
|  | ||||
| import static io.github.amithkoujalgi.ollama4j.core.utils.Utils.getObjectMapper; | ||||
|  | ||||
| import com.fasterxml.jackson.core.JsonProcessingException; | ||||
| import java.io.File; | ||||
| import java.util.List; | ||||
| import lombok.AllArgsConstructor; | ||||
| import lombok.Data; | ||||
| import lombok.NoArgsConstructor; | ||||
| import lombok.NonNull; | ||||
| import lombok.RequiredArgsConstructor; | ||||
|  | ||||
| /** | ||||
|  * Defines a single Message to be used inside a chat request against the ollama /api/chat endpoint. | ||||
|  * | ||||
|  * @see <a href="https://github.com/ollama/ollama/blob/main/docs/api.md#generate-a-chat-completion">Generate chat completion</a> | ||||
|  */ | ||||
| @Data | ||||
| @AllArgsConstructor | ||||
| @RequiredArgsConstructor | ||||
| @NoArgsConstructor | ||||
| public class OllamaChatMessage { | ||||
|  | ||||
|     @NonNull | ||||
|     private OllamaChatMessageRole role; | ||||
|  | ||||
|     @NonNull | ||||
|     private String content; | ||||
|  | ||||
|     private List<File> images; | ||||
|      | ||||
|       @Override | ||||
|   public String toString() { | ||||
|     try { | ||||
|       return getObjectMapper().writerWithDefaultPrettyPrinter().writeValueAsString(this); | ||||
|     } catch (JsonProcessingException e) { | ||||
|       throw new RuntimeException(e); | ||||
|     } | ||||
|   } | ||||
| } | ||||
| @@ -0,0 +1,19 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.chat; | ||||
|  | ||||
| import com.fasterxml.jackson.annotation.JsonValue; | ||||
|  | ||||
| /** | ||||
|  * Defines the possible Chat Message roles. | ||||
|  */ | ||||
| public enum OllamaChatMessageRole { | ||||
|     SYSTEM("system"), | ||||
|     USER("user"), | ||||
|     ASSISTANT("assistant"); | ||||
|  | ||||
|     @JsonValue | ||||
|     private String roleName; | ||||
|  | ||||
|     private OllamaChatMessageRole(String roleName){ | ||||
|         this.roleName = roleName; | ||||
|     } | ||||
| } | ||||
| @@ -0,0 +1,68 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.chat; | ||||
|  | ||||
| import java.io.File; | ||||
| import java.util.ArrayList; | ||||
| import java.util.List; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Options; | ||||
|  | ||||
| /** | ||||
|  * Helper class for creating {@link OllamaChatRequestModel} objects using the builder-pattern. | ||||
|  */ | ||||
| public class OllamaChatRequestBuilder { | ||||
|  | ||||
|     private OllamaChatRequestBuilder(String model, List<OllamaChatMessage> messages){ | ||||
|         request = new OllamaChatRequestModel(model, messages); | ||||
|     } | ||||
|  | ||||
|     private OllamaChatRequestModel request; | ||||
|  | ||||
|     public static OllamaChatRequestBuilder getInstance(String model){ | ||||
|         return new OllamaChatRequestBuilder(model, new ArrayList<>()); | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestModel build(){ | ||||
|         return request; | ||||
|     } | ||||
|  | ||||
|     public void reset(){ | ||||
|         request = new OllamaChatRequestModel(request.getModel(), new ArrayList<>()); | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withMessage(OllamaChatMessageRole role, String content, File... images){ | ||||
|         List<OllamaChatMessage> messages = this.request.getMessages(); | ||||
|         messages.add(new OllamaChatMessage(role,content,List.of(images))); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withMessages(List<OllamaChatMessage> messages){ | ||||
|         this.request.getMessages().addAll(messages); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withOptions(Options options){ | ||||
|         this.request.setOptions(options); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withFormat(String format){ | ||||
|         this.request.setFormat(format); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withTemplate(String template){ | ||||
|         this.request.setTemplate(template); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withStreaming(){ | ||||
|         this.request.setStream(true); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
|     public OllamaChatRequestBuilder withKeepAlive(String keepAlive){ | ||||
|         this.request.setKeepAlive(keepAlive); | ||||
|         return this; | ||||
|     } | ||||
|  | ||||
| } | ||||
| @@ -0,0 +1,47 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.chat; | ||||
|  | ||||
| import java.util.List; | ||||
|  | ||||
| import com.fasterxml.jackson.core.JsonProcessingException; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.OllamaRequestBody; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Options; | ||||
|  | ||||
| import static io.github.amithkoujalgi.ollama4j.core.utils.Utils.getObjectMapper; | ||||
|  | ||||
| import lombok.AllArgsConstructor; | ||||
| import lombok.Data; | ||||
| import lombok.NonNull; | ||||
| import lombok.RequiredArgsConstructor; | ||||
|  | ||||
| /** | ||||
|  * Defines a Request to use against the ollama /api/chat endpoint. | ||||
|  * | ||||
|  * @see <a | ||||
|  *     href="https://github.com/ollama/ollama/blob/main/docs/api.md#generate-a-chat-completion">Generate | ||||
|  *     Chat Completion</a> | ||||
|  */ | ||||
| @Data | ||||
| @AllArgsConstructor | ||||
| @RequiredArgsConstructor | ||||
| public class OllamaChatRequestModel implements OllamaRequestBody { | ||||
|  | ||||
|   @NonNull private String model; | ||||
|  | ||||
|   @NonNull private List<OllamaChatMessage> messages; | ||||
|  | ||||
|   private String format; | ||||
|   private Options options; | ||||
|   private String template; | ||||
|   private boolean stream; | ||||
|   private String keepAlive; | ||||
|  | ||||
|   @Override | ||||
|   public String toString() { | ||||
|     try { | ||||
|       return getObjectMapper().writerWithDefaultPrettyPrinter().writeValueAsString(this); | ||||
|     } catch (JsonProcessingException e) { | ||||
|       throw new RuntimeException(e); | ||||
|     } | ||||
|   } | ||||
| } | ||||
| @@ -0,0 +1,21 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.chat; | ||||
|  | ||||
| import com.fasterxml.jackson.annotation.JsonProperty; | ||||
|  | ||||
| import java.util.List; | ||||
| import lombok.Data; | ||||
|  | ||||
| @Data | ||||
| public class OllamaChatResponseModel { | ||||
|     private String model; | ||||
|     private @JsonProperty("created_at") String createdAt; | ||||
|     private OllamaChatMessage message; | ||||
|     private boolean done; | ||||
|     private List<Integer> context; | ||||
|     private @JsonProperty("total_duration") Long totalDuration; | ||||
|     private @JsonProperty("load_duration") Long loadDuration; | ||||
|     private @JsonProperty("prompt_eval_duration") Long promptEvalDuration; | ||||
|     private @JsonProperty("eval_duration") Long evalDuration; | ||||
|     private @JsonProperty("prompt_eval_count") Integer promptEvalCount; | ||||
|     private @JsonProperty("eval_count") Integer evalCount; | ||||
| } | ||||
| @@ -0,0 +1,32 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.chat; | ||||
|  | ||||
| import java.util.List; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.OllamaResult; | ||||
|  | ||||
| /** | ||||
|  * Specific chat-API result that contains the chat history sent to the model and appends the answer as {@link OllamaChatResult} given by the | ||||
|  * {@link OllamaChatMessageRole#ASSISTANT} role. | ||||
|  */ | ||||
| public class OllamaChatResult extends OllamaResult{ | ||||
|  | ||||
|     private List<OllamaChatMessage> chatHistory; | ||||
|  | ||||
|     public OllamaChatResult(String response, long responseTime, int httpStatusCode, | ||||
|             List<OllamaChatMessage> chatHistory) { | ||||
|         super(response, responseTime, httpStatusCode); | ||||
|         this.chatHistory = chatHistory; | ||||
|         appendAnswerToChatHistory(response); | ||||
|     } | ||||
|  | ||||
|     public List<OllamaChatMessage> getChatHistory() { | ||||
|         return chatHistory; | ||||
|     }  | ||||
|  | ||||
|     private void appendAnswerToChatHistory(String answer){ | ||||
|         OllamaChatMessage assistantMessage = new OllamaChatMessage(OllamaChatMessageRole.ASSISTANT, answer); | ||||
|         this.chatHistory.add(assistantMessage); | ||||
|     } | ||||
|      | ||||
|      | ||||
| } | ||||
| @@ -0,0 +1,44 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.request; | ||||
|  | ||||
| import org.slf4j.Logger; | ||||
| import org.slf4j.LoggerFactory; | ||||
|  | ||||
| import com.fasterxml.jackson.core.JsonProcessingException; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.BasicAuth; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatResponseModel; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Utils; | ||||
|  | ||||
| /** | ||||
|  * Specialization class for requests | ||||
|  */ | ||||
| public class OllamaChatEndpointCaller extends OllamaEndpointCaller{ | ||||
|  | ||||
|     private static final Logger LOG = LoggerFactory.getLogger(OllamaChatEndpointCaller.class); | ||||
|  | ||||
|     public OllamaChatEndpointCaller(String host, BasicAuth basicAuth, long requestTimeoutSeconds, boolean verbose) { | ||||
|         super(host, basicAuth, requestTimeoutSeconds, verbose); | ||||
|     } | ||||
|  | ||||
|     @Override | ||||
|     protected String getEndpointSuffix() { | ||||
|         return "/api/chat"; | ||||
|     } | ||||
|  | ||||
|     @Override | ||||
|     protected boolean parseResponseAndAddToBuffer(String line, StringBuilder responseBuffer) { | ||||
|                 try { | ||||
|                     OllamaChatResponseModel ollamaResponseModel = Utils.getObjectMapper().readValue(line, OllamaChatResponseModel.class); | ||||
|                     responseBuffer.append(ollamaResponseModel.getMessage().getContent()); | ||||
|                     return ollamaResponseModel.isDone(); | ||||
|                 } catch (JsonProcessingException e) { | ||||
|                     LOG.error("Error parsing the Ollama chat response!",e); | ||||
|                     return true; | ||||
|                 }          | ||||
|     } | ||||
|  | ||||
|  | ||||
|  | ||||
|      | ||||
|  | ||||
| } | ||||
| @@ -0,0 +1,150 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.request; | ||||
|  | ||||
| import java.io.BufferedReader; | ||||
| import java.io.IOException; | ||||
| import java.io.InputStream; | ||||
| import java.io.InputStreamReader; | ||||
| import java.net.URI; | ||||
| import java.net.http.HttpClient; | ||||
| import java.net.http.HttpRequest; | ||||
| import java.net.http.HttpResponse; | ||||
| import java.nio.charset.StandardCharsets; | ||||
| import java.time.Duration; | ||||
| import java.util.Base64; | ||||
|  | ||||
| import org.slf4j.Logger; | ||||
| import org.slf4j.LoggerFactory; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.OllamaAPI; | ||||
| import io.github.amithkoujalgi.ollama4j.core.exceptions.OllamaBaseException; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.BasicAuth; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.OllamaErrorResponseModel; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.OllamaResult; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.OllamaRequestBody; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Utils; | ||||
|  | ||||
| /** | ||||
|  * Abstract helperclass to call the ollama api server. | ||||
|  */ | ||||
| public abstract class OllamaEndpointCaller { | ||||
|      | ||||
|     private static final Logger LOG = LoggerFactory.getLogger(OllamaAPI.class); | ||||
|  | ||||
|     private String host; | ||||
|     private BasicAuth basicAuth; | ||||
|     private long requestTimeoutSeconds; | ||||
|     private boolean verbose; | ||||
|  | ||||
|     public OllamaEndpointCaller(String host, BasicAuth basicAuth, long requestTimeoutSeconds, boolean verbose) { | ||||
|         this.host = host; | ||||
|         this.basicAuth = basicAuth; | ||||
|         this.requestTimeoutSeconds = requestTimeoutSeconds; | ||||
|         this.verbose = verbose; | ||||
|     } | ||||
|  | ||||
|     protected abstract String getEndpointSuffix(); | ||||
|  | ||||
|     protected abstract boolean parseResponseAndAddToBuffer(String line, StringBuilder responseBuffer); | ||||
|  | ||||
|      | ||||
|     /** | ||||
|      * Calls the api server on the given host and endpoint suffix asynchronously, aka waiting for the response. | ||||
|      *  | ||||
|      * @param body POST body payload | ||||
|      * @return result answer given by the assistant | ||||
|      * @throws OllamaBaseException any response code than 200 has been returned | ||||
|      * @throws IOException in case the responseStream can not be read | ||||
|      * @throws InterruptedException in case the server is not reachable or network issues happen | ||||
|      */ | ||||
|     public OllamaResult generateSync(OllamaRequestBody body)  throws OllamaBaseException, IOException, InterruptedException{ | ||||
|  | ||||
|         // Create Request | ||||
|     long startTime = System.currentTimeMillis(); | ||||
|     HttpClient httpClient = HttpClient.newHttpClient(); | ||||
|     URI uri = URI.create(this.host + getEndpointSuffix()); | ||||
|     HttpRequest.Builder requestBuilder = | ||||
|         getRequestBuilderDefault(uri) | ||||
|             .POST( | ||||
|                 body.getBodyPublisher()); | ||||
|     HttpRequest request = requestBuilder.build(); | ||||
|     if (this.verbose) LOG.info("Asking model: " + body.toString()); | ||||
|     HttpResponse<InputStream> response = | ||||
|         httpClient.send(request, HttpResponse.BodyHandlers.ofInputStream()); | ||||
|      | ||||
|          | ||||
|         int statusCode = response.statusCode(); | ||||
|     InputStream responseBodyStream = response.body(); | ||||
|     StringBuilder responseBuffer = new StringBuilder(); | ||||
|     try (BufferedReader reader = | ||||
|         new BufferedReader(new InputStreamReader(responseBodyStream, StandardCharsets.UTF_8))) { | ||||
|       String line; | ||||
|       while ((line = reader.readLine()) != null) { | ||||
|         if (statusCode == 404) { | ||||
|             LOG.warn("Status code: 404 (Not Found)"); | ||||
|           OllamaErrorResponseModel ollamaResponseModel = | ||||
|               Utils.getObjectMapper().readValue(line, OllamaErrorResponseModel.class); | ||||
|           responseBuffer.append(ollamaResponseModel.getError()); | ||||
|         } else if (statusCode == 401) { | ||||
|             LOG.warn("Status code: 401 (Unauthorized)"); | ||||
|           OllamaErrorResponseModel ollamaResponseModel = | ||||
|               Utils.getObjectMapper() | ||||
|                   .readValue("{\"error\":\"Unauthorized\"}", OllamaErrorResponseModel.class); | ||||
|           responseBuffer.append(ollamaResponseModel.getError()); | ||||
|         } else { | ||||
|           boolean finished = parseResponseAndAddToBuffer(line,responseBuffer); | ||||
|             if (finished) { | ||||
|               break; | ||||
|             } | ||||
|         } | ||||
|       } | ||||
|     } | ||||
|  | ||||
|     if (statusCode != 200) { | ||||
|         LOG.error("Status code " + statusCode); | ||||
|       throw new OllamaBaseException(responseBuffer.toString()); | ||||
|     } else { | ||||
|       long endTime = System.currentTimeMillis(); | ||||
|       OllamaResult ollamaResult = | ||||
|           new OllamaResult(responseBuffer.toString().trim(), endTime - startTime, statusCode); | ||||
|       if (verbose) LOG.info("Model response: " + ollamaResult); | ||||
|       return ollamaResult; | ||||
|     } | ||||
|   } | ||||
|  | ||||
|     /** | ||||
|    * Get default request builder. | ||||
|    * | ||||
|    * @param uri URI to get a HttpRequest.Builder | ||||
|    * @return HttpRequest.Builder | ||||
|    */ | ||||
|   private HttpRequest.Builder getRequestBuilderDefault(URI uri) { | ||||
|     HttpRequest.Builder requestBuilder = | ||||
|         HttpRequest.newBuilder(uri) | ||||
|             .header("Content-Type", "application/json") | ||||
|             .timeout(Duration.ofSeconds(this.requestTimeoutSeconds)); | ||||
|     if (isBasicAuthCredentialsSet()) { | ||||
|       requestBuilder.header("Authorization", getBasicAuthHeaderValue()); | ||||
|     } | ||||
|     return requestBuilder; | ||||
|   } | ||||
|  | ||||
|   /** | ||||
|    * Get basic authentication header value. | ||||
|    * | ||||
|    * @return basic authentication header value (encoded credentials) | ||||
|    */ | ||||
|   private String getBasicAuthHeaderValue() { | ||||
|     String credentialsToEncode = this.basicAuth.getUsername() + ":" + this.basicAuth.getPassword(); | ||||
|     return "Basic " + Base64.getEncoder().encodeToString(credentialsToEncode.getBytes()); | ||||
|   } | ||||
|  | ||||
|   /** | ||||
|    * Check if Basic Auth credentials set. | ||||
|    * | ||||
|    * @return true when Basic Auth credentials set | ||||
|    */ | ||||
|   private boolean isBasicAuthCredentialsSet() { | ||||
|     return this.basicAuth != null; | ||||
|   } | ||||
|      | ||||
| } | ||||
| @@ -0,0 +1,40 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.models.request; | ||||
|  | ||||
| import org.slf4j.Logger; | ||||
| import org.slf4j.LoggerFactory; | ||||
|  | ||||
| import com.fasterxml.jackson.core.JsonProcessingException; | ||||
|  | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.BasicAuth; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.OllamaResponseModel; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.Utils; | ||||
|  | ||||
| public class OllamaGenerateEndpointCaller extends OllamaEndpointCaller{ | ||||
|  | ||||
|     private static final Logger LOG = LoggerFactory.getLogger(OllamaGenerateEndpointCaller.class); | ||||
|  | ||||
|     public OllamaGenerateEndpointCaller(String host, BasicAuth basicAuth, long requestTimeoutSeconds, boolean verbose) { | ||||
|         super(host, basicAuth, requestTimeoutSeconds, verbose);    | ||||
|     } | ||||
|  | ||||
|     @Override | ||||
|     protected String getEndpointSuffix() { | ||||
|         return "/api/generate"; | ||||
|     } | ||||
|  | ||||
|     @Override | ||||
|     protected boolean parseResponseAndAddToBuffer(String line, StringBuilder responseBuffer) { | ||||
|                 try { | ||||
|                     OllamaResponseModel ollamaResponseModel = Utils.getObjectMapper().readValue(line, OllamaResponseModel.class); | ||||
|                     responseBuffer.append(ollamaResponseModel.getResponse()); | ||||
|                     return ollamaResponseModel.isDone(); | ||||
|                 } catch (JsonProcessingException e) { | ||||
|                     LOG.error("Error parsing the Ollama chat response!",e); | ||||
|                     return true; | ||||
|                 }          | ||||
|     } | ||||
|  | ||||
|      | ||||
|      | ||||
|      | ||||
| } | ||||
| @@ -0,0 +1,28 @@ | ||||
| package io.github.amithkoujalgi.ollama4j.core.utils; | ||||
|  | ||||
| import java.net.http.HttpRequest.BodyPublisher; | ||||
| import java.net.http.HttpRequest.BodyPublishers; | ||||
|  | ||||
| import com.fasterxml.jackson.annotation.JsonIgnore; | ||||
| import com.fasterxml.jackson.core.JsonProcessingException; | ||||
|  | ||||
| /** | ||||
|  * Interface to represent a OllamaRequest as HTTP-Request Body via {@link BodyPublishers}. | ||||
|  */ | ||||
| public interface OllamaRequestBody { | ||||
|      | ||||
|     /** | ||||
|      * Transforms the OllamaRequest Object to a JSON Object via Jackson. | ||||
|      *  | ||||
|      * @return JSON representation of a OllamaRequest | ||||
|      */ | ||||
|     @JsonIgnore | ||||
|     default BodyPublisher getBodyPublisher(){ | ||||
|                 try { | ||||
|           return BodyPublishers.ofString( | ||||
|                       Utils.getObjectMapper().writeValueAsString(this)); | ||||
|         } catch (JsonProcessingException e) { | ||||
|           throw new IllegalArgumentException("Request not Body convertible.",e); | ||||
|         } | ||||
|     } | ||||
| } | ||||
| @@ -5,7 +5,10 @@ import static org.junit.jupiter.api.Assertions.*; | ||||
| import io.github.amithkoujalgi.ollama4j.core.OllamaAPI; | ||||
| import io.github.amithkoujalgi.ollama4j.core.exceptions.OllamaBaseException; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.OllamaResult; | ||||
| import io.github.amithkoujalgi.ollama4j.core.types.OllamaModelType; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatMessageRole; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatRequestBuilder; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatRequestModel; | ||||
| import io.github.amithkoujalgi.ollama4j.core.models.chat.OllamaChatResult; | ||||
| import io.github.amithkoujalgi.ollama4j.core.utils.OptionsBuilder; | ||||
| import java.io.File; | ||||
| import java.io.IOException; | ||||
| @@ -16,26 +19,14 @@ import java.net.http.HttpConnectTimeoutException; | ||||
| import java.util.List; | ||||
| import java.util.Objects; | ||||
| import java.util.Properties; | ||||
| import lombok.Data; | ||||
| import org.junit.jupiter.api.BeforeEach; | ||||
| import org.junit.jupiter.api.Order; | ||||
| import org.junit.jupiter.api.Test; | ||||
|  | ||||
| class TestRealAPIs { | ||||
|   OllamaAPI ollamaAPI; | ||||
|  | ||||
|   private Properties loadProperties() { | ||||
|     Properties properties = new Properties(); | ||||
|     try (InputStream input = | ||||
|         getClass().getClassLoader().getResourceAsStream("test-config.properties")) { | ||||
|       if (input == null) { | ||||
|         throw new RuntimeException("Sorry, unable to find test-config.properties"); | ||||
|       } | ||||
|       properties.load(input); | ||||
|       return properties; | ||||
|     } catch (IOException e) { | ||||
|       throw new RuntimeException("Error loading properties", e); | ||||
|     } | ||||
|   } | ||||
|   Config config; | ||||
|  | ||||
|   private File getImageFileFromClasspath(String fileName) { | ||||
|     ClassLoader classLoader = getClass().getClassLoader(); | ||||
| @@ -44,9 +35,9 @@ class TestRealAPIs { | ||||
|  | ||||
|   @BeforeEach | ||||
|   void setUp() { | ||||
|     Properties properties = loadProperties(); | ||||
|     ollamaAPI = new OllamaAPI(properties.getProperty("ollama.api.url")); | ||||
|     ollamaAPI.setRequestTimeoutSeconds(20); | ||||
|     config = new Config(); | ||||
|     ollamaAPI = new OllamaAPI(config.getOllamaURL()); | ||||
|     ollamaAPI.setRequestTimeoutSeconds(config.getRequestTimeoutSeconds()); | ||||
|   } | ||||
|  | ||||
|   @Test | ||||
| @@ -85,10 +76,10 @@ class TestRealAPIs { | ||||
|   void testPullModel() { | ||||
|     testEndpointReachability(); | ||||
|     try { | ||||
|       ollamaAPI.pullModel(OllamaModelType.LLAMA2); | ||||
|       ollamaAPI.pullModel(config.getModel()); | ||||
|       boolean found = | ||||
|           ollamaAPI.listModels().stream() | ||||
|               .anyMatch(model -> model.getModelName().equals(OllamaModelType.LLAMA2)); | ||||
|               .anyMatch(model -> model.getModel().equalsIgnoreCase(config.getModel())); | ||||
|       assertTrue(found); | ||||
|     } catch (IOException | OllamaBaseException | InterruptedException | URISyntaxException e) { | ||||
|       throw new RuntimeException(e); | ||||
| @@ -102,7 +93,7 @@ class TestRealAPIs { | ||||
|     try { | ||||
|       OllamaResult result = | ||||
|           ollamaAPI.generate( | ||||
|               OllamaModelType.LLAMA2, | ||||
|               config.getModel(), | ||||
|               "What is the capital of France? And what's France's connection with Mona Lisa?", | ||||
|               new OptionsBuilder().build()); | ||||
|       assertNotNull(result); | ||||
| @@ -120,7 +111,7 @@ class TestRealAPIs { | ||||
|     try { | ||||
|       OllamaResult result = | ||||
|           ollamaAPI.generate( | ||||
|               OllamaModelType.LLAMA2, | ||||
|               config.getModel(), | ||||
|               "What is the capital of France? And what's France's connection with Mona Lisa?", | ||||
|               new OptionsBuilder().setTemperature(0.9f).build()); | ||||
|       assertNotNull(result); | ||||
| @@ -131,6 +122,46 @@ class TestRealAPIs { | ||||
|     } | ||||
|   } | ||||
|  | ||||
|   @Test | ||||
|   @Order(3) | ||||
|   void testChat() { | ||||
|     testEndpointReachability(); | ||||
|     try { | ||||
|       OllamaChatRequestBuilder builder = OllamaChatRequestBuilder.getInstance(config.getModel()); | ||||
|       OllamaChatRequestModel requestModel = builder.withMessage(OllamaChatMessageRole.USER, "What is the capital of France?") | ||||
|              .withMessage(OllamaChatMessageRole.ASSISTANT, "Should be Paris!") | ||||
|              .withMessage(OllamaChatMessageRole.USER,"And what is the second larges city?") | ||||
|              .build(); | ||||
|  | ||||
|       OllamaChatResult chatResult = ollamaAPI.chat(requestModel); | ||||
|       assertNotNull(chatResult); | ||||
|       assertFalse(chatResult.getResponse().isBlank()); | ||||
|       assertEquals(4,chatResult.getChatHistory().size()); | ||||
|     } catch (IOException | OllamaBaseException | InterruptedException e) { | ||||
|       throw new RuntimeException(e); | ||||
|     } | ||||
|   } | ||||
|  | ||||
|   @Test | ||||
|   @Order(3) | ||||
|   void testChatWithSystemPrompt() { | ||||
|     testEndpointReachability(); | ||||
|     try { | ||||
|       OllamaChatRequestBuilder builder = OllamaChatRequestBuilder.getInstance(config.getModel()); | ||||
|       OllamaChatRequestModel requestModel = builder.withMessage(OllamaChatMessageRole.SYSTEM, "You are a silent bot that only says 'NI'. Do not say anything else under any circumstances!") | ||||
|              .withMessage(OllamaChatMessageRole.USER,"What is the capital of France? And what's France's connection with Mona Lisa?") | ||||
|              .build(); | ||||
|  | ||||
|       OllamaChatResult chatResult = ollamaAPI.chat(requestModel); | ||||
|       assertNotNull(chatResult); | ||||
|       assertFalse(chatResult.getResponse().isBlank()); | ||||
|       assertTrue(chatResult.getResponse().startsWith("NI")); | ||||
|       assertEquals(3,chatResult.getChatHistory().size()); | ||||
|     } catch (IOException | OllamaBaseException | InterruptedException e) { | ||||
|       throw new RuntimeException(e); | ||||
|     } | ||||
|   } | ||||
|  | ||||
|   @Test | ||||
|   @Order(3) | ||||
|   void testAskModelWithOptionsAndImageFiles() { | ||||
| @@ -139,7 +170,7 @@ class TestRealAPIs { | ||||
|     try { | ||||
|       OllamaResult result = | ||||
|           ollamaAPI.generateWithImageFiles( | ||||
|               OllamaModelType.LLAVA, | ||||
|               config.getImageModel(), | ||||
|               "What is in this image?", | ||||
|               List.of(imageFile), | ||||
|               new OptionsBuilder().build()); | ||||
| @@ -158,7 +189,7 @@ class TestRealAPIs { | ||||
|     try { | ||||
|       OllamaResult result = | ||||
|           ollamaAPI.generateWithImageURLs( | ||||
|               OllamaModelType.LLAVA, | ||||
|               config.getImageModel(), | ||||
|               "What is in this image?", | ||||
|               List.of( | ||||
|                   "https://t3.ftcdn.net/jpg/02/96/63/80/360_F_296638053_0gUVA4WVBKceGsIr7LNqRWSnkusi07dq.jpg"), | ||||
| @@ -171,3 +202,29 @@ class TestRealAPIs { | ||||
|     } | ||||
|   } | ||||
| } | ||||
|  | ||||
| @Data | ||||
| class Config { | ||||
|   private String ollamaURL; | ||||
|   private String model; | ||||
|   private String imageModel; | ||||
|   private int requestTimeoutSeconds; | ||||
|  | ||||
|   public Config() { | ||||
|     Properties properties = new Properties(); | ||||
|     try (InputStream input = | ||||
|         getClass().getClassLoader().getResourceAsStream("test-config.properties")) { | ||||
|       if (input == null) { | ||||
|         throw new RuntimeException("Sorry, unable to find test-config.properties"); | ||||
|       } | ||||
|       properties.load(input); | ||||
|       this.ollamaURL = properties.getProperty("ollama.url"); | ||||
|       this.model = properties.getProperty("ollama.model"); | ||||
|       this.imageModel = properties.getProperty("ollama.model.image"); | ||||
|       this.requestTimeoutSeconds = | ||||
|           Integer.parseInt(properties.getProperty("ollama.request-timeout-seconds")); | ||||
|     } catch (IOException e) { | ||||
|       throw new RuntimeException("Error loading properties", e); | ||||
|     } | ||||
|   } | ||||
| } | ||||
|   | ||||
| @@ -1,2 +1,4 @@ | ||||
| ollama.api.url=http://192.168.29.223:11434 | ||||
| ollama.model=llava | ||||
| ollama.url=http://localhost:11434 | ||||
| ollama.model=qwen:0.5b | ||||
| ollama.model.image=llava | ||||
| ollama.request-timeout-seconds=120 | ||||
		Reference in New Issue
	
	Block a user