--- sidebar_position: 1 --- # Introduction Let's get started with **Ollama4j**. ## 🦙 What is Ollama? [Ollama](https://ollama.ai/) is an advanced AI tool that allows users to easily set up and run large language models locally (in CPU and GPU modes). With Ollama, users can leverage powerful language models such as Llama 2 and even customize and create their own models. ## 👨‍💻 Why Ollama4j? Ollama4j was built for the simple purpose of integrating Ollama with Java applications. ```mermaid flowchart LR o4j[Ollama4j] o[Ollama Server] o4j -->|Communicates with| o; m[Models] p[Your Java Project] subgraph Your Java Environment direction TB p -->|Uses| o4j end subgraph Ollama Setup direction TB o -->|Manages| m end ``` ## Getting Started ### What you'll need - **[Ollama](https://ollama.ai/download)** - **[Oracle JDK](https://www.oracle.com/java/technologies/javase/jdk11-archive-downloads.html)** or **[Open JDK](https://jdk.java.net/archive/)** 11.0 or above. - **[Maven](https://maven.apache.org/download.cgi)** ### Start Ollama server The easiest way of getting started with Ollama server is with [Docker](https://docs.docker.com/get-started/overview/). But if you choose to run the Ollama server directly, **[download](https://ollama.ai/download)** the distribution of your choice and follow the installation process. #### With Docker ##### Run in CPU mode: ```bash docker run -it -v ~/ollama:/root/.ollama -p 11434:11434 ollama/ollama ``` ##### Run in GPU mode: ```bash docker run -it --gpus=all -v ~/ollama:/root/.ollama -p 11434:11434 ollama/ollama ``` You can type this command into Command Prompt, Powershell, Terminal, or any other integrated terminal of your code editor. The command runs the Ollama server locally at **http://localhost:11434/**. ### Setup your project Get started by **creating a new Maven project** on your favorite IDE. Add the dependency to your project's `pom.xml`. ```xml io.github.amithkoujalgi ollama4j 1.0.27 ``` Find the latest version of the library [here](https://central.sonatype.com/artifact/io.github.amithkoujalgi/ollama4j). You might want to include an implementation of [SL4J](https://www.slf4j.org/) logger in your `pom.xml` file. For example, Use `slf4j-jdk14` implementation: ```xml org.slf4j slf4j-jdk14 2.0.9 ``` or use `logback-classic` implementation: ```xml ch.qos.logback logback-classic 1.3.11 ``` or use other suitable implementations. Create a new Java class in your project and add this code. ```java public class OllamaAPITest { public static void main(String[] args) { String host = "http://localhost:11434/"; OllamaAPI ollamaAPI = new OllamaAPI(host); ollamaAPI.setVerbose(true); boolean isOllamaServerReachable = ollamaAPI.ping(); System.out.println("Is Ollama server alive: " + isOllamaServerReachable); } } ```