Dify Supported Access Ollama Deployment of large-scale language model inference and embedding capabilities.
Quick Access
- Download Ollama
interviews Ollama Installation and ConfigurationFor more information, see the Ollama Local Deployment tutorial. - Run Ollama and Chat with Llama
ollama run llama3.1
After successful startup, ollama starts an API service on local port 11434, which can be accessed via the
http://localhost:11434
Access.
Additional models are available at https://ollama.ai/library for details. - Accessing Ollama in Dify
existSettings > Model Providers > Ollama
Fill in the blanks:
- Model Name:
llama3.1
- Base URL:
http://:11434
The address of the Ollama service that can be accessed is required here.
If Dify is deployed as a docker, it is recommended to fill in the LAN IP address, for example:http://192.168.1.100:11434
or the docker host IP address, for example:http://172.17.0.1:11434
TheNote: The Find LAN IP Address command:
- On Linux/macOS, use the command
ip addr show
maybeifconfig
The - On Windows, use the
ipconfig
command to find similar addresses. - Typically, this address is displayed under the eth0 or wlan0 interface, depending on whether you are using a wired or wireless network.
If deployed for local source code, fill in the
http://localhost:11434
The - On Linux/macOS, use the command
- Model Type:
dialogues
- Model context length:
4096
The maximum context length of the model, if not clear you can fill in the default value 4096. - greatest token Cap:
4096
The maximum number of tokens for the content returned by the model, which may be consistent with the model context length if not otherwise specified by the model. - Whether or not Vision is supported:
be
When the model supports picture understanding (multimodal) check this box, e.g.llava
The
Click "Save" to verify that the model is correct and can be used in your application.
The Embedding model is accessed in a similar way to LLM, by changing the model type to Text Embedding. - Model Name:
- Using the Ollama Model
Go to the App Prompt Programming page that you need to configure, and select the Ollama vendor under thellama3.1
model, configure the model parameters and use it.
FAQ
⚠️ If you are using Docker to deploy Dify and Ollama, you may encounter the following error.
httpconnectionpool(host=127.0.0.1, port=11434): max retries exceeded with url:/cpi/chat (Caused by NewConnectionError(': fail to establish a new connection:[Errno 111] Connection refused'))
httpconnectionpool(host=localhost, port=11434): max retries exceeded with url:/cpi/chat (Caused by NewConnectionError(': fail to establish a new connection:[Errno 111] Connection refused'))
This error is because the Docker container cannot access the Ollama service. The localhost usually refers to the container itself, not the host or other containers. To resolve this issue, you need to expose the Ollama service to the network.
Setting Environment Variables on a Mac
in the event that Ollama
act as macOS
application is running, then the following command should be used to set the environment variableslaunchctl
::
- This is accomplished by calling the
launchctl setenv
Setting environment variables:launchctl setenv OLLAMA_HOST "0.0.0.0"
- Restart the Ollama application.
- If the above steps do not work, you can use the following method:
The problem is that inside docker, you should connect to thehost.docker.internal
to access the docker's hosts, so set thelocalhost
Replace withhost.docker.internal
The service is ready to take effect:http://host.docker.internal:11434
Setting Environment Variables on Linux
If Ollama is running as a systemd service, you should use the systemctl
Setting environment variables:
- This is accomplished by calling the
systemctl edit ollama.service
Edit the systemd service. This will open an editor. - For each environment variable, the
[Service]
Add a line under the sectionEnvironment
::[Service] Environment="OLLAMA_HOST=0.0.0.0"
- Save and exit.
- heavy load (on a truck)
systemd
and restart Ollama:systemctl daemon-reload systemctl restart ollama
Setting Environment Variables on Windows
On Windows, Ollama inherits your user and system environment variables.
- First exit the program by clicking on Ollama in the taskbar.
- Edit system environment variables from the control panel
- Edit or create new variables for your user account, such as
OLLAMA_HOST
,OLLAMA_MODELS
etc. - Click OK/Apply to save the application
- In a new terminal window, run
ollama
How do I expose Ollama on my network?
Ollama binds to port 11434 on 127.0.0.1 by default. via the OLLAMA_HOST
Environment variables change the binding address.