Public Ollama Models
Public Ollama Models 20250525
How to chat with Ollama models
Select an IP and model from the table below, then use them in this command:
# Start a conversation with a model
# Replace <IP> with an IP from the table below
# Replace <MODEL> with one of the models listed for that IP
curl -X POST http://<IP>:11434/api/chat -d '{
"model": "<MODEL>",
"messages": [{
"role": "user",
"content": "Hello, how are you?"
}]
}'
Available Models
IP | Models |
---|---|
123.60.64.205 | llama3.2:3b-instruct-q5_K_M deepseek-r1:1.5b |
101.132.74.29 | llama3.2:3b-instruct-q5_K_M |
139.196.196.43 | qwen3:32b qwen3:30b-a3b qwen2.5:72b-instruct deepseek-r1:70b |
223.166.61.66 | qwen3:4b qwen3:32b mxbai-embed-large:latest bge-m3:latest nomic-embed-text:latest qwen2.5-coder:32b qwen2.5-coder:14b qwen2.5:32b qwen2.5:latest mistral-small3.1:latest codellama:13b codellama:34b codellama:70b qwq:latest deepseek-v2:16b deepseek-coder-v2:latest gemma3:12b gemma3:latest qwen2.5:3b gemma3:1b deepseek-r1:14b deepseek-r1:32b deepseek-r1:8b deepseek-r1:7b deepseek-r1:1.5b |
106.75.235.214 | llama3.1:8b gemma3:27b qwq:latest deepseek-r1:70b |
117.50.171.129 | qwq:latest qwen2.5-coder:32b |
117.50.180.35 | bge-m3:latest |
116.232.26.17 | qwen3:0.6b qwen3:1.7b |
117.50.176.34 | qwen2.5:1.5b qwen2.5:7b qwen2.5:14b |
222.64.131.253 | llama3.2:3b-instruct-q5_K_M qwen2.5:32b-instruct-q4_K_M qwq:32b-q4_K_M qwen2.5:32b qwen2.5-coder:14b sqlcoder:15b deepseek-r1:14b qwen2.5:14b bge-m3:567m qwen2.5:7b nomic-embed-text:latest qwen2.5:0.5b |
123.60.5.218 | mario:latest llama3-backup:latest llama3.2:latest llama3.2:3b-instruct-q5_K_M deepseek-r1:8b |
218.1.223.134 | llama3.2:3b-instruct-q5_K_M smollm2:135m deepseek-r1:1.5b deepseek-r1:7b deepseek-r1:8b |
101.132.88.19 | nomic-embed-text:latest llama3.2:3b llama3.2:1b deepseek-r1:1.5b |
114.95.209.211 | nomic-embed-text:latest qwen2.5-coder:latest deepseek-r1:7b |
Disclaimer
These Ollama model endpoints are publicly exposed interfaces found on the internet. They are listed here for informational purposes only. Please be aware that:
- These endpoints are not maintained or controlled by us
- The availability and stability of these services cannot be guaranteed
- Use these services at your own risk
- We take no responsibility for any issues or damages that may arise from using these endpoints
免责声明
本文列出的 Ollama 模型接口均来自互联网上公开暴露的端点。请注意:
- 这些端点并非由我们维护或控制
- 无法保证这些服务的可用性和稳定性
- 使用这些服务需自行承担风险
- 对于使用这些端点可能产生的任何问题或损失,我们不承担任何责任