Remote HTTP access to self-hosted Ollama AI models
In my previous post in the Ollama series with OpenAI OSS models, I explained how to install Ollama on an Ubuntu EC2 instance. If you followed my guide, you now have everything set up to remotely access gpt-oss-20b on your self-hosted Ollama installation via HTTP. I will also showcase how to connect to remote AI models in VS Code using popular coding assistants like Cline and Roo Code, as well as through AI apps such as the local Ollama client, BoltAI, and Anything LLM.