Running OpenAI gpt-oss-20b and gpt‑oss-120b in AWS on your own EC2 instance
OpenAI recently released its first open-weight models, gpt-oss-20b and gpt‑oss-120b, which you can download and run with Ollama. If security and privacy are essential when using AI models, you might not want to use the vendor's chat UI or API. The most powerful AI models require significant computing power and memory, and many desktop devices cannot run these models. Although Ollama was not designed with remote connections in mind, it is still possible. In this series, I will explain how to install OpenAI's open-weight models using Ollama and connect to them remotely from various applications, including VS Code GitHub Copilot. In this first post, I will outline how to choose the right size for your virtual machine, using EC2 instance types in AWS as an example.