Vue normale

Il y a de nouveaux articles disponibles, cliquez pour rafraîchir la page.
À partir d’avant-hierFlux principal

Install Ollama with OpenAI gpt-oss-20b on Ubuntu within an EC2 instance

19 août 2025 à 18:33
In my previous post, I explained why you might want to run OpenAI gpt-oss-20b or gpt-oss-120b on your own virtual machine and how to select an appropriate EC2 instance type. Today, I will show you how to install Ollama via script or snap on Ubuntu within an EC2 instance. Most steps can be adapted for other cloud providers, on-premises setups, or Linux distributions with minor modifications. I will also briefly discuss the benefits of running open-weight models on an EC2 instance with Ollama compared to Bedrock, SageMaker, vLLM, or Hugging Face Transformers.

Source

Running OpenAI gpt-oss-20b and gpt‑oss-120b in AWS on your own EC2 instance

18 août 2025 à 18:17
OpenAI recently released its first open-weight models, gpt-oss-20b and gpt‑oss-120b, which you can download and run with Ollama. If security and privacy are essential when using AI models, you might not want to use the vendor's chat UI or API. The most powerful AI models require significant computing power and memory, and many desktop devices cannot run these models. Although Ollama was not designed with remote connections in mind, it is still possible. In this series, I will explain how to install OpenAI's open-weight models using Ollama and connect to them remotely from various applications, including VS Code GitHub Copilot. In this first post, I will outline how to choose the right size for your virtual machine, using EC2 instance types in AWS as an example.

Source

❌
❌