Windows-MCP is an open-source MCP server that enables AI agents to control the Windows operating system, supporting tasks such as file navigation, application launching, and GUI automation. It features mouse/keyboard input tools, clipboard management, and window and app state detection. I tested Windows-MCP, one of the extensions listed in Anthropic Claude Desktop that can be installed using the free plan.
Reports indicate that ChatGPT Agent Mode might be capable of solving CAPTCHAs, which are often the final line of defense against persistent bots like AI crawlers on many websites. However, these claims originate from Reddit, known for the frequent hallucinations of its users. To verify, I decided to test it myself.
The new ChatGPT Agent Mode combines the Operator's ability to browse and interact with websites with Deep Research's ability to analyze and synthesize information from multiple sources. I tested ChatGPT's Agent Mode and was quite disappointed. Honestly, I find it surprising that OpenAI made this feature available to the public, considering that they have led the way with their chatbot so far. In my view, ChatGPT's Agent Mode is a flawed concept because it forces an AI agent to use a web browser, a relatively primitive tool designed for inferior human intelligence.
Are you skeptical about AI? Do you believe that AI is overhyped? Then, the latest episode of the Big Technology podcast is for you. AI champion Alex Kantrowitz and fierce AI critic Ed Zitron tackle the gap between AI hype and reality. However, both prominent journalists overlook an essential aspect when assessing AI's economic potential.
I tested three MCP servers in Anthropic Claude Sonnet 4: Claude Gmail Search, Zapier Gmail MCP server, and GongRzhe’s Gmail MCP Server. I also submitted the same search prompt to Gemini in Gmail. The results varied greatly. Only two AI tools passed my test, one failed, and the most promising MCP server turned out to be useless.
Anthropic has introduced two new concepts for connecting AI models to resources through MCP: desktop extensions and connectors. Extensions are available only on Claude Desktop (Windows and Mac), while connectors are supported on both the desktop and the Claude.ai web app. All desktop extensions are local MCP servers, but not all local MCP servers qualify as extensions. Conversely, all remote MCP servers are considered connectors; however, not all connectors are remote MCP servers. Although some theoretical documentation about these new concepts exists, I mainly categorized them based on how Claude Desktop organizes the different types of MCP servers.
VS Code version 1.102 introduced several useful new features in GitHub Copilot. Command allow/deny lists let you specify which commands the Copilot agent can execute without asking for permission. The resubmit feature enables you to edit a previous request and resubmit the prompt. Additionally, the new VS Code version includes various updates for managing MCP servers, including a curated catalog of MCP servers.
A new study by METR got a lot of attention online. It claims that developers often overestimate the productivity boost from AI tools. In a coding test, the researchers found that AI caused a 19% "slowdown" for developers. In my opinion, it's not AI coding that's overhyped, but rather the study's findings.
The number of GitHub Copilot concepts continues to grow, and their similar names can easily confuse. This post offers an overview of all the concepts, along with a link to find more information. The main confusion stems from the Copilot tools on GitHub.com and GitHub Copilot in VS Code. Although there is some overlap, keeping these two areas separate is essential.
Google's new Gemini CLI allows up to 60 model requests per minute and 1,000 daily requests. Professional developers might hit these limits quickly, but many IT admins will find the free plan sufficient. Compared to Anthropic Claude Code and OpenAI Codex CLI, I find Google's open-source AI-powered terminal more useful for the AI-augmented admin.
The brand-new Warp 2.0 no longer wants to be just an AI-powered terminal. The Warp team appears to be taking inspiration from Anthropic’s Claude Code and OpenAI’s Codex CLI. Both AI-driven coding tools target devs who prefer to code in a terminal instead of working in an Integrated Development Environment (IDE). Zach Lloyd, Warp's CEO, is eager to join the AI agent race, believing that features like the Agent Management Panel, diffs, and a built-in editor can transform a CLI tool into a vibe coding platform or what he calls an Agentic Development Environment (ADE).
Recently, leading AI companies announced support for remote MCP, allowing an MCP client to connect to a remote MCP server via Streamable HTTP, which still supports Server-Sent Events (SSE). The Streamable Data Input/Output (SDIO) protocol remains essential for local integrations and remote services, which are accessible only through conventional APIs. In my opinion, the excitement surrounding new model releases has lost its significance. AI model developers seem to primarily focus on optimizing their models for AI benchmarks. However, the genuine innovations occur in the technologies surrounding AI models—most notably in AI agents and the supporting infrastructure in which remote MCP servers will play an important role.
Microsoft Edit is a new command-line editor for Windows and Linux. Currently, you cannot install Microsoft Edit using apt-get on Ubuntu, so you must either download the latest version from GitHub or install it via snap. I will also briefly compare Microsoft Edit with GNU nano, the standard text editor in the Linux terminal. It turns out that Microsoft Edit has at least one crucial advantage over nano.
The VS Code May 2025 update (version 1.101), released a few days ago, introduces a range of new features primarily focused on GitHub Copilot. My favorite addition is the custom chat mode, which lets you create a personalized mode alongside the built-in options: Ask, Edit, and Agent. In this post, I'll also explore the new chat tool sets and the undoing edits feature. While the Simple Browser feature was included in the previous release, it deserves discussion. Furthermore, I will briefly summarize the MCP enhancements.
The MCP catalog in Docker Desktop has been available for a couple of weeks now. However, MCP support only works properly in the just-released Docker Desktop 4.42. The Docker Desktop MCP server catalog significantly simplifies installing MCP servers for any MCP client. I tested it with Gordon (Docker Desktop chatbot), Anthropic Claude Desktop, and VS Code GitHub Copilot on macOS, but the process should be identical on Windows.
AWS now allows automatic deletion of associated Amazon EBS snapshots when deregistering EC2 AMIs, reducing storage costs and simplifying cleanup. Previously, snapshots had to be manually deleted, risking orphaned resources. This feature is long overdue. In a previous post, I introduced a Bash script that deleted the snapshot when an AMI was deregistered. The script should still work, but now you can automatically delete snapshots in the AWS Console and the AWS CLI.
Warp, the top terminal app with AI support, has introduced new features in its GA release, including MCP support and Refine for AI responses. The Preview release now allows image uploads and the saving of reusable AI prompts. I also found a few new features that weren’t mentioned in the Launch Log 3 post.
Mountpoint for Amazon S3 (mountpoint-s3) now allows you to mount an S3 bucket in AWS at EC2 instance boot time via fstab. However, it remains non-POSIX compliant, so standard filesystem operations like deleting or renaming files still do not work. In contrast, the open-source alternative s3fs-fuse offers substantial POSIX support but comes with its own drawbacks.
Creating an AI agent in Amazon Bedrock involves a relatively complex process. The example in this post has been simplified for clarity. Once you have your first Agent setup working, exploring all the features that AI agents in AWS offer becomes much easier. The Bedrock Agent I discuss uses a Lambda function, which can be triggered by the AI model when the user prompt indicates its usefulness. An AWS Lambda function is a serverless compute service that runs your code in response to events without provisioning or managing servers. These functions may involve complex logic, enabling the AI agent to perform sophisticated tasks autonomously.