Host Your Own Coding Agents with OpenHands using AMD Ryzen AI

8 min read

Written by

Xingyao Wang, Graham Neubig

Published on

November 19, 2025

Coding agents like OpenHands can automate many software development tasks. However, running them typically requires expensive cloud APIs or dedicated data center infrastructure, raising concerns about cost and privacy. An alternative is to run coding agents directly on your workstation. With advances in open-weight models and edge AI processing, smaller locally-runnable models are now achieving competitive results—approaching their larger counterparts on benchmarks like SWE-Bench Verified—making self-hosted coding agents a viable option.

The AMD Stack for Language Model Serving

The AMD Ryzen™ AI Max+ 395 processor combines 16 Zen 5 CPU cores with a Radeon™ 8060S integrated GPU and an XDNA 2 Neural Processing Unit (NPU). Together, these components deliver up to 126 AI TOPS (Tera Operations Per Second), providing the compute needed to run language models locally. The processor includes 64 MB of L3 cache and supports high-bandwidth LPDDR5X memory, enabling inference with models like Qwen3-Coder-30B.

To utilize this hardware, AMD has developed Lemonade Server, a software stack that enables local large language models to run with NPU and GPU acceleration. Lemonade supports the OpenAI API standard, making it compatible with existing AI applications and tools. Lemonade provides accelerated and efficient coding agent execution by leveraging the integrated NPU, GPU, and CPU on the Ryzen AI™. The server works across platforms and supports popular inference engines, including optimized backends for AMD hardware.

By combining the Ryzen AI Max+ 395's hardware with Lemonade's software stack, developers can run models like Qwen3-Coder-30B locally while maintaining reasonable performance. This approach enables private, cost-effective AI development for those who prefer to keep their code on-premises.

Setting Up AMD Ryzen AI with Lemonade

AMD and OpenHands are collaborating to make it easier to run coding agents locally on AMD hardware. Here's how to get started:

Prerequisites

Before starting, ensure you have:

• A compatible Linux or Windows environment
• Administrative access to install required software

Step 1: Install Lemonade Prerequisites

For detailed installation instructions, see the Lemonade Getting Started guide. On Windows, Lemonade works out of the box. On Linux, ensure that the ROCm tools are installed before proceeding.

Step 2: Install Lemonade

Lemonade is AMD's optimized software stack for running large language models on AMD hardware. To install and run Lemonade, first download and install Lemonade then run


lemonade-server serve --host 0.0.0.0 --ctx-size 32768

When you start the Lemonade server, you'll see output like this:


🍋 Open http://0.0.0.0:8000 in your browser for:
🍋 💬 chat
🍋 💻 model management
🍋 📄 docs

The Lemonade server will automatically download the Qwen3-Coder-30B-A3B-Instruct model (approximately 18.6GB) on first run.

Step 3: Install OpenHands

There are many ways to use OpenHands through the CLI, GUI, and SDK and for detailed instructions you can see the OpenHands documentation. Here, we'll use the CLI, which is easy to install and start using through uv:


uvx tool install openhands

This will start the OpenHands interactive command-line interface.

Step 4: Configure OpenHands to Use Lemonade

To connect OpenHands to your local Lemonade server, you'll need to configure the LLM settings. In the OpenHands CLI:

1. Type /settings and press Enter to access the configuration menu

2. Select "LLM (Basic)" to configure a custom LLM endpoint

3. When prompted, enter the following configuration:

Provider: lemonade

Model: Qwen3-Coder-30B-A3B-Instruct-GGUF

4. Confirm to save the settings

Your OpenHands CLI is now configured to use your local Lemonade server! The settings are persisted for all future sessions.

Step 5: Start Your First Conversation

Back at the main CLI menu, start a new conversation by typing your request. For example:

"Create a website that showcases Ryzen AI and the ability to run the OpenHands coding agents locally through the Lemonade software stack. Make the website fun with a theme of lemonade and laptops."

The OpenHands agent will begin working on your task using the local Qwen model running on your AMD hardware. You'll see the agent writing code, testing, and iterating until everything is complete.

Once the agent completes its work, you can view and interact with the generated code in the working directory.

Why Self-Host with AMD and OpenHands?

When using coding agents, many developers rely on powerful but closed API-based models like Claude and GPT. While these models are effective, there are compelling reasons to use open-weight models and host them yourself:

1. Privacy: Your code and intellectual property never leave your machine

2. Cost: No per-token API costs, especially beneficial for long-running tasks

3. Compliance: Meet regulatory requirements for government, healthcare, and finance

4. Performance: Low latency with local inference

5. Flexibility: Full control over model selection and configuration

6. Offline Capability: Work without internet connectivity

Conclusion

Going forward, at OpenHands we're working on developing additional local model capabilities and methods for you to adapt models on your own data using AMD Ryzen AI.

We hope that this post has been useful, and if you'd like to try setting things up yourself, please join our Slack community or check out our documentation for more information. And thanks to AMD for the collaboration on this integration!

Citation
Host Your Own Coding Agents with OpenHands using AMD Ryzen AI

Get useful insights in our blog

Insights and updates from the OpenHands team

Thank you for your submission!

Oops! Something went wrong while submitting the form.
Building the open standard for autonomous software development.

OpenHands is the foundation for secure, transparent, model-agnostic coding agents - empowering every software team to build faster with full control.

Build with SDK
Try it live
© 2025 OpenHands - All rights reserved