OpenCLaw on AMD Developer Cloud: Free Deployment with Qwen 3.5 and SGLang

Mar 16, 2026

OpenClaw stands out because it gives users meaningful control over what their agent can do. Instead of prescribing a fixed experience, it allows you to enable the skills, tools, and capabilities that match your needs and your comfort level. In practice, that means the power of the agent is shaped by what you choose to make available to it.

That design makes OpenClaw interesting not just as an AI assistant, but as a flexible foundation for personal automation. With the right configuration, it can help automate daily tasks, support recurring routines, and turn familiar messaging interfaces into a more capable personal workflow layer.

This blog focuses on a practical way to get started: running OpenClaw for free with Qwen 3.5 and SGLang on AMD Developer Cloud using a single AMD Instinct™ MI300X GPU. It is a straightforward path for developers who want to experiment with a capable self-hosted agent stack on AMD hardware without a large upfront cost.

To address these concerns, this guide demonstrates how to use a powerful open-source model that exceeds the limits of typical consumer-grade GPUs by running it on enterprise data center hardware, specifically the AMD Instinct MI300X with a massive 192GB of memory, at no cost.

Complimentary AMD Developer Cloud Credits through the AMD AI Developer Program 

The AMD AI Developer Program provides members with $100 of free AMD Developer Cloud credits, which is enough for approximately 50 hours of usage to get started. Members who showcase useful applications and publicly share their projects may also apply for additional credits.

Beyond the credits, signing up grants you a one-month DeepLearning.AI PRO membership, automatic entry into monthly AMD hardware sweepstakes, and access to free AMD training courses.

Getting Started 

Step 1: Sign up for the AMD AI Developer Program

  • Existing Users: If you already have an AMD account, simply sign in and enroll in the program.

  • New Users: Select “Create an account” on the login page to set up your AMD account and enroll simultaneously.

Step 2: Activate Your Credits

After you join the program, navigate to the “AMD AI Developer Portal” and find your free credit activation link  on your profile page. 

Figure 1 AMD AI Developer Program Member Site Landing Page
Figure 1: AMD AI Developer Program Portal

Step 3: Create a GPU Droplet
Once signed in, you will be directed to the Create a GPU Droplet page, where your credits should appear at the top.

  1. Select Hardware: Choose a single MI300X instance.
  2. Select Image: Choose the ROCm Software image.
  3. Configure Access: Add your SSH key (instructions for creating one are provided in that section of the page).
Figure 2 Creating an AMD Developer Cloud GPU Droplet with MI300X
Figure 2: Creating an AMD Developer Cloud GPU Droplet with MI300X

Once the droplet is created, you can access it via your terminal or by clicking on Web Console button:

		ssh root@<YOUR_DROPLET_IP> 
	

Running the Model with SGLang

We will use SGLang, one of the most popular frameworks for LLMs, to run our model. The following steps must be followed in your droplet that you created in the previous step.

Step 1: Configure Firewall and Launch 
First, open port 8090 to allow traffic to your model:

ufw allow 8090

Step 2: Run the Model in SGLang Docker Container
In this example, we will pull the latest docker image for MI300X GPU and start a container to server Qwen3.5 model. Run the following commands on your droplet to start your model inference:

		docker run -d \ 
    --name sglang_server \ 
    --ipc=host \ 
    --privileged \ 
    --device=/dev/kfd \ 
    --device=/dev/dri \ 
    -p 8090:8090 \ 
    lmsysorg/sglang:v0.5.9-rocm700-mi30x \ 
    python3 -m sglang.launch_server \ 
    --model-path Qwen/Qwen3.5-122B-A10B-FP8 \ 
    --served-model-name qwen3-5-122b \ 
    --host 0.0.0.0 --port 8090 \ 
    --tp-size 1 \ 
    --api-key abc-123 \ 
    --attention-backend triton \ 
    --reasoning-parser qwen3 \ 
    --tool-call-parser qwen3_coder \ 
    --trust-remote-code 
	

Note: Replace abc-123 in the command below with a secure, unique API Key.

This server exposes an OpenAI-compatible endpoint that OpenClaw can connect to directly during installation. The address of your endpoint will be in this format: 

http://<your-droplet-ip>:8090/v1

Connecting OpenClaw to Your Local LLM

Now that the model is running, we will install OpenClaw and connect it directly to the SGLang endpoint. Recent versions of OpenClaw simplify this process significantly by allowing you to configure the model during the initial installation

Step 1: Install OpenClaw
Run the following command on your local machine or droplet (Mac/Linux):

curl -fsSL https://openclaw.ai/install.sh | bash

This launches the OpenClaw interactive onboarding process.
For all other OS installation options, please check OpenClaw’s official page. During installation, OpenClaw will guide you through a short onboarding process. For a minimal working setup, select the options shown in the screenshots below.

Step 2: Select QuickStart Mode
When prompted during installation, choose: Onboarding mode: QuickStart

OpenClaw Onboarding
Figure 3: OpenClaw Onboarding

Step 3: Configure the SGLang Model
When prompted for the model provider, select: Model/auth provider: SGLang

Figure 4: OpenClaw CLI onboarding showing model provider
Figure 4: OpenClaw CLI onboarding showing model provider

You will then be asked to enter the following values.

SGLang Base URL: http://<your-droplet-ip>:8090/v1
SGLang API Keyabc-123 (Replace this with the same API key you used when launching the SGLang server.)
SGLang Modelqwen3-5-122b (This must match the --served-model-name used when starting the SGLang server.)

Figure 5: OpenClaw CLI onboarding showing SGLang config
Figure 5: OpenClaw CLI onboarding showing SGLang config

Once entered, OpenClaw automatically configures the provider and sets the model as the default agent model.

Step 4: Skip Optional Features
For a minimal working setup, you can skip the remaining optional features during onboarding:

  • Chat channel integrations (Telegram, Slack, etc.)
  • Web search provider
  • Skills configuration
  • Hooks
OpenClaw onboarding prompts for optional features such as chat channels, web search, and agent skills
Figure 6: OpenClaw onboarding prompts for optional features such as chat channels, web search, and agent skills

These features can always be enabled later.

Step 5: Open the OpenClaw Dashboard
At the end of onboarding, OpenClaw prints a dashboard link similar to:
 http://127.0.0.1:18789/token=<generated-token>

If running on a remote droplet, create an SSH tunnel before opening any links in your browser:
ssh -N -L 18789:127.0.0.1:18789 root@<your-droplet-ip>

Then open in your browser:
http://127.0.0.1:18789/token=<generated-token>

 

OpenClaw dashboard access instructions and SSH tunneling for remote servers
Figure 7: OpenClaw dashboard access instructions and SSH tunneling for remote servers

From here you can start chatting with your OpenClaw agent powered by your SGLang-hosted Qwen 3.5 model.

OpenClaw dashboard showing successful connection to our model
Figure 8: OpenClaw dashboard showing successful connection to our model

Conclusion

You are now ready to start chatting with your agent running on enterprise-grade hardware. You can use this same method to run other open-source models from Hugging Face. Don't forget to share your projects with the community to earn more free GPU credits. Happy building!

Share:

Article By


Related Blogs