Back to articles

AWS demonstrates building AI agents using Strands Agents SDK with models deployed on SageMaker AI endpoints and MLflow observability

Amazon AI Blog · April 27, 2026

AWS demonstrates building AI agents using Strands Agents SDK with models deployed on SageMaker AI endpoints and MLflow observability

AI Summary

  • AWS published a guide showing how to build AI agents by combining Strands Agents SDK (an open source SDK for building AI agents) with foundation models deployed on SageMaker AI endpoints, integrating them with SageMaker Serverless MLflow for agent tracing and A/B testing across model variants.
  • Organizations deploying models on SageMaker AI gain infrastructure control over compute instances, networking, and scaling; support for different models including custom fine-tuned or open-source alternatives like Llama or Mistral; and cost predictability through reserved instances and spot pricing—capabilities that managed foundation model services do not provide.
  • The post demonstrates deploying Qwen3-4B model from SageMaker JumpStart as a SageMaker AI endpoint, then creating a SageMaker AI Model provider within Strands Agents to run agents against the deployed endpoint with OpenAI-compatible chat completions APIs; a Jupyter notebook with complete code is available in the GitHub repo.

Related Articles

Stay ahead with AI news

Get curated AI news from 200+ sources delivered daily to your inbox. Free to use.

Get Started Free