Logo
Back to Blog
AI & AutomationApril 29, 202614 min read

Amazon Bedrock AgentCore CLI & Managed Harness: Deploy AI Agents in 3 API Calls

AWS just shipped a managed agent harness and CLI for Amazon Bedrock AgentCore that lets you stand up production AI agents with 3 API calls — no orchestration code needed. We cover the CLI setup, harness architecture, prebuilt coding skills, pricing, and real-world deployment patterns.

Lushbinary Team

Lushbinary Team

AI & Cloud Solutions

Amazon Bedrock AgentCore CLI & Managed Harness: Deploy AI Agents in 3 API Calls

Building AI agents that actually work in production has always been the hard part. You pick a model, write some prompts, and then spend weeks building the infrastructure around it — session management, tool routing, rate limiting, observability, and graceful degradation when things break. On April 22, 2026, AWS shipped an update to Amazon Bedrock AgentCore that eliminates most of that plumbing.

The new managed agent harness and AgentCore CLI let you stand up a working autonomous agent through configuration rather than orchestration code. Three API calls. No backend scaffolding. The harness itself is free — you pay only for the compute your agent actually consumes, billed per-second with zero charges during I/O wait.

This guide covers everything you need to go from zero to a deployed agent: CLI installation, harness architecture, prebuilt coding skills, the MCP Gateway for tool connections, pricing breakdown, and production patterns we recommend at Lushbinary.

📋 What This Guide Covers

  1. What Changed on April 22, 2026
  2. AgentCore Architecture Overview
  3. Installing & Using the AgentCore CLI
  4. The Managed Agent Harness
  5. AgentCore Runtime: Per-Second Billing
  6. MCP Gateway & Tool Connections
  7. Prebuilt Coding Skills
  8. Pricing Breakdown
  9. Production Deployment Patterns
  10. AgentCore vs LangGraph vs CrewAI
  11. Why Lushbinary for AI Agent Development

1What Changed on April 22, 2026

AWS introduced three major additions to Amazon Bedrock AgentCore:

  • Managed Agent Harness — A pre-built infrastructure layer that handles orchestration, compute provisioning, tool connections, and session state. You configure an agent and invoke it; the harness does the rest.
  • AgentCore CLI — An npm-distributed command-line tool that scaffolds agent projects, deploys them to AgentCore Runtime, and invokes them directly from your terminal.
  • Prebuilt Coding Skills — Ready-to-use agent capabilities for code generation, review, and refactoring that plug into the harness without custom development.

🎤 AWS re:Invent 2025 Context

AgentCore was first announced at re:Invent 2025 (December 2, 2025) alongside shell command execution and managed session storage. The April 2026 update builds on that foundation with the managed harness and CLI, making the "3 API calls to production" workflow possible.

2AgentCore Architecture Overview

AgentCore is structured as a set of composable services you can use independently or together:

Developer (CLI / SDK / Console)Managed Agent Harness (FREE)Orchestration • Session State • Tool Routing • GuardrailsAgentCore Runtime (Per-Second Billing)Serverless Compute • microVM • I/O Wait = FreeMCP GatewayMemoryObservabilityModels: Claude Opus 4.7 • GPT-5.5 • Llama 4 • Custom EndpointsTools: MCP Servers • AWS Services • External APIs • Shell Commands

The key insight: the harness sits between your code and the runtime, handling all the orchestration logic that teams typically spend weeks building. You define what the agent should do; AgentCore handles how it runs.

3Installing & Using the AgentCore CLI

The CLI is distributed as an npm package. Prerequisites: Node.js (any recent LTS), Python 3.10+, and configured AWS credentials.

# Install globally

npm install -g @aws/bedrock-agentcore-cli

# Scaffold a new agent project

agentcore init my-support-agent

# Deploy to AgentCore Runtime

agentcore deploy

# Invoke the agent

agentcore invoke --prompt "Investigate the latency spike in us-east-1"

The init command generates a Python project with a pre-configured agent definition, tool bindings, and deployment configuration. The generated code uses the AgentCore SDK to define agent behavior:

# agent.py (generated by CLI)

from agentcore import Agent, Tool

agent = Agent(

model="anthropic.claude-opus-4-7",

instructions="You are a DevOps assistant...",

tools=[Tool.shell(), Tool.code_interpreter()],

timeout_seconds=3600,

max_steps=75,

)

4The Managed Agent Harness

The harness is the centerpiece of this update. It's a pre-built orchestration layer that handles:

  • Session management — Maintains conversation state, tool call history, and context across invocations without you managing a database
  • Tool orchestration — Routes tool calls to the right MCP servers, AWS services, or shell commands with automatic retry and error handling
  • Guardrails — Enforces token limits, step limits, and timeout boundaries to prevent runaway agents
  • Model routing — Supports any Bedrock model or custom inference profile, with the ability to switch models mid-session

The "3 API calls" workflow looks like this:

# 1. Create agent configuration

POST /agents

{ "model": "...", "tools": [...], "instructions": "..." }

# 2. Create a session

POST /agents/{agentId}/sessions

# 3. Invoke the agent

POST /agents/{agentId}/sessions/{sessionId}/invoke

{ "prompt": "Your task here" }

5AgentCore Runtime: Per-Second Billing

The runtime is where your agent code actually executes. It's a serverless environment built on microVMs with a billing model designed specifically for agentic workloads:

ResourceBilling ModelKey Detail
CPUActual consumption per secondZero CPU during I/O wait = $0
MemoryPeak memory per second128MB minimum billing
StorageS3 Standard rates (code) or ECR (containers)Managed session storage in preview
HarnessFreeNo extra charge for orchestration

This matters because agentic workloads typically spend 30-70% of their time in I/O wait — waiting for LLM responses, API calls, or database queries. With traditional compute (EC2, Lambda), you pay for that idle time. With AgentCore Runtime, idle time is free.

💡 Cost Implication

For a typical agent that spends 50% of its time waiting on LLM responses, AgentCore Runtime can be ~50% cheaper than running the same workload on Lambda or ECS. The savings scale with how I/O-heavy your agent is.

6MCP Gateway & Tool Connections

The AgentCore Gateway provides a centralized layer for managing how AI agents connect to tools and MCP servers across your organization. It consolidates authentication, observability, and policy enforcement into a single endpoint.

  • Centralized auth — OAuth 2.0 Authorization Code flow for MCP server connections, managed at the gateway level
  • Policy enforcement — Define which agents can access which tools, with IAM-based access control
  • Observability — Every tool call is logged with latency, success/failure, and token usage
  • Rate limiting — Protect downstream services from agent-driven traffic spikes

You register MCP servers with the gateway once, and any agent in your account can use them (subject to IAM policies). No per-agent configuration of tool endpoints.

7Prebuilt Coding Skills

The April update also shipped prebuilt coding skills that plug directly into the harness. These give agents the ability to:

  • Execute shell commands in a sandboxed environment (launched March 17, 2026)
  • Read and write files with persistent session storage (launched March 25, 2026)
  • Generate, review, and refactor code using the agent's model
  • Run test suites and interpret results

Combined with the managed harness, this means you can deploy a coding agent that clones a repo, makes changes, runs tests, and submits a PR — all without writing orchestration code yourself.

8Pricing Breakdown

AgentCore uses consumption-based pricing with no upfront commitments:

ComponentCost
Managed HarnessFree
AgentCore Runtime (CPU)Per-second, actual consumption only
AgentCore Runtime (Memory)Per-second, peak memory (128MB min)
MCP GatewayPer-request pricing
Agent RegistryFree during preview
Model inferenceStandard Bedrock model pricing

New AWS customers receive up to $200 in Free Tier credits. The "pay only for active compute" model makes AgentCore particularly cost-effective for agents that do a lot of waiting (which is most of them).

9Production Deployment Patterns

Based on our experience deploying AI agents for clients, here are the patterns that work best with AgentCore:

Pattern 1: DevOps Incident Responder

An agent that monitors CloudWatch alarms, investigates incidents by correlating logs and metrics, and posts findings to Slack. Uses shell commands to run diagnostic scripts and the MCP Gateway to connect to PagerDuty and Jira.

Pattern 2: Code Review Agent

Triggered by GitHub webhooks, this agent clones PRs, runs static analysis, executes tests, and posts review comments. The prebuilt coding skills handle file I/O and shell execution; the harness manages the multi-step workflow.

Pattern 3: Customer Support Triage

An agent connected to Zendesk via MCP that reads tickets, queries internal documentation, drafts responses, and escalates complex issues to humans. Session storage maintains context across multi-turn conversations.

🏗️ Architecture Tip

Use the managed harness for orchestration and AgentCore Runtime for compute, but keep your business logic in the agent code itself. The harness handles the "how" (routing, retries, state); your code defines the "what" (domain logic, decision criteria, output format).

10AgentCore vs LangGraph vs CrewAI

How does AgentCore compare to popular open-source agent frameworks?

FeatureAgentCoreLangGraphCrewAI
InfrastructureFully managedSelf-hosted or LangSmithSelf-hosted
BillingPer-second, active onlyCompute + LangSmith feesCompute only
Model supportAll Bedrock modelsAny (via LangChain)Any (via LiteLLM)
MCP nativeYes (Gateway)Via adaptersVia adapters
Session persistenceBuilt-inCheckpointingManual
Best forAWS-native teamsComplex graph workflowsMulti-agent teams

AgentCore's sweet spot is teams already on AWS who want managed infrastructure without vendor lock-in on the model layer. If you need complex graph-based workflows with conditional branching, LangGraph gives you more control. If you need multi-agent collaboration patterns, CrewAI is purpose-built for that. For a deeper comparison, see our agent framework comparison guide.

11Why Lushbinary for AI Agent Development

At Lushbinary, we've been building production AI agents on AWS since the early days of Bedrock. We help teams:

  • Design agent architectures that balance cost, latency, and reliability
  • Implement MCP Gateway integrations connecting agents to internal tools and data
  • Set up observability and guardrails for safe autonomous operation
  • Migrate from self-hosted frameworks (LangChain, CrewAI) to managed AgentCore
  • Build custom coding agents for CI/CD, code review, and incident response

🚀 Free Consultation

Want to deploy AI agents on AWS without the infrastructure headaches? Lushbinary specializes in AgentCore deployments and MCP integrations. We'll scope your agent architecture, recommend the right model and tool setup, and give you a realistic timeline — no obligation.

❓ Frequently Asked Questions

What is Amazon Bedrock AgentCore?

A managed platform for building, deploying, and operating AI agents at scale without infrastructure management. It includes a serverless runtime, MCP Gateway for tool connections, intelligent memory, and observability — all billed per-second based on actual resource consumption.

How much does Amazon Bedrock AgentCore cost?

The managed harness is free. You pay only for underlying resources: AgentCore Runtime bills per-second based on actual CPU consumption and peak memory (128MB minimum). I/O wait time is free. New AWS customers get up to $200 in Free Tier credits.

What is the AgentCore CLI and how do I install it?

The AgentCore CLI is an npm package that scaffolds agent projects, deploys them to AgentCore Runtime, and invokes them. Install with 'npm install -g @aws/bedrock-agentcore-cli'. Requires Node.js, Python 3.10+, and configured AWS credentials.

Can I deploy an AI agent with just 3 API calls?

Yes. The managed harness handles orchestration, compute, tool connections, and session state. You configure the agent (model, tools, instructions), create a session, and invoke it — 3 calls total. No orchestration code required.

What models does AgentCore support?

AgentCore works with any model available on Amazon Bedrock including Claude Opus 4.7, Claude Sonnet 4, GPT-5.5, GPT-5.4, Llama 4, and others. You can also bring your own model endpoint via custom inference profiles.

📺 Related re:Invent Sessions

📚 Sources

Content was rephrased for compliance with licensing restrictions. Pricing and feature details sourced from official AWS documentation as of April 2026. Pricing may change — always verify on the AWS pricing page.

Ready to Deploy AI Agents on AWS?

Let Lushbinary design and deploy your AgentCore-powered AI agents. From architecture to production — we handle the infrastructure so you can focus on business logic.

Ready to Build Something Great?

Get a free 30-minute strategy call. We'll map out your project, timeline, and tech stack — no strings attached.

Let's Talk About Your Project

Contact Us

Amazon BedrockAgentCoreAI AgentsAWS CLIManaged HarnessServerless AIAgentic AIMCP GatewayClaudeGPTAgent DeploymentAWS Infrastructure

ContactUs