Amazon Bedrock regularly releases new foundation model (FM) versions with better capabilities, accuracy, and safety. Understanding the model lifecycle is essential for effective planning and management of AI applications built on Amazon Bedrock. Before migrating your applications, you can test these models through the Amazon Bedrock console or API to evaluate their performance and compatibility.
This post shows you how to manage FM transitions in Amazon Bedrock, so you can make sure your AI applications remain operational as models evolve. We discuss the three lifecycle states, how to plan migrations with the new extended access feature, and practical strategies to transition your applications to newer models without disruption.
A model offered on Amazon Bedrock can exist in one of three states: Active, Legacy, or End-of-Life (EOL). Their current status is visible both on the Amazon Bedrock console and in API responses. For example, when you make a GetFoundationModel or ListFoundationModels call, the state of the model will be shown in the modelLifecycle field in the response.
The following diagram illustrates the details around each model state.

The state details are as follows:
Active, you can use it for inference through APIs like InvokeModel or Converse, customize it (if supported), and request quota increases through AWS Service Quotas.Legacy state, Amazon Bedrock will notify customers with at least 6 months’ advance notice before the EOL date, providing essential time to plan and execute a migration to newer or alternative model versions. During the Legacy period, existing customers can continue using the model, though new customers might be unable to access it, and existing customers might lose access for inactive accounts if they do not call the model for a period of 15 days or more. Organizations should note that creating new provisioned throughput by model units becomes unavailable, and model customization capabilities might face restrictions. For models with EOL dates after February 1, 2026, Amazon Bedrock introduces an additional phase within the Legacy state:
Legacy status, the model enters this extended access phase. Active users can continue using it for at least another 3 months until EOL. During extended access, quota increase requests through AWS Service Quotas are not expected to be approved, so plan your capacity needs before a model enters this phase. During this period, pricing may be adjusted (see Pricing during extended access below), and customers will receive notifications about the transition date and any changes.After a model launches on Amazon Bedrock, it remains available for at least 12 months after launch and stays in Legacy state for at least 6 months before EOL. This timeline helps customers plan migrations without rushing.
During the extended access period, pricing may be adjusted by the model provider. If pricing changes are planned, you will be notified in the initial legacy announcement and before any subsequent changes take effect, so there will be no surprise retroactive price increases. Customers with existing private pricing agreements with model providers or those using provisioned throughput will continue to operate under their current pricing terms during the extended access period. This makes sure customers who have made specific arrangements with model providers or invested in provisioned capacity will not be unexpectedly affected by any pricing changes.
Customers will receive a notification 6 months prior to a model’s EOL date when the model provider transitions a model to Legacy state. This proactive communication approach ensures that customers have sufficient time to plan and execute their migration strategies before a model becomes EOL.
Notifications include details about the model being deprecated, important dates, extended access availability, and when the model will be EOL. AWS uses multiple channels to ensure these important communications reach the right people, including:

To make sure you receive these notifications, verify and configure your account contact email addresses. By default, notifications are sent to your account’s root user email and alternate contacts (operations, security, and billing). You can review and update these contacts on your AWS Account page in the Alternate contacts section. To add additional recipients or delivery channels (such as Slack or email distribution lists), go to the AWS User Notifications console and choose AWS managed notifications subscriptions to manage your delivery channels and account contacts. If you are not receiving expected notifications, check that your email addresses are correctly configured in these settings and that notification emails from [email protected] are not being filtered by your email provider.
When migrating to a newer model, update your application code and check that your service quotas can handle expected volume. Planning ahead helps you transition smoothly with minimal disruption.
Start planning as soon as a model enters Legacy state:
Test your migration thoroughly:
anthropic.claude-3-5-sonnet-20240620-v1:0 to anthropic.claude-sonnet-4-5-20250929-v1:0 or global cross-Region inference global.anthropic.claude-sonnet-4-5-20250929-v1:0. Update prompt structures according to new model’s best practices. For more detailed guidance, refer to Migrate from Anthropic’s Claude Sonnet 3.x to Claude Sonnet 4.x on Amazon Bedrock.Thorough testing is critical for a successful migration:
The model lifecycle policy in Amazon Bedrock gives you clear stages for managing FM evolution. Transition periods offer extended access options, and provisions for fine-tuned models help you balance innovation with stability.
Stay informed about model states through the AWS Health Dashboard, plan migrations when models enter the Legacy state, and test newer versions thoroughly. These guidelines can help you maintain continuity in your AI applications while using improved capabilities in newer models.
If you have further questions or concerns, reach out to your AWS team. We want to help you and facilitate a smooth transition as you continue to take advantage of the latest advancements in FM technology.
For continued learning and implementation support, explore the official AWS Bedrock documentation for comprehensive guides and API references. Additionally, visit the AWS Machine Learning Blog and AWS Architecture Center for real-world case studies, migration best practices, and reference architectures that can help optimize your model lifecycle management strategy.
Saurabh Trikande is a Senior Product Manager for Amazon Bedrock and Amazon SageMaker Inference. He is passionate about working with customers and partners, motivated by the goal of democratizing AI. He focuses on core challenges related to deploying complex AI applications, inference with multi-tenant models, cost optimizations, and making the deployment of generative AI models more accessible. In his spare time, Saurabh enjoys hiking, learning about innovative technologies, following TechCrunch, and spending time with his family.
Melanie Li, PhD, is a Senior Generative AI Specialist Solutions Architect at AWS based in Sydney, Australia, where her focus is on working with customers to build solutions using state-of-the-art AI/ML tools. She has been actively involved in multiple generative AI initiatives across APJ, harnessing the power of LLMs. Prior to joining AWS, Dr. Li held data science roles in the financial and retail industries.
Derrick Choo is a Senior Solutions Architect at AWS who accelerates enterprise digital transformation through cloud adoption, AI/ML, and generative AI solutions. He specializes in full-stack development and ML, designing end-to-end solutions spanning frontend interfaces, IoT applications, data integrations, and ML models, with a particular focus on computer vision and multi-modal systems.
Jared Dean is a Principal AI/ML Solutions Architect at AWS. Jared works with customers across industries to develop machine learning applications that improve efficiency. He is interested in all things AI, technology, and BBQ.
Julia Bodia is Principal Product Manager for Amazon Bedrock.
Pooja Rao is a Senior Program Manager at AWS, leading quota and capacity management and supporting business development for the Bedrock Go-To-Market team. Outside of work, she enjoys reading, traveling, and spending time with her family.
Manuel Rioux est fièrement propulsé par WordPress