Amazon Bedrock adds reinforcement fine-tuning simplifying how developers build smarter, more accurate AI models

By Dustin Ward

Organizations face a challenging trade-off when adapting AI models to their specific business needs: settle for generic models that produce average results, or tackle the complexity and expense of advanced model customization. Traditional approaches force a choice between poor performance with smaller models or the high costs of deploying larger model variants and managing complex…

New serverless customization in Amazon SageMaker AI accelerates model fine-tuning

By Dustin Ward

Today, I’m happy to announce new serverless customization in Amazon SageMaker AI for popular AI models, such as Amazon Nova, DeepSeek, GPT-OSS, Llama, and Qwen. The new customization capability provides an easy-to-use interface for the latest fine-tuning techniques like reinforcement learning, so you can accelerate the AI model customization process from months to days. With…

Introducing checkpointless and elastic training on Amazon SageMaker HyperPod

By Dustin Ward

Today, we’re announcing two new AI model training features within Amazon SageMaker HyperPod: checkpointless training, an approach that mitigates the need for traditional checkpoint-based recovery by enabling peer-to-peer state recovery, and elastic training, enabling AI workloads to automatically scale based on resource availability. Checkpointless training – Checkpointless training eliminates disruptive checkpoint-restart cycles, maintaining forward training…

Announcing replication support and Intelligent-Tiering for Amazon S3 Tables

By Dustin Ward

Today, we’re announcing two new capabilities for Amazon S3 Tables: support for the new Intelligent-Tiering storage class that automatically optimizes costs based on access patterns, and replication support to automatically maintain consistent Apache Iceberg table replicas across AWS Regions and accounts without manual sync. Organizations working with tabular data face two common challenges. First, they…