2025 - Week 33 (11 Aug - 17 Aug)

2025 - Week 33 (11 Aug - 17 Aug)

Amazon Athena now supports CREATE TABLE AS SELECT with Amazon S3 Tables

Published Date: 2025-08-15 18:44:00

Amazon Athena now supports CREATE TABLE AS SELECT (CTAS) statements with Amazon S3 Tables. Using CTAS statements makes it simple to create a new table and populate it with data using the results of a SELECT query. You can now use CTAS statements in Athena to query existing datasets and create a new table in S3 Tables with the query results, all in a single SQL statement. S3 Tables deliver the first cloud object store with built-in Apache Iceberg support and streamline storing tabular data at scale. With today’s launch, you can quickly and efficiently convert existing datasets stored in Parquet, CSV, JSON, and other formats, including Apache Iceberg, Hudi, and Delta Lake, into fully-managed tables that are continually optimized for performance and cost. Once created, use Athena to analyze your data, JOIN it with other datasets, and evolve it over time using INSERT and UPDATE operations. Using CTAS, you can partition the data on the fly, giving you flexibility to optimize query performance for different use cases. You can use CTAS to create S3 Tables in all AWS Regions where both Athena and S3 Tables are supported. To learn more, see the Amazon Athena User Guide.

Amazon EC2 R8g instances now available in AWS Asia Pacific (Jakarta)

Published Date: 2025-08-15 18:03:00

Starting today, Amazon Elastic Compute Cloud (Amazon EC2) R8g instances are available in AWS Asia Pacific (Jakarta)region. These instances are powered by AWS Graviton4 processors and deliver up to 30% better performance compared to AWS Graviton3-based instances. Amazon EC2 R8g instances are ideal for memory-intensive workloads such as databases, in-memory caches, and real-time big data analytics. These instances are built on the AWS Nitro System, which offloads CPU virtualization, storage, and networking functions to dedicated hardware and software to enhance the performance and security of your workloads. AWS Graviton4-based Amazon EC2 instances deliver the best performance and energy efficiency for a broad range of workloads running on Amazon EC2. AWS Graviton4-based R8g instances offer larger instance sizes with up to 3x more vCPU (up to 48xlarge) and memory (up to 1.5TB) than Graviton3-based R7g instances. These instances are up to 30% faster for web applications, 40% faster for databases, and 45% faster for large Java applications compared to AWS Graviton3-based R7g instances. R8g instances are available in 12 different instance sizes, including two bare metal sizes. They offer up to 50 Gbps enhanced networking bandwidth and up to 40 Gbps of bandwidth to the Amazon Elastic Block Store (Amazon EBS). To learn more, see Amazon EC2 R8g Instances. To explore how to migrate your workloads to Graviton-based instances, see AWS Graviton Fast Start program and Porting Advisor for Graviton. To get started, see the AWS Management Console.

Amazon DynamoDB now supports a CloudWatch Contributor Insights mode exclusively for throttled keys

Published Date: 2025-08-15 16:00:00

DynamoDB now supports the ability to selectively emit events for throttled keys to CloudWatch Contributor Insights, enabling you to monitor throttled keys without emitting events for all accessed keys. By emitting events for throttled keys exclusively, you no longer need to pay for all of your successful request events. Cloudwatch Contributor Insights for DynamoDB can help you understand your traffic patterns by providing information about your most accessed and throttled keys in a table or global secondary index. This information can be used to understand your application usage patterns or diagnose throttling-related issues. By choosing to only emit events for throttled keys, you can reduce the amount you spend to receive these insights. The new mode to exclusively emit throttled key events to CloudWatch Contributor Insights is available in all commercial AWS Regions, the AWS GovCloud (US) Regions, and the China Regions. To get started, see the following list of resources:

Amazon DynamoDB now supports more granular throttle error exceptions

Published Date: 2025-08-15 16:00:00

DynamoDB now supports more granular throttling exceptions along with their corresponding Amazon CloudWatch metrics. The additional fields in the new throttling exceptions identify the specific resources and reasons for throttling events, making it easier to understand and diagnose throttling-related issues. You can see the new Amazon CloudWatch metrics immediately, and upon upgrading your SDK to the newest version, you will also see the new granular throttling exceptions. Every throttling exception now contains a list of reasons why the request was throttled, as well as the Amazon Resource Name (ARN) of the table or index that was throttled. These new throttle exception reasons help you understand why you were throttled and enable you to take corrective actions like adjusting your configured throughput, switching your table to on-demand capacity mode, or optimizing data access patterns. The more granular throttling exceptions and their respective metrics are available in all commercial AWS Regions, the AWS GovCloud (US) Regions, and the China Regions. To get started see the following list of resources:

AWS Certificate Manager supports AWS PrivateLink

Published Date: 2025-08-15 15:00:00

AWS Certificate Manager (ACM) now supports AWS PrivateLink so that you can access ACM APIs from your Amazon Virtual Private Cloud (VPC) without traversing the public internet. This feature can help you meet compliance requirements by allowing you to access and use ACM APIs entirely within the AWS network. ACM simplifies the process of provisioning and managing public and private TLS certificates, wherever you need to securely terminate traffic; Whether it’s with integrated AWS services such as Amazon CloudFront, Load Balancing or with hybrid workloads. You can now create interface endpoints in AWS Private Link to connect your VPC to ACM. Communication between your VPC and ACM is then conducted entirely within the AWS network, providing a secure pathway for your data. To get started, you can create an AWS PrivateLink to connect to ACM using the AWS Management Console or AWS Command Line Interface (AWS CLI) commands or AWS CloudFormation. This new feature is available in all AWS Regions including AWS GovCloud (US) and China Regions where AWS Certificate Manager Service and AWS PrivateLink are available. For more information, please refer to the AWS PrivateLink documentation.

Amazon Managed Service for Prometheus adds support resource policies

Published Date: 2025-08-15 13:30:00

Amazon Managed Service for Prometheus, a fully managed Prometheus-compatible monitoring service, now supports resource-based policies, making it easier to build applications that work across accounts. With resource-based policies, you can specify which Identity and Access Management (IAM) principals have access to ingest or query your Amazon Managed Service for Prometheus workspace.

To allow cross-account ingestion into an Amazon Managed Service for Prometheus workspace or query the metrics using PromQL from a different account, customers so far had to assume an IAM role in the workspace owner account. With this launch, you now can attach a resource-based policy to an Amazon Managed Service for Prometheus workspace and allow-list non-workspace owner to perform any actions using Prometheus-compatible APIs.

This feature is now available in all regions where Amazon Managed Service for Prometheus is generally available.

To learn more about Amazon Managed Service for Prometheus collector, visit the user guide or product page.

Amazon Neptune now integrates with Cognee for graph-native memory in GenAI Applications

Published Date: 2025-08-15 13:00:00

Today, we’re announcing the integration of Amazon Neptune Analytics with Cognee, a leading agentic memory framework designed to help AI agents structure, retrieve, and reason over information. With this launch, customers can use Neptune as the graph store behind Cognee’s memory layer, enabling long-term memory and reasoning capabilities for agentic AI applications. This integration allows Cognee users to store and query memory graphs at scale, unlocking advanced use cases where AI agents become more personalized and effective over time by learning from ongoing interactions. Neptune supports multi-hop graph reasoning and hybrid retrieval across graph, vector, and keyword modalities—helping Cognee deliver richer, more context-aware AI experiences. Cognee enables a self-improving memory system that helps developers build cost-efficient, personalized generative AI applications. To learn more about the Neptune–Cognee integration, visit the User Guide and the sample notebook.

Amazon RDS for Db2 now supports cross-region automated backups for encrypted databases

Published Date: 2025-08-15 13:00:00

Amazon Relational Database Service (RDS) for Db2 now supports cross-region automated backups for encrypted databases, providing customers with an additional layer of data protection while safeguarding their mission critical Db2 workloads against regional outages. Customers can now securely copy encrypted database snapshots to regions outside of their primary AWS region for improved disaster recovery. The feature can be enabled by simply turning on encryption for RDS for Db2 instances, and configuring backup replication to the desired AWS region. To learn more about Amazon RDS for Db2 cross-region automated backups and the supported destination regions, visit the documentation page. Amazon RDS for Db2 makes it simple to set up, operate, and scale Db2 deployments in the cloud. See Amazon RDS for Db2 Pricing for up-to-date pricing of instances, storage, backup, data transfer, and regional availability.

AWS Billing and Cost Management Console adds new recommended actions

Published Date: 2025-08-15 13:00:00

Starting today, customers can view 6 new recommended actions added to the existing list of 15 recommended actions available in the AWS Billing and Cost Management Console recommended actions widget. These recommended actions include notifications across AWS payments and tax settings, such as an expired payment method or if tax registration numbers are invalid. All recommended actions are now categorized as critical, advisory, or informational, enabling customers to prioritize and timely resolve any identified billing issues.  Using recommended actions on the AWS Billing and Cost Management Console, customers can quickly learn of and mitigate AWS billing or payment issues, identify cost saving opportunities, and avoid surprises by acting on time-sensitive information. Each recommended action includes a specific call-to-action, which allows customers to optimize their AWS spend and prevent any disruptions to their AWS account and billing status. Customers can access these recommendation actions through the recommended actions widget in the console or via a new public API at no additional cost. The recommended actions widget and API is available in all AWS commercial regions, excluding China. To get started with new recommended actions, visit the recommended actions widgets on the AWS Billing and Cost Management Console. To learn more, see recommended actions in the AWS Billing and Cost Management user guide

AWS Managed Microsoft AD increases directory sharing limits

Published Date: 2025-08-15 07:00:00

AWS has increased the account sharing limits for AWS Managed Microsoft AD directory sharing, allowing customers to share their directories with significantly more AWS accounts. The Standard Edition limit has increased from 5 to 25 accounts, while the Enterprise Edition limit has expanded from 125 to 500 accounts. These enhanced limits remove previous technical constraints and enable organizations to scale their directory infrastructure more effectively across their AWS environments. The increased limits help enterprise customers consolidate their Active Directory infrastructure and reduce operational complexity by supporting larger AWS account footprints from a single managed directory. Organizations can now centralize authentication and management across hundreds of AWS accounts, which in turn helps eliminate the need for complex workarounds with multiple directory deployments. This feature enhancement is available in all AWS Regions where AWS Managed Microsoft AD is currently supported. To learn more about AWS Managed Microsoft AD directory sharing, see the AWS Directory Service documentation . For detailed information about directory sharing capabilities and setup, visit the AWS Managed Microsoft AD directory sharing page . For regional availability, see the AWS Region table.

Amazon RDS for MariaDB now supports community MariaDB minor versions 11.4.8, 10.11.14 and 10.6.23

Published Date: 2025-08-15 07:00:00

Amazon Relational Database Service (Amazon RDS) for MariaDB now supports community MariaDB minor versions 11.4.8, 10.11.14 and 10.6.23. We recommend that you upgrade to the latest minor versions to fix known security vulnerabilities in prior versions of MariaDB, and to benefit from the bug fixes, performance improvements, and new functionality added by the MariaDB community. You can leverage automatic minor version upgrades to automatically upgrade your databases to more recent minor versions during scheduled maintenance windows. You can also leverage Amazon RDS Managed Blue/Green deployments for safer, simpler, and faster updates to your MariaDB instances. Learn more about upgrading your database instances, including automatic minor version upgrades and Blue/Green Deployments, in the Amazon RDS User Guide. Amazon RDS for MariaDB makes it straightforward to set up, operate, and scale MariaDB deployments in the cloud. Learn more about pricing details and regional availability at Amazon RDS for MariaDB. Create or update a fully managed Amazon RDS database in the Amazon RDS Management Console.

Amazon VPC now supports IPv4 ingress routing for large IP Pools

Published Date: 2025-08-15 07:00:00

Amazon VPC now allows customers to route inbound internet traffic destined for large pools of public IP addresses, to a single elastic network interface (ENI) within a VPC. Prior to this enhancement, internet gateways only accepted traffic destined to public IP addresses that were associated with network interfaces in the VPC. There are limits to the number of IP addresses that can be associated with network interfaces. These limits depend on the instance type and can be found in our documentation. There are use cases in Telco, Internet of Things (IoT) and other industries that require customers to route inbound traffic destined for public IP pools, larger than the allowed limits, to a single network interface. Customers would earlier perform address translation to consolidate traffic for such large number of IP addresses. This enhancement removes the need to perform address translation on inbound internet connections for these Telco and IoT use cases. Customers can bring their own public IP pools (BYOIP documentation) and configure their VPC Internet Gateway to accept traffic belonging to this BYOIP pool and route it to a network interface. They can also use this feature with VPC Route Server and dynamically update their routes in events of failure. Refer to our public documentation for details on VPC Route Server. This enhancement is now available across all AWS commercial, AWS China and GovCloud regions. To learn more about this feature, please refer to our documentation.

Amazon Connect Cases now supports rules to automatically update cases upon creation

Published Date: 2025-08-15 07:00:00

Amazon Connect Cases now supports Contact Lens Rules that automatically update cases when created, streamlining case workflows and reducing manual tasks. For example, you can configure rules to auto-assign refund cases to the billing team, auto-close cases that don’t require follow-up, or automatically set priority based on case reason. Amazon Connect Cases is available in the following AWS regions: US East (N. Virginia), US West (Oregon), Canada (Central), Europe (Frankfurt), Europe (London), Asia Pacific (Seoul), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), and Africa (Cape Town) AWS regions. To learn more and get started, visit the Amazon Connect Cases webpage and documentation.

SageMaker HyperPod now supports fine-grained quota allocation of compute resources

Published Date: 2025-08-14 21:00:00

SageMaker HyperPod task governance now supports fine-grained compute quota allocation of GPU, Trainium accelerator, vCPU, and vCPU memory within an instance. Administrators can allocate fine-grained compute quota across teams, optimizing compute resource distribution and staying within budget. Data scientists often execute LLM tasks, like training or inference, that do not require entire HyperPod instances, leading to underutilization of accelerated compute resources. HyperPod task governance enables administrators to manage compute quota allocation across teams. With this capability, administrators can now strategically allocate compute resources, ensuring fair access, preventing resource monopolization, and maximizing cluster utilization. This capability enables fine-grained compute quota allocation in addition to instance-level allocation, aligning with organizational workload demands. SageMaker HyperPod task governance is available in all AWS Regions where HyperPod is available: US East (N. Virginia), US West (N. California), US West (Oregon), Asia Pacific (Mumbai), Asia Pacific (Singapore), Asia Pacific (Sydney), and Asia Pacific (Tokyo), Europe (Frankfurt), Europe (Ireland), Europe (London), Europe (Stockholm), and South America (São Paulo). To learn more, visit SageMaker HyperPod webpage, and HyperPod task governance documentation.

Amazon U7i instances now available in the AWS US East (Ohio) Region

Published Date: 2025-08-14 16:49:00

Starting today, Amazon EC2 High Memory U7i instances with 12TB of memory (u7i-12tb.224xlarge) are now available in the US East (Ohio) region. U7i-12tb instances are part of AWS 7th generation and are powered by custom fourth generation Intel Xeon Scalable Processors (Sapphire Rapids). U7i-12tb instances offer 12TiB of DDR5 memory enabling customers to scale transaction processing throughput in a fast-growing data environment. U7i-12tb instances offer 896 vCPUs, support up to 100Gbps Elastic Block Storage (EBS) for faster data loading and backups, deliver up to 100Gbps of network bandwidth, and support ENA Express. U7i instances are ideal for customers using mission-critical in-memory databases like SAP HANA, Oracle, and SQL Server. To learn more about U7i instances, visit the High Memory instances page.

Amazon EC2 R8g instances now available in AWS Asia Pacific (Jakarta)

Published Date: 2025-08-14 16:14:00

Starting today, Amazon Elastic Compute Cloud (Amazon EC2) R8g instances are available in AWS Asia Pacific (Jakarta)region. These instances are powered by AWS Graviton4 processors and deliver up to 30% better performance compared to AWS Graviton3-based instances. Amazon EC2 R8g instances are ideal for memory-intensive workloads such as databases, in-memory caches, and real-time big data analytics. These instances are built on the AWS Nitro System, which offloads CPU virtualization, storage, and networking functions to dedicated hardware and software to enhance the performance and security of your workloads. AWS Graviton4-based Amazon EC2 instances deliver the best performance and energy efficiency for a broad range of workloads running on Amazon EC2. AWS Graviton4-based R8g instances offer larger instance sizes with up to 3x more vCPU (up to 48xlarge) and memory (up to 1.5TB) than Graviton3-based R7g instances. These instances are up to 30% faster for web applications, 40% faster for databases, and 45% faster for large Java applications compared to AWS Graviton3-based R7g instances. R8g instances are available in 12 different instance sizes, including two bare metal sizes. They offer up to 50 Gbps enhanced networking bandwidth and up to 40 Gbps of bandwidth to the Amazon Elastic Block Store (Amazon EBS). To learn more, see Amazon EC2 R8g Instances. To explore how to migrate your workloads to Graviton-based instances, see AWS Graviton Fast Start program and Porting Advisor for Graviton. To get started, see the AWS Management Console.

Amazon OpenSearch UI is now available in seven new regions

Published Date: 2025-08-14 16:10:00

Amazon OpenSearch Service expands its modernized operational analytics experience to seven new regions, including Asia Pacific (Hyderabad), Asia Pacific (Osaka), Asia Pacific (Seoul), Europe (Milan), Europe (Zurich), Europe (Spain), and US-West (N. California) enabling users to gain insights across data spanning managed domains and serverless collections from a single endpoint. The expansion includes Workspaces to enhance collaboration and productivity, allowing teams to create dedicated spaces. Discover is revamped to provide a unified log exploration experience supporting languages such as Piped-Processing-Language (PPL) and SQL, in addition to DQL and Lucene. Discover now features a data selector to support multiple sources, new visual design and query autocomplete for improved usability. This experience ensures users can access the latest UI enhancements, regardless of version of underlying managed cluster or collection. The expanded OpenSearch analytics helps users gain insights from their operational data by providing purpose-built features for observability, security analytics, and search use cases. With the enhanced Discover interface, users can now analyze data from multiple sources without switching tools, improving efficiency. Workspaces enable better collaboration by creating dedicated environments for teams to work on dashboards, saved queries, and other relevant content. Availability of the latest UI updates across all versions ensures uninterrupted access to the newest features and tools. OpenSearch UI can connect to OpenSearch domains (above version 1.3) and OpenSearch serverless collections. It is now available in 22 AWS commercial regions. To get started, create an OpenSearch application in AWS Management Console. Learn more at Amazon OpenSearch Service Developer Guide.

PostgreSQL 18 Beta 3 is now available in Amazon RDS Database Preview Environment

Published Date: 2025-08-14 16:00:00

Amazon RDS for PostgreSQL 18 Beta 3 is now available in the Amazon RDS Database Preview Environment, allowing you to evaluate the pre-release of PostgreSQL 18 on Amazon RDS for PostgreSQL. You can deploy PostgreSQL 18 Beta 3 in the Amazon RDS Database Preview Environment that has the benefits of a fully managed database. PostgreSQL 18 includes "skip scan" support for multicolumn B-tree indexes and improves WHERE clause handling for OR and IN conditions. It introduces parallel GIN index builds and updates join operations. Observability improvements show buffer usage counts and index lookups during query execution, along with per-connection I/O utilization metric. Please refer the RDS PostgreSQL release documentation for more details. Amazon RDS Database Preview Environment database instances are retained for a maximum period of 60 days and are automatically deleted after the retention period. Amazon RDS database snapshots that are created in the preview environment can only be used to create or restore database instances within the preview environment. You can use the PostgreSQL dump and load functionality to import or export your databases from the preview environment. Amazon RDS Database Preview Environment database instances are priced as per the pricing in the US East (Ohio) Region.

SageMaker HyperPod now supports Topology Aware Scheduling of LLM tasks

Published Date: 2025-08-14 16:00:00

SageMaker HyperPod task governance now supports Topology Aware Scheduling (TAS), enabling data scientists to schedule their large language model (LLM) tasks on an optimal network topology that minimizes network communication and enhances training efficiency.

LLM training and fine-tuning tasks that are distributed across multiple accelerated compute instances frequently exchange large volumes of data between them. Multiple network hops between instances can result in higher communication latency, impacting LLM task performance. SageMaker HyperPod task governance now enables data scientists to use network topology information when scheduling tasks with specific topology preferences. Using network topology in HyperPod, SageMaker HyperPod task governance automatically schedules tasks in optimal locations, reducing instance-to-instance communication and enhancing training efficiency.

SageMaker HyperPod task governance is available in all AWS Regions where HyperPod is available: US West (N. California), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Europe (Frankfurt), Europe (Ireland), Europe (Stockholm). To learn more, visit SageMaker HyperPod webpage, and SageMaker HyperPod task governance documentation.

Amazon EC2 I7ie instances are now available in additional AWS regions

Published Date: 2025-08-14 16:00:00

Amazon Web Services (AWS) announces the availability of Amazon EC2 I7ie instances in the AWS Europe (Stockholm), Asia Pacific (Jakarta), and US West (N. California) regions. Designed for large storage I/O intensive workloads, these new instances are powered by 5th generation Intel Xeon Scalable processors with an all-core turbo frequency of 3.2 GHz, offering up to 40% better compute performance and 20% better price performance over previous generation I3en instances. I7ie instances offer up to 120TB local NVMe storage density—the highest available in the cloud for storage optimized instances—and deliver up to twice as many vCPUs and memory compared to prior generation instances. Powered by 3rd generation AWS Nitro SSDs, these instances achieve up to 65% better real-time storage performance, up to 50% lower storage I/O latency, and 65% lower storage I/O latency variability compared to I3en instances. Additionally, torn write prevention feature support up to 16KB block sizes, enables customers to eliminate performance bottlenecks for database workloads. I7ie instances are high-density storage-optimized instances, for workloads that demand rapid local storage with high random read/write performance and consistently low latency for accessing large data sets. These instances are offered in eleven different sizes including 2 metal sizes, providing flexibility for customers computational needs. They deliver up to 100 Gbps of network performance bandwidth, and 60 Gbps of dedicated bandwidth for Amazon Elastic Block Store (EBS), ensuring fast and efficient data transfer for applications. To learn more, visit the I7ie instances page.

AWS Batch now supports AWS Graviton-based Spot compute with AWS Fargate

Published Date: 2025-08-14 14:00:00

AWS Batch for ECS Fargate now supports AWS Graviton-based compute with AWS Fargate Spot. This capability helps you run fault-tolerant Arm-based applications with up to 70% discount compared to Fargate prices. AWS Graviton processors are custom-built by AWS to deliver the best price-performance for cloud workloads. AWS Batch for ECS Fargate enables customers to deploy and build workloads at scale in a serverless manner. Starting today, customers can further optimize for costs by running fault-tolerant Arm-based workloads on AWS Fargate Spot. To get started, create a new Fargate configured Compute Environment (CE), select ARM64 as the cpuArchitecture, and choose FARGATE_SPOT as the type. You can then connect it to existing job queues or create a new one for your workload. AWS Batch will leverage spare AWS Graviton-based compute capacity available in the AWS cloud for running your service or task. You can now get the simplicity of serverless compute with familiar cost optimization levers of Spot capacity with Graviton-based compute. This capability is now available for AWS Batch in all commercial and the AWS GovCloud (US) Regions. To learn more, see Batch’s updated RuntimePlatform API and AWS Batch for ECS Fargate documentation

Amazon FSx for NetApp ONTAP now supports decreasing your SSD storage capacity

Published Date: 2025-08-14 13:00:00

Amazon FSx for NetApp ONTAP, a fully managed shared storage service built on NetApp’s popular ONTAP file system, now allows you to decrease your file system’s solid-state drive (SSD) storage capacity, enabling you to more efficiently run project-based workloads with varying active working sets. You can provision SSD capacity upfront to meet peak usage needs—for periodic reporting, analytics, or large-scale data ingestion and processing—and then easily decrease SSD capacity to reduce storage costs. An FSx for ONTAP file system offers two storage tiers: a provisioned high-performance SSD tier for your workload’s active working set and a fully elastic capacity pool cost-optimized for infrequently accessed data. Until now, you could only increase your file system’s SSD capacity as your workload’s active working set grew. Starting today, you can decrease your file system's SSD capacity in-place with just a few clicks in the Amazon FSx console, allowing you to deliver optimal performance during peak usage for workloads such as Electronic Design Automation and media processing, then scale down SSD capacity once data processing is complete. You can also accelerate data migrations by temporarily increasing SSD capacity to enable faster data ingestion, then right-sizing SSD capacity after data has been tiered to the capacity pool. You can decrease SSD storage capacity on all FSx for ONTAP second-generation file systems in all AWS Regions where FSx for ONTAP second-generation file systems are available. For more information, see the FSx for ONTAP user guide.

AWS Cloud Map adds support for cross-account service discovery

Published Date: 2025-08-14 13:00:00

AWS Cloud Map now supports cross-account service discovery through integration with AWS Resource Access Manager (AWS RAM). This enhancement lets you seamlessly manage and discover cloud resources—such as Amazon ECS tasks, Amazon EC2 instances, and Amazon DynamoDB tables—across AWS accounts. By sharing your AWS Cloud Map namespace via AWS RAM, workloads in other accounts can discover and manage resources registered in that namespace. This enhancement simplifies resource sharing, reduces duplication, and promotes consistent service discovery across environments for organizations with multi-account architectures. You can now share your AWS Cloud Map namespaces using AWS RAM with individual AWS accounts, specific Organizational Units (OUs), or your entire AWS Organization. To get started, create a resource share in AWS RAM, add the namespaces you want to share, and specify the principals (accounts, OUs, or the organization) that should have access. This enables platform engineers to maintain a centralized service registry—or a small set of registries—and share them across multiple accounts, simplifying service discovery. Application developers can then build services that rely on a consistent, shared registry without worrying about availability or synchronization across accounts. AWS Cloud Map's cross-account service discovery support improves operational efficiency and makes it easier to scale service discovery as your organization grows by reducing duplication and streamlining access to namespaces. This feature is available now in all commercial AWS Regions via the AWS Management Console, API, SDK, CLI, and CloudFormation. To learn more, please refer to the AWS Cloud Map documentation.

Amazon Braket introduces support for program sets

Published Date: 2025-08-14 07:00:00

Amazon Braket now supports program sets, enabling quantum researchers to run complex workloads requiring hundreds of quantum circuit executions up to 24X faster. This new feature allows customers to submit up to 100 quantum programs or a single parametric circuit with up to 100 parameter values within a single quantum task. Program sets help minimize the time between subsequent circuit executions reducing quantum task processing overhead for complex algorithms. Program sets are particularly valuable for researchers working with variational quantum algorithms (VQA), quantum machine learning models, and error mitigation techniques. Customers can create program sets using two approaches: submitting multiple independent circuits together, or submitting a single parametric circuit with parameter sets. Amazon Braket handles compilation and execution orchestration, returning results that include the status and outcomes for each quantum program. If individual programs within a program set fail during execution, customers receive partial results for successfully completed programs and error information for failed executions. When submitting a program set, you pay a single per-task fee accompanied by a per-shot fee based on the total number of successful shots across your workload in a program set. Program sets are initially available on all superconducting quantum processing units (QPUs) from Rigetti in the US West (N. California) Region and IQM in the Europe (Stockholm) Region. Customers are able to submit program sets to Braket directly via the Amazon Braket SDK, from Qiskit via the Qiskit-Braket provider, or from PennyLane via the Amazon Braket PennyLane Plugin. To learn more about program sets, visit the Amazon Braket developer guide, explore our new example notebooks, and visit our updated Amazon Braket management console.

Amazon RDS for PostgreSQL supports minor versions 17.6, 16.10, 15.14, 14.19, and 13.22

Published Date: 2025-08-14 07:00:00

Amazon Relational Database Service (RDS) for PostgreSQL now supports the latest minor versions 17.6, 16.10, 15.14, 14.19, and 13.22. We recommend that you upgrade to the latest minor versions to fix known security vulnerabilities in prior versions of PostgreSQL, and to benefit from the bug fixes added by the PostgreSQL community. This release also includes updates for PostgreSQL extensions such as pg_repack 1.5.2, oracle_fdw 2.8.0, and pgactive 2.1.5 You can use automatic minor version upgrades to automatically upgrade your databases to more recent minor versions during scheduled maintenance windows. You can also use Amazon RDS Blue/Green deployments for RDS for PostgreSQL using physical replication for your minor version upgrades. Learn more about upgrading your database instances, including automatic minor version upgrades and Blue/Green Deployments in the Amazon RDS User Guide . Amazon RDS for PostgreSQL makes it simple to set up, operate, and scale PostgreSQL deployments in the cloud. See Amazon RDS for PostgreSQL Pricing for pricing details and regional availability. Create or update a fully managed Amazon RDS database in the Amazon RDS Management Console. 

Amazon Q Business launches Agentic RAG to enhance accuracy and explainability

Published Date: 2025-08-14 07:00:00

Amazon Q Business, the generative AI-powered assistant for finding information, gaining insight, and taking action at work, today introduced Agentic Retrieval-Augmented Generation (RAG) for Q Business applications. The new feature significantly enhances the ability of Q Business to provide more accurate and explainable responses to complex, multi-step RAG queries. Using Agentic RAG, Q Business uses AI agents to dynamically plan the retrieval and response generation for user’s queries that target enterprise data. Agentic RAG breaks down complex queries into simpler ones, executes them in parallel to retrieve data, intelligently invokes data retrieval tools, and combines the data to generate comprehensive responses. The built-in AI agents critique and validate the responses, and retry retrievals if necessary, resulting in higher accuracy. Through Agentic RAG, Q Business delivers a more intuitive dialogue experience, proactively resolving data conflicts through targeted clarifying questions and maintaining momentum with contextually relevant follow-ups – all while keeping users informed of the actions that Q Business takes. Agentic RAG is available in all AWS Regions where Amazon Q Business is offered. For any of their queries targeting their company knowledge, users can leverage this feature by toggling the “Advanced Search” option in the built-in web application. For more details, refer to the documentation. For more information about Amazon Q Business and its features, please visit the Amazon Q product page.

AWS Config now supports 10 new resource types

Published Date: 2025-08-14 07:00:00

AWS Config now supports 10 additional AWS resource types. This expansion provides greater coverage over your AWS environment, enabling you to more effectively discover, assess, audit, and remediate an even broader range of resources. With this launch, if you have enabled recording for all resource types, then AWS Config will automatically track these new additions. The newly supported resource types are also available in Config rules and Config aggregators. You can now use AWS Config to monitor the following newly supported resource types in all AWS Regions where the supported resources are available:

Resource Types:

AWS::Backup::RestoreTestingPlan

AWS::CloudFront::RealtimeLogConfig

AWS::EC2::SecurityGroupVpcAssociation

AWS::EC2::VerifiedAccessInstance

AWS::KafkaConnect::CustomPlugin

AWS::OpenSearchServerless::SecurityConfig

AWS::Redshift::Integration

AWS::Route53Profiles::ProfileAssociation

AWS::SSMIncidents::ResponsePlan

AWS::Transfer::Server

AWS Systems Manager Automation enhances runbook execution control and updates free tier

Published Date: 2025-08-14 07:00:00

AWS Systems Manager Automation now offers three new features that enhance runbook execution control and success rates. Additionally, we are announcing updates to our free tier structure, as part of our ongoing commitment to providing simple, standardized, and cost-effective services to customers. There are three key features added. First, customers can now easily re-execute runbooks directly from the Automation console with pre-populated parameters, streamlining repeated operations. Second, customers will be able to automatically retry throttled API calls during high-concurrency scenarios to improve execution reliability. Lastly, customers will be able to specify nested organizational units (OUs) in their target selection for more fine-grained control over their resources across accounts. Additionally, the existing free tier for Automation (100,000 steps and 5,000 seconds of script duration per month) will no longer be available for new customers and will end on December 31, 2025 for existing customers. Customers can try Automation capabilities for free by starting a new AWS account under a free plan, where they can use up to $200 in Free Tier credits on eligible AWS Services. Automation pricing remains the same at $0.002 per step executed and $0.00003 per second of scripts executed. Systems Manager Automation is available in all AWS Commercial and AWS GovCloud (US) Regions. To learn more about Automation, review the documentation and Systems Manager pricing page. For more details on available plans on new customer benefits, visit the AWS Free Tier page.

Accelerate Amazon WorkSpaces deployment with streamlined Bring Your Own License (BYOL) process

Published Date: 2025-08-14 07:00:00

WorkSpaces has improved the BYOL process, offering customers a more efficient and faster way to import their Windows images to use with WorkSpaces. With this streamlined approach, customers can enable the BYOL feature in their AWS account without contacting AWS Support. The new process supports importing either customized virtual machine (VM) images or Windows ISO files directly into WorkSpaces. Leveraging an integrated Amazon EC2 Image Builder pipeline, the system automatically constructs a WorkSpaces-compatible image from the imported source. During this process, most compatibility issues are identified and resolved automatically, reducing the manual troubleshooting efforts previously required. In cases where compatibility issues cannot be fixed automatically, customers now have the ability to access an EC2 instance to address these issues directly. This eliminates the need to upload a new image, further streamlining the import process. These improvements collectively reduce the overall time needed to get started with BYOL images on WorkSpaces, while minimizing the associated troubleshooting efforts. The improved BYOL process is available in all regions where WorkSpaces BYOL is currently available. To learn more, see the Amazon WorkSpaces BYOL documentation.

Amazon EC2 M8g instances now available in AWS Asia Pacific (Seoul)

Published Date: 2025-08-14 05:27:00

Starting today, Amazon Elastic Compute Cloud (Amazon EC2) M8g instances are available in AWS Asia Pacific (Seoul) region. These instances are powered by AWS Graviton4 processors and deliver up to 30% better performance compared to AWS Graviton3-based instances. Amazon EC2 M8g instances are built for general-purpose workloads, such as application servers, microservices, gaming servers, midsize data stores, and caching fleets. These instances are built on the AWS Nitro System, which offloads CPU virtualization, storage, and networking functions to dedicated hardware and software to enhance the performance and security of your workloads. AWS Graviton4-based Amazon EC2 instances deliver the best performance and energy efficiency for a broad range of workloads running on Amazon EC2. These instances offer larger instance sizes with up to 3x more vCPUs and memory compared to Graviton3-based Amazon M7g instances. AWS Graviton4 processors are up to 40% faster for databases, 30% faster for web applications, and 45% faster for large Java applications than AWS Graviton3 processors. M8g instances are available in 12 different instance sizes, including two bare metal sizes. They offer up to 50 Gbps enhanced networking bandwidth and up to 40 Gbps of bandwidth to the Amazon Elastic Block Store (Amazon EBS). To learn more, see Amazon EC2 M8g Instances. To explore how to migrate your workloads to Graviton-based instances, see AWS Graviton Fast Start program and Porting Advisor for Graviton. To get started, see the AWS Management Console.

AWS Security Incident Response now supports membership coverage for individual AWS organizational units

Published Date: 2025-08-14 04:00:00

AWS Security Incident Response now allows you to choose membership coverage for specific AWS organizational units (OUs) within an AWS Organization. While memberships previously covered all accounts in the selected AWS Organization, you now have the flexibility to choose which OUs to cover making it easier to try out the service and support your existing IR processes. 

AWS Security Incident Response combines the power of automated monitoring and investigation, accelerated communication and coordination, and direct 24/7 access to the AWS Customer Incident Response Team (CIRT) to quickly prepare for, respond to, and recover from security events. This new feature allows you to enable Security Incident Response for specific types of workloads, such as production workloads, or for specific business subsidiaries that are grouped as OUs in their AWS Organization. You can also start with a pilot OU to evaluate the service's effectiveness and refine processes before expanding coverage to your entire AWS Organization. All accounts within the selected OUs, including accounts under child OUs, are covered by the membership. Accounts added or removed from selected OUs will automatically update the coverage, making it easy for you to manage your AWS Security Incident Response membership.

Get started today by visiting AWS Security Incident Response via the AWS Management Console, AWS Command Line Interface (CLI), or AWS Software Development Kit (SDK). To learn more, see the AWS Security Incident Response documentation. For additional information on managing OUs with AWS Organizations, visit the AWS Organizations page.

Amazon SageMaker Studio now supports trusted identity propagation

Published Date: 2025-08-13 18:00:00

We are excited to announce that Amazon SageMaker Studio now supports trusted identity propagation (TIP), enabling admins to trace actions taken in SageMaker Studio back to a human user. It also enables administrators to manage permissions based on user identity to AWS Lake Formation and Amazon S3 Access Grants. Starting today, SageMaker Studio now supports trusted identity propagation for AWS services, including AWS Lake Formation, Amazon S3, Amazon EMR, Amazon EMR-Serverless, Amazon Redshift and Amazon Athena. Customers can enable trusted identity propagation for SageMaker Studio domains, either at domain creation or while updating an existing domain. With this capability, customers can apply fine grained access controls using S3 Access Grants, Lake Formation, or Redshift Data APIs to govern user access to data in their SageMaker Studio notebooks. Customers’ Training and Processing jobs in Studio notebooks will integrate with S3 Access Grants. With this change, administrators can trace creation of user interactive sessions as well as user background sessions across Studio applications (JupyterLab and CodeEditor) via AWS CloudTrail events. As an example, administrators can now track which user creates SageMaker Pipeline steps to run experiments using SageMaker MLFlow, trains ML models using SageMaker Training, and deploys using SageMaker Inference. This feature is available in all AWS Regions where Amazon SageMaker Studio with AWS IAM Identity Center support is currently available, excluding China regions, and GovCloud (US) regions. To learn more, visit our documentation.

AWS Resource Explorer now Supports Filtering for Multiple Values

Published Date: 2025-08-13 17:47:00

Today, AWS Resource Explorer introduces support for filtering on multiple values in both the Search and ListResources APIs. This new feature allows customers to build targeted queries to find resources more easily. The new filtering feature allows customers to search for resources based on multiple criteria simultaneously, improving upon the previous process where they had to query each individual facet and combine the search results. For example, customers can now find all storage-types tagged as non-production by querying “tag.key:environment tag.value:dev,test,integ service:dynamodb,rds,ebs...”. Importantly, customers with large AWS Organizations can now scope their Resource Explorer Views to specific listed accounts, rather than being scoped to an entire Organization or Organizational Unit. To scope multi-account views to a specific set of accounts, add a view filter with the “account:” operator, and a comma-separated list of account IDs to be included in search results.

This feature is available in all AWS Regions where AWS Resource Explorer is supported. For specific regional availability, please refer to the AWS Regional Services List. To learn more about this new feature visit the AWS Resource Explorer documentation page. You can start using this enhanced querying capability immediately through the AWS Management Console, AWS CLI, or AWS SDKs.

Amazon DocumentDB announces Extended Support for Version 3.6

Published Date: 2025-08-13 17:19:00

Amazon DocumentDB (with MongoDB compatibility) offers Extended Support for version 3.6, allowing customers to maintain critical workloads on version 3.6 for up to three years beyond the standard support end date on March 30, 2026. This is designed for customers who may need more time to plan and implement version upgrades, especially when navigating application dependencies or managing enterprise-scale deployments. Standard support for Amazon DocumentDB version 3.6 will end on March 30, 2026. After this date, clusters not upgraded to a supported version will incur Extended Support charges. During the Extended Support period, AWS will continue to provide critical security updates for Common Vulnerabilities and Exposures (CVEs). To perform an upgrade from Amazon DocumentDB version 3.6 to 5.0, customers can use an in-place major version upgrade. Upgrading to version 5.0 of Amazon DocumentDB enables customers to benefit from features such as I/O-Optimized, Graviton2, and Serverless instances, providing price per performance improvements. To get started and learn more about the Extended Support, see our blog.

Amazon DynamoDB now supports more frequent throughput mode updates from provisioned to on-demand capacity

Published Date: 2025-08-13 17:00:00

You can now update your Amazon DynamoDB table throughput mode from provisioned capacity to on-demand mode up to four times in a rolling 24-hour period. This enhancement makes it simpler for customers who have use cases which require loading large volumes of data into their DynamoDB tables multiple times per day, or want greater flexibility to manage their workload requirements and optimize costs. The increased quota also simplifies CloudFormation deployments and rollbacks, allowing templates to execute more readily without being blocked by previous changes. Previously, you could only set or update your table throughput mode to on-demand capacity once in a 24-hour period. DynamoDB on-demand mode offers a truly serverless database experience with pay-per-request pricing and automatic scaling to accommodate the most demanding workloads without capacity planning. Many customers prefer the simplicity of on-demand mode to build modern, serverless applications that can start small and scale to millions of requests per second. There is no charge for switching capacity modes, and the increased quota automatically applies to all DynamoDB tables. On-demand throughput is available in all AWS Regions. For more information about DynamoDB on-demand throughput, see the Developer Guide.

Announcing expanded support for Cilium with Amazon EKS Hybrid Nodes

Published Date: 2025-08-13 17:00:00

Today, Amazon Elastic Kubernetes Service (Amazon EKS) expands support for Cilium as the Container Networking Interface (CNI) for Amazon EKS Hybrid Nodes. Cilium is a Cloud-Native Computing Foundation (CNCF) graduated project that provides core networking capabilities for Kubernetes workloads. Now, you can receive support from AWS for a broader set of Cilium features when using Cilium with Amazon EKS Hybrid Nodes including application ingress, in-cluster load balancing, Kubernetes network policies, and kube-proxy replacement mode. Kubernetes clusters require a CNI for connectivity between pods running in the cluster, but most Kubernetes applications require additional components, such as ingress controllers and load balancers, to serve and secure network traffic with other external systems or users. These additional capabilities are integrated features of Cilium, built on Cilium’s eBPF-powered networking and security. Now, Amazon EKS Hybrid Nodes users can receive support from AWS for Cilium’s Ingress and Gateway features, Border Gateway Protocol (BGP) Control Plane, Load Balancer IP Address Management (LB IPAM), kube-proxy replacement, and Kubernetes network policies. AWS supports the Amazon VPC CNI for Amazon EKS nodes in AWS Cloud, which is optimized for Amazon VPC networking with built-in features such as enhanced subnet discovery, Kubernetes network policies, and multiple network interfaces per pod. Cilium support for Amazon EKS Hybrid Nodes is available in all AWS Regions where Amazon EKS Hybrid Nodes is available. To learn more about Cilium support for Amazon EKS Hybrid Nodes, see Configure CNI for hybrid nodes in the Amazon EKS User Guide.

Amazon FSx for OpenZFS now supports Internet Protocol Version 6 (IPv6)

Published Date: 2025-08-13 10:18:00

Amazon FSx now offers customers the option to use Internet Protocol version 6 (IPv6) for access to Amazon FSx for OpenZFS file systems. More and more customers are adopting IPv6 to mitigate IPv4 address exhaustion in their private networks or to satisfy government mandates such as the US Office of Management and Budget (OMB) M-21-07 memorandum. With this launch, customers can now access their file systems using IPv4, IPv6, or dual-stack clients without the need for complex infrastructure to handle IPv6 to IPv4 address translation. IPv6 support for new FSx for OpenZFS file systems is now available in all AWS Commercial, AWS GovCloud (US), and AWS China regions where Amazon FSx is available, with IPv6 support for existing FSx for OpenZFS file systems coming in an upcoming weekly maintenance window. To learn more, visit the Amazon FSx user guide.

AWS IAM Identity Center introduces support for user background sessions with Amazon SageMaker Studio

Published Date: 2025-08-13 07:00:00

 AWS IAM Identity Center introduces support for user background sessions, a new feature which allows Amazon SageMaker Studio users to initiate long-running jobs that continue to run in the background even after a user logs off from their computer. User background sessions leverage permissions tied to a user's identity and their corporate directory group associations while accessing data across multiple AWS services including Amazon Athena, Amazon EMR, AWS Lake Formation, Amazon Redshift, and Amazon S3.

Building on IAM Identity Center's trusted identity propagation (TIP) capability, this feature allows SageMaker Studio jobs to run for up to 90 days without requiring users to stay signed in. When launching jobs such as SageMaker Training, Processing, or Pipeline executions, IAM Identity Center creates a new user background session that runs independently until job completion or the background session duration limit is reached, whichever comes first. Administrators can track the creation of user background sessions via AWS CloudTrail events, update the background session duration for their IAM Identity Center instance and terminate background sessions for different users in their organization.

AWS IAM Identity Center helps you securely connect your workforce identities and manage their access centrally across AWS applications. For more information on user background sessions, see here. For the list of supported regions and compatible AWS services supported with user background sessions, refer to the Amazon SageMaker Studio page, linked here

Amazon EC2 G6 instances now available in the AWS GovCloud (US-East) Region

Published Date: 2025-08-12 21:08:00

Starting today, the Amazon Elastic Compute Cloud (Amazon EC2) G6 instances powered by NVIDIA L4 GPUs are now available in AWS GovCloud (US-East). G6 instances can be used for a wide range of graphics-intensive and machine learning use cases.

Customers can use G6 instances for deploying ML models for natural language processing, language translation, video and image analysis, speech recognition, and personalization as well as graphics workloads, such as creating and rendering real-time, cinematic-quality graphics and game streaming. G6 instances feature up to 8 NVIDIA L4 Tensor Core GPUs with 24 GB of memory per GPU and third generation AMD EPYC processors. They also support up to 192 vCPUs, up to 100 Gbps of network bandwidth, and up to 7.52 TB of local NVMe SSD storage.

Amazon EC2 G6 instances are already available today in the AWS US East (N. Virginia and Ohio) , US West (Oregon), Europe (Frankfurt, London, Paris, Spain, Stockholm and Zurich), Asia Pacific (Mumbai, Tokyo, Malaysia, Seoul and Sydney), South America (Sao Paulo) and Canada (Central) regions. Customers can purchase G6 instances as On-Demand Instances, Reserved Instances, Spot Instances, or as part of Savings Plans.

To get started, visit the AWS Management Console, AWS Command Line Interface (CLI), and AWS SDKs. To learn more, visit the G6 instance page.

Amazon EC2 Single GPU P5 instances are now generally available

Published Date: 2025-08-12 19:09:00

Today, AWS announces new Amazon Elastic Compute Cloud (Amazon EC2) P5 instance size with one NVIDIA H100 GPU that allows businesses to right-size their machine learning (ML) and high-performance computing (HPC) resources with cost-effectiveness.

The new instance size enables customers to start small and scale in granular increments, providing more flexible control over infrastructure costs. Customers developing small to medium Large Language Models (LLMs) such as chatbots or specialized language translation tools can now run inference tasks more economically. Customers can also use these instances to deploy HPC applications for pharmaceutical discovery, fluid flow analysis, and financial modeling without committing to expensive, large-scale GPU deployments.

P5.4xlarge instances are now available through Amazon EC2 Capacity Blocks for ML in the following AWS Regions: US East (North Virginia, Ohio), US West (Oregon), Europe (London), Asia Pacific (Mumbai, Sydney, Tokyo) and South America (Sao Paulo) regions. These instances can be purchased On-Demand, Spot or through Savings Plans in Europe (London), Asia Pacific (Mumbai, Jakarta, Tokyo), and South America (Sao Paulo) regions.

To learn more about P5.4xlarge instances, visit Amazon EC2 P5 instances.

Amazon SageMaker AI now supports P6e-GB200 UltraServers

Published Date: 2025-08-12 19:07:00

Today, Amazon SageMaker AI announces support for P6e-GB200 UltraServers in SageMaker HyperPod and Training Jobs. With P6e-GB200 UltraServers, you can leverage up to 72 NVIDIA Blackwell GPUs under one NVLink domain to accelerate training and deployment of foundational models at trillion-parameter scale. P6e-GB200 UltraServers are available in two sizes: ml.u-p6e-gb200x72 (72 GPUs within NVLink) and ml.u-p6e-gb200x36 (36 GPUs within NVLink). P6e-GB200 UltraServers deliver over 20x compute and over 11x memory under one NVIDIA NVLink compared to P5en instances. Within each NVLink domain you can leverage 360 petaflops of FP8 compute (without sparsity) and 13.4 TB of total high bandwidth memory (HBM3e). When you use P6e-GB200 UltraServers on SageMaker AI, you get the GB200’s superior performance combined with SageMaker's managed infrastructure such as security, built-in fault tolerance, topology aware scheduling (SageMaker HyperPod EKS & Slurm), integrated monitoring capabilities, and native integration with other SageMaker AI and AWS services. The UltraServers are available through SageMaker Flexible Training Plans in the Dallas Local Zone ("us-east-1-dfw-2a"), an extension of the US East (N. Virginia) AWS Region. For on-demand reservation of GB200 UltraServers, please reach out to your account manager. Amazon SageMaker AI lets you easily train and deploy machine learning models at scale using fully managed infrastructure optimized for performance and cost. To get started with UltraServers on SageMaker AI, visit the documentation.

Announcing new incentives for ISVs selling in AWS Marketplace

Published Date: 2025-08-12 18:25:00

Amazon Web Services, Inc. (AWS) announces the launch of the AWS Marketplace Private Offer Promotion Program (MPOPP) in AWS Partner Central to support independent software vendors (ISVs) with driving new customer acquisition. This program is designed to accelerate sales through AWS Marketplace by offering AWS Promotional Credits to customers as an incentive for purchasing from participating ISVs. MPOPP offers benefits for AWS Partners at different stages in their AWS Marketplace journey. New AWS Marketplace Sellers can benefit from immediate funding support, and established sellers can benefit from special incentives for driving AWS Marketplace renewals. Eligible Partners can submit self-service requests for funds through the AWS Partner Funding Portal year-round, enabling funding to be targeted for next business day delivery following Private Offer acceptance. The simplified funding template can help accelerate deal closure and provides better speed-to-market with a fully automated approval process. Following deal completion and AWS Marketplace transaction verification, Promotional Credits will be issued to the customer's AWS account based on the Total Contract Value (TCV) and applicable program rates, streamlining the entire process from planning to credit disbursement. To learn more about the MPOPP, eligibility, and benefits, visit the AWS Partner Funding Benefits Guide (AWS Partner Central login required).

Amazon SageMaker HyperPod now supports custom AMIs (Amazon Machine Images)

Published Date: 2025-08-12 18:06:00

Amazon SageMaker HyperPod now supports custom AMIs, enabling customers to deploy clusters with pre-configured, security-hardened environments that meet their specific organizational requirements. Customers deploying AI/ML workloads on HyperPod need customized environments that meet strict security, compliance, and operational requirements while maintaining fast cluster startup times, but often struggle with complex lifecycle configuration scripts that slow deployment and create inconsistencies across cluster nodes. This capability allows customers to build upon HyperPod's performance-optimized base AMIs while incorporating customized security agents, compliance tools, proprietary libraries, and specialized drivers directly into the image, delivering faster startup times, improved reliability, and enhanced security compliance. Security teams can embed organizational policies directly into base images, allowing AI/ML teams to use pre-approved environments that accelerate time-to-training while meeting enterprise security standards. You can specify custom AMIs when creating new HyperPod clusters using the CreateCluster API, adding instance groups with UpdateCluster API, or patching existing clusters with UpdateClusterSoftware API. Custom AMIs must be built using HyperPod's public base AMIs to maintain compatibility with distributed training libraries and cluster management capabilities. This feature is available in all AWS Regions where Amazon SageMaker HyperPod is supported. To learn more about custom AMI support, see the Amazon SageMaker HyperPod User Guide.

Anthropic’s Claude Sonnet 4 in Amazon Bedrock Expanded Context Window

Published Date: 2025-08-12 16:48:00

Anthropic’s Claude Sonnet 4 in Amazon Bedrock is launching today with a significantly expanded context window in public preview. The context window has been increased from 200,000 to 1 million tokens, representing a 5x expansion. This enhancement allows Claude to process and reason over much larger amounts of text in a single request, opening up new possibilities for comprehensive analysis and generation tasks. This expanded context window for Sonnet 4 brings many benefits to customers. For large-scale code analysis, users can now load entire codebases, including source files, tests, and documentation, enabling Sonnet 4 to understand project architecture, identify cross-file dependencies, and suggest improvements that account for the complete system design. In document synthesis, the model can now process extensive document sets like legal contracts, lengthy research papers, large datasets, or technical specifications in a single API call, analyzing relationships across hundreds of documents while maintaining full context. Additionally, this expansion allows for the creation of more sophisticated context-aware agents that can maintain coherence across hundreds of tool calls and multi-step workflows, including complete API documentation and interaction histories. The expanded context window for Claude Sonnet 4 is now available in public preview in Amazon Bedrock in US West (Oregon), US East (N. Virginia), and US East (Ohio) AWS regions. Prompts over 200,000 tokens will incur approximately twice the token price for input and 1.5 times for output. To get started with the expanded context window for Claude Sonnet 4, visit the Amazon Bedrock console.

AWS Deadline Cloud introduces new cost-saving compute option

Published Date: 2025-08-12 16:00:00

AWS Deadline Cloud is a fully managed service that simplifies render management for teams creating computer-generated graphics and visual effects for films, television, broadcasting, web content, and design. Today, we’re excited to announce a new wait and save feature for Deadline Cloud service-managed fleets that can reduce rendering costs with prices starting as low as $0.006 per vCPU-hour. This new feature is ideal for non time-sensitive rendering workloads with flexible completion times. Submitting jobs using this wait and save approach allows you to achieve significant cost savings so you can do more creative iteration and exploration on your next project. This feature complements existing AWS Deadline Cloud compute options in its service-managed fleets, giving you more flexibility to optimize your resource utilization across different priorities and budgets. AWS Deadline Cloud wait and save is available in all AWS Regions where AWS Deadline Cloud is offered. To learn more about this new cost-saving feature and how it can help optimize your rendering workloads, visit the AWS Deadline Cloud product page or review the AWS Deadline Cloud documentation.

AWS Direct Connect announces 100G expansion in Cape Town, South Africa

Published Date: 2025-08-12 16:00:00

Today, AWS announced the expansion of 100 Gbps dedicated connections at the AWS Direct Connect location in the Teraco CT1 data center near Cape Town, South Africa. You can now establish private, direct network access to all public AWS Regions (except those in China), AWS GovCloud Regions, and AWS Local Zones from this location. This is the second AWS Direct Connect location in South Africa to provide 100 Gbps connections with MACsec encryption capabilities.

The Direct Connect service enables you to establish a private, physical network connection between AWS and your data center, office, or colocation environment. These private connections can provide a more consistent network experience than those made over the public internet.

For more information on the over 142 Direct Connect locations worldwide, visit the locations section of the Direct Connect product detail pages. Or, visit our getting started page to learn more about how to purchase and deploy Direct Connect.

Amazon OpenSearch Serverless now supports kNN Byte vector and new data types

Published Date: 2025-08-12 15:00:00

Amazon OpenSearch Serverless has introduced several new features including kNN Byte vector support, radial search capabilities for Vector collections, and new data types and mapping parameters such as strict_allow_templates, wildcard field type, and kuromoji_completion analyzer. These enhancements deliver significant benefits for search and analytics workloads. The kNN Byte vector support helps reduce costs through lower memory and storage requirements while improving latency and performance. The additional features like nested fields for storing multiple vectors in a single document and new mapping parameters provide greater flexibility and control in managing search operations without the complexity of infrastructure management. Please refer to the AWS Regional Services List for more information about Amazon OpenSearch Service availability. To learn more about OpenSearch Serverless, see the documentation. 

Amazon Q Business launches Response Events for enhanced chat transparency

Published Date: 2025-08-12 07:00:00

Amazon Q Business, the generative AI-powered assistant for finding information, gaining insight, and taking action at work, today introduced Response Events, a new feature that provides real-time visibility into the assistant's query processing steps. This capability allows users to observe how their queries are processed, creating transparency and trust in their interactions with Amazon Q Business. Previously, users had no visibility into how Amazon Q Business processed their queries and generated responses, making it challenging to understand the reasoning behind answers. This lack of transparency created a black-box experience that limited organizations' ability to audit and trust AI-generated responses. With Response Events, users can now track processing steps in real-time across multiple components: RAG for enterprise-specific knowledge, uploaded files in chat sessions, and interactions with both built-in and custom plugins if enabled. Response Events are available by default for all chat conversations within Amazon Q Business and in all AWS Regions where Amazon Q Business is available. To learn more about Amazon Q Business how it can enhance your organization's productivity, visit the Amazon Q Business webpage.

New Sharing and Targeting Capabilities for EC2 On-Demand Capacity Reservations in Cluster Placement Groups

Published Date: 2025-08-12 07:00:00

Today, we are introducing multiple enhancements to Amazon EC2 On-Demand Capacity Reservations in Cluster Placement Groups (CPG-ODCRs). CPG-ODCRs provide customers with assured capacity and offer low latency and high throughput between instances within the same Cluster Placement Group (CPG). Now, customers using CPG-ODCRs can benefit from two additional capabilities that make them easier to use. First, customers can now add ODCRs belonging to different CPGs to Resource Groups which will enable customers to manage and target groups of reservations spread across multiple Placement Groups. Second, customers can share CPG-ODCRs across multiple AWS accounts through AWS Resource Access Manager, which allow them to create central pools of capacity and use them efficiently across workloads in different accounts. Customers can get started with these capabilities of CPG-ODCR by using the AWS CLI/APIs or by visiting the AWS Management console. These capabilities are now available in all AWS regions except China, and they are available at no additional cost. To learn more about these capabilities, please refer to the Capacity Reservations user guide.

Amazon SageMaker HyperPod now provides a new cluster setup experience

Published Date: 2025-08-11 21:00:00

SageMaker HyperPod now provides a new cluster creation experience that sets up all the resources needed for large-scale AI/ML workloads—including networking, storage, compute, and IAM permissions in just a few clicks. SageMaker HyperPod clusters are purpose-built for scalability and resilience, designed to accelerate large-scale distributed training and deployment of complex machine learning models like LLMs and diffusion models, as well as customization of Amazon Nova foundation models. The new cluster creation experience for SageMaker HyperPod introduces both quick and custom setup paths that make it easier for both beginners and advanced AWS customers to get started. Previously, customers needed to manually configure networking, IAM roles, storage, and compute. With the new quick setup, model builders, who may not have AWS infrastructure expertise, can now launch a fully-operational cluster optimized for large-scale AI workloads in just a few clicks using a streamlined single-page interface that provisions all dependencies including VPCs, subnets, FSx storage, EKS/Slurm orchestrator, and essential (k8s) operators. For platform engineering teams who may want to modify the default settings, the custom setup path provides full control over every configuration—from specific subnet configurations to selective operator installations—from within the same console experience. Teams can also export an auto-generated CloudFormation template for repeatable production deployments. You can create clusters using either the AWS Console or CloudFormation in all AWS Regions where SageMaker HyperPod is supported. To learn more, see the user guide.

Amazon Connect Outbound Campaigns now supports multi-profile campaigns and enhanced phone number retry sequencing

Published Date: 2025-08-11 19:36:00

Amazon Connect Outbound Campaigns now supports account-based campaigns, allowing you to reach multiple people associated with the same account. For example, when calling about a joint bank account, if the first person is unavailable, the system automatically tries to reach other authorized members of the account. You can also define a prioritized contact sequence across multiple phone numbers, for example, mobile first, then home, then work. If the first number is unreachable, Connect will automatically try the next number in the sequence. Previously, campaigns targeted one profile and retried a single phone number. With these updates, you can target multiple profiles within the same campaign, enabling outreach to all associated contacts in an account. You can also configure fallback phone numbers within each profile, automatically moving to the next preferred phone number if the first attempt is unsuccessful. Together, these capabilities help you create more flexible and effective engagement workflows that improve right-party contact rates and simplify campaign management. This feature is available in all AWS Regions where Amazon Connect Outbound Campaigns is supported. To get started, refer to the Amazon Connect Customer Profiles documentation to learn how to ingest customer data, and the Outbound Campaigns documentation for guidance on creating campaigns.. 

Amazon RDS for Oracle now supports July 2025 Spatial Patch Bundle

Published Date: 2025-08-11 19:27:00

Amazon Relational Database Service (Amazon RDS) for Oracle now supports the Spatial Patch Bundle (SPB) for the July 2025 Release Update (RU) for Oracle Database version 19c. This update delivers important fixes for Oracle Spatial and Graph functionality, helping ensure reliable and optimal performance for your spatial operations. You can now create new DB instances or upgrade existing ones to engine version ‘19.0.0.0.ru-2025-07.spb-1.r1’. The SPB engine version will be visible in the AWS Console by selecting the "Spatial Patch Bundle Engine Versions" checkbox in the engine version selector, making it simple to identify and implement the latest spatial patches for your database environment. To learn more about Oracle SPBs supported on Amazon RDS for each engine version, see the Amazon RDS for Oracle Release notes. For more information about the AWS Regions where Amazon RDS for Oracle is available, see the AWS Region table.

Amazon RDS for Oracle now supports the July 2025 Release Update (RU)

Published Date: 2025-08-11 17:51:00

Amazon Relational Database Service (Amazon RDS) for Oracle now supports the July 2025 Release Update (RU) for Oracle Database versions 19c and 21c. These RUs include bug and security fixes and are available for RDS for Oracle Standard Edition 2 and Enterprise Edition. Review the Oracle release notes for July RU for details. We recommend upgrading to this RU as it includes security fixes. You can upgrade with just a few clicks in the Amazon RDS Management Console or by using the AWS SDK or CLI. You can also enable auto minor version upgrade (AmVU) to automatically upgrade your database instances. Learn more about upgrading your database instances from the Amazon RDS User Guide. This new minor version is available in all AWS regions where Amazon RDS for Oracle is available. See Amazon RDS for Oracle Pricing for pricing details and regional availability.

Amazon Web Services (AWS) Advanced Go Driver is generally available

Published Date: 2025-08-11 16:10:00

The Amazon Web Services (AWS) Advanced Go Driver is now generally available for use with Amazon RDS and Amazon Aurora PostgreSQL and MySQL-compatible database clusters. This database driver provides support for faster switchover and failover times, Federated Authentication, and authentication with AWS Secrets Manager or AWS Identity and Access Management (IAM). The Amazon Web Services (AWS) Advanced Go Driver "wraps" the pgx PostgreSQL driver or the native MySQL driver to provide enhanced functionality. You can install the PostgreSQL and MySQL packages for Windows, Mac, or Linux, by following the installation guides in GitHub. The driver relies on monitoring the database cluster status and being aware of the cluster topology to determine the new writer. This approach reduces writer failover times to single digit seconds compared to the open-source drivers on their own (pgx and native MySQL). The driver is available as an open-source project under the Apache 2.0 license. For implementation details, please refer to the getting started instructions on GitHub.

AWS IoT Core introduces DeleteConnection API to streamline MQTT connection management

Published Date: 2025-08-11 14:00:00

AWS IoT Core now offers the DeleteConnection API, enabling programmatic disconnection of MQTT clients using their client IDs. This new capability helps enable developers to terminate MQTT connections with options to clear persistent sessions and suppress publication of Last Will and Testament messages—messages that the MQTT broker automatically publishes on a client’s behalf when it disconnects unexpectedly. Upon disconnection, the service generates lifecycle events, providing enhanced operational visibility into device connection states. The DeleteConnection API helps developers manage device connectivity more effectively, whether redirecting devices across endpoints, troubleshooting connection issues, or handling problematic device behavior. The DeleteConnection API is now available in all AWS Regions where AWS IoT Core is supported. To learn more, visit the AWS IoT Core documentation and AWS IoT Core API reference guide.

To view or add a comment, sign in

Explore topics