AWS launches EC2 capacity blocks for short-term GPU compute

Amazon Web Services (AWS) has announced the general availability of EC2 Capacity Blocks for machine learning (ML), enabling customers to reserve GPU capacity for short-duration ML projects.

EC2 Capacity Blocks can be utilized with P5 instances, which use Nvidia H100 Tensor Core GPUs. The EC2 UltraClusters are interconnected with second-generation Elastic Fabric Adapter (EFA) networking.

Customers can schedule their EC2 Capacity Blocks up to eight weeks in advance for a duration of one to 14 days, in cluster sizes ranging from one to 64 instances.

The price of an EC2 Capacity Block depends on available supply and demand for EC2 Capacity Blocks at the time users purchase a reservation. The operating system price is billed at per-second granularity. Find pricing details here.

EC2 Capacity Blocks are initially available in the AWS US East (Ohio) Region, with plans for future expansion to other AWS Regions and Local Zones.

[Image courtesy: AWS]

Just in

Snowflake to acquire TruEra AI observability platform

Snowflake has entered into a definitive agreement to acquire TruEra, providers of an AI observability platform. Financial terms of the transaction were not disclosed.

Biggest Windows 11 update in 2 years nearly finalized, enters Release Preview — Ars Technica

Microsoft has released Windows 11 24H2 build 26100.712 to its Release Preview testing channel for Windows Insiders, a sign that the update is nearly complete and that the company has shifted into bug-fixing mode ahead of general availability, writes Andrew Cunningham in Ars Technica. 

Elon Musk’s xAI is working on making Grok multimodal — The Verge

Elon Musk’s AI company, xAI, is making progress on adding multimodal inputs to its Grok chatbot, according to public developer documents, writes Kylie Robison in The Verge.