OpenAI has reached a $3.8 billion long-term partnership with Amazon AWS, and will use AWS cloud infrastructure over the next seven years, particularly Amazon EC2 UltraServers and thousands of high-performance NVIDIA GPUs. This marks OpenAI's shift from primarily relying on Microsoft Azure to a diversified cloud service strategy to meet its growing computing needs.
Amazon Web Services (AWS) has launched a new High Performance Computing (HPC) management service - AWS Parallel Computing Service, aimed at simplifying the process for businesses to access and utilize supercomputing resources, lowering the barrier to entry. This service utilizes the open-source tool Slurm to manage clusters, allowing users to easily set up and manage clusters of Amazon Elastic Compute Cloud (EC2) instances on AWS without complex system administrator support. Businesses only need one AWS account to experiment and test the advantages of their workloads in large-scale computing.
AWS will introduce NVIDIA GH200 Grace Hopper Superchips for the first time, providing scalable AI computing in the cloud. The collaboration includes hosting the first NVIDIA DGX Cloud with GH200 NVL32 on AWS, accelerating generative AI and language model training. New Amazon EC2 instances will be launched, including the P5e instance for generative AI and G6 and G6e instances suitable for various applications. The collaboration also involves software development, including NV
Amazon AWS has launched the Amazon EC2 Capacity Blocks service, allowing users to reserve Nvidia H100 GPU resources on demand. Users can reserve 1 to 64 instances, each with 8 GPUs, for a maximum of 14 days. Prices will dynamically change based on supply and demand for GPU resources, enabling users to adjust according to their needs and budget. This service is initially available in the AWS US East region and aims to alleviate the shortage of GPU resources for AI projects. This new on-demand purchasing and scheduling method allows users to...
The AWS EC2 Pricing MCP Server provides real-time EC2 price query services and supports direct use through Docker or Python.
An MCP server based on the AWS SDK for retrieving and listing running EC2 instances in a specified region.
AWS Security MCP is a service based on the model context protocol that allows AI assistants to independently check and analyze security issues in AWS infrastructure through natural language queries. It supports multiple AWS security services, including IAM, EC2, S3, etc., and provides security recommendations and threat modeling reports.
MCP - AWS is an AI - based application that manages AWS EC2 instances through OpenAI agents and the MCP server, supporting natural language commands to create and terminate instances.
Learn how to build and deploy a remote MCP server on AWS EC2 for remote use.
The AWS MCP server is a platform that provides tools for interacting with AWS services, supporting operations on core services such as S3, EC2, and RDS, as well as cost analysis and monitoring functions.
A tool that provides AWS resource creator tracking for AI assistants through the MCP protocol, which can quickly query the creator, time, method, and cost of resources such as EC2, RDS, and S3, helping to reduce cloud resource waste.