Scale AI training and inference for drug discovery through Amazon EKS and Karpenter

Large language model inference over confidential data using AWS Nitro Enclaves

Introducing three new NVIDIA GPU-based Amazon EC2 instances

Amazon EC2 DL2q instance for cost-efficient, high-performance AI inference is now generally available

Enable pod-based GPU metrics in Amazon CloudWatch

Maximize Stable Diffusion performance and lower inference costs with AWS Inferentia2

Accelerate PyTorch with DeepSpeed to train large language models with Intel Habana Gaudi-based DL1 EC2 instances