Accelerated PyTorch inference with torch.compile on AWS Graviton processors

Host the Whisper Model on Amazon SageMaker: exploring inference options

Enable faster training with Amazon SageMaker data parallel library

Optimize AWS Inferentia utilization with FastAPI and PyTorch models on Amazon EC2 Inf1 & Inf2 instances

Fine-tune GPT-J using an Amazon SageMaker Hugging Face estimator and the model parallel library

Host ML models on Amazon SageMaker using Triton: CV model with PyTorch backend