Member-only story

Multimodal Batch Inference on Amazon Bedrock with Anthropic Claude 3.5 Sonnet

Explore Amazon Bedrock’s batch inference capabilities with multimodal models like Anthropic Claude 3.5 Sonnet to generate cost-effective bulk image titles, descriptions, and keywords.

Gary A. Stafford
14 min readNov 20, 2024

AWS initially announced Batch inference for Amazon Bedrock in Preview by AWS in November 2023. In August 2024, AWS announced the general availability of batch inference for Amazon Bedrock in all supported AWS regions for supported models. The big news was that batch inference is 50% of on-demand inference pricing. With the return on investment for many customers’ generative AI projects still being analyzed, spending 50% less for inference can significantly impact project feasibility.

Amazon Bedrock Console showing the Batch inference tab

AWS documentation states, “With batch inference, you can submit multiple prompts and generate responses asynchronously. Batch inference helps you process a large number of requests efficiently by sending a single request and generating the responses in an Amazon S3 bucket. The official AWS documentation and code examples on GitHub demonstrate using Amazon Bedrock batch inference with common text-to-text use cases, such as analyzing CloudTrail logs and customer call transcripts. However, Amazon Bedrock batch…

--

--

Gary A. Stafford
Gary A. Stafford

Written by Gary A. Stafford

Area Principal Solutions Architect @ AWS | 10x AWS Certified Pro | Polyglot Developer | DataOps | GenAI | Technology consultant, writer, and speaker

No responses yet