AWS announces powerful new capabilities for Amazon Bedrock and biggest expansion of models to date
During his re:Invent keynote, Swami Sivasubramanian, VP of AI and Data at AWS, introduced new innovations for Amazon Bedrock, a fully managed service for building and scaling generative AI applications. These updates offer customers greater flexibility to deploy production-ready AI faster.
Key innovations include:
- Access to 100+ models from leading AI companies via the Amazon Bedrock Marketplace.
- New capabilities to better manage prompts at scale.
- Enhanced Amazon Bedrock Knowledge Bases, with structured data retrieval and GraphRAG.
- Amazon Bedrock Data Automation for efficient unstructured data processing.
The Amazon Bedrock Marketplace is available today, with other features like inference management and data automation in preview. Models from Luma AI, Poolside, and Stability AI are coming soon.
New Amazon Bedrock Data Automation
AWS announced Amazon Bedrock Data Automation, a capability to extract, transform, and generate data from unstructured content using a single API. It processes documents, images, audio, and videos, transforming them into structured formats for use cases like intelligent document processing, video analysis, and retrieval-augmented generation (RAG). Customers can generate outputs using predefined defaults, such as video scene descriptions or audio transcripts, or customize outputs to match their data schema for seamless integration into databases. Integrated with Knowledge Bases, it enhances RAG applications by parsing embedded text and images, improving accuracy and relevance.
With confidence scores and responses grounded in original content, Amazon Bedrock Data Automation mitigates hallucinations and increases transparency. Now available in preview, it streamlines unstructured data handling at scale.
New capabilities for Amazon Bedrock Knowledge Bases
Amazon Bedrock Knowledge Bases simplifies foundation model customization with contextual data using retrieval-augmented generation (RAG). Expanding beyond data sources like Amazon OpenSearch Serverless and Amazon Aurora, AWS introduces two powerful features.
Structured Data Retrieval enables querying structured data from sources like Amazon S3 and Redshift using natural language prompts, which are translated into SQL queries. This reduces generative AI app development time from months to days, breaking down data silos and enhancing response accuracy.
GraphRAG uses Amazon Neptune to create and traverse knowledge graphs, linking relationships between data for precise, relevant responses. Without requiring graph expertise, it reveals connections and improves transparency in response generation.
Structured data retrieval and GraphRAG in Amazon Bedrock Knowledge Bases are available in preview.
New Amazon Bedrock capabilities to help customers more effectively manage prompts at scale
Developers often face trade-offs between accuracy, cost, and latency when selecting models. To simplify this process, AWS introduces two new features for Amazon Bedrock to optimize prompt management:
Prompt Caching reduces latency and costs by securely caching frequently used prompts. This minimizes repeated processing, cutting costs by up to 90% and latency by up to 85%. For instance, a law firm’s AI chat app can cache sections of a document queried multiple times, processing them once and reusing the results, significantly lowering costs.
Intelligent Prompt Routing dynamically routes prompts to the most suitable model within a family based on response quality and cost predictions. This feature reduces costs by up to 30% while maintaining accuracy, ensuring an optimal balance for customers.
Access to more than 100 popular, emerging, and specialized models with Amazon Bedrock Marketplace
Through the new Amazon Bedrock Marketplace capability, AWS is giving access to more than 100 popular, emerging, and specialized models, so customers can find the right set of models for their use case. This includes popular models such as Mistral AI’s Mistral NeMo Instruct 2407, Technology Innovation Institute’s Falcon RW 1B, and NVIDIA NIM microservices, along with a wide array of specialized models, including Writer’s Palmyra-Fin for the financial industry, Upstage’s Solar Pro for translation, Camb.ai’s text-to-audio MARS6, and EvolutionaryScale’s ESM3 generative model for biology.
Once a customer finds a model they want, they select the appropriate infrastructure for their scaling needs and easily deploy on AWS through fully managed endpoints. Customers can then securely integrate the model with Amazon Bedrock’s unified application programming interfaces (APIs), leverage tools like Guardrails and Agents, and benefit from built-in security and privacy features.
Amazon Bedrock Marketplace is available today.
New in Amazon Bedrock: The broadest selection of models from leading AI companies
Luma AI’s multimodal models, including the Luma Ray 2, are transforming video content creation with generative AI. AWS will be the first cloud provider to offer Ray 2, enabling customers to generate high-quality, realistic videos from text and images with cinematic quality. Users can experiment with camera angles and create videos for industries like architecture, fashion, film, and music. AWS will also provide access to Poolside’s Malibu and Point models, specializing in code generation, testing, and real-time code completion, as well as Poolside’s Assistant for integrated development environments (IDEs). Additionally, AWS will offer Stable Diffusion 3.5 Large from Stability AI, a model for generating high-quality images from text. Amazon Bedrock will also soon feature Amazon Nova models, offering industry-leading intelligence and performance across a range of tasks.
New Amazon SageMaker AI capabilities reimagine how customers build and scale generative AI and machine learning models
AWS announced four new innovations for Amazon SageMaker AI to help customers accelerate generative AI model development. These include three updates for SageMaker HyperPod, which scales generative AI model training across thousands of AI accelerators, cutting training time by up to 40%. New training recipes help customers quickly get started with popular models like Llama and Mistral, simplifying the optimization process. Flexible training plans allow customers to manage compute capacity, meeting timelines and budgets. With the new SageMaker HyperPod task governance innovation, customers can maximize accelerator utilization for model training, fine-tuning, and inference, reducing model development costs by up to 40%. Additionally, a new integration with partner applications like Comet and Fiddler simplifies deploying generative AI and ML tools within SageMaker, enhancing flexibility and reducing onboarding time. All innovations are now generally available.
All of the new SageMaker innovations are generally available to customers today. Learn more about Amazon SageMaker.