AWS re:Invent 2024 – Day 3 Recap

AWS announces powerful new capabilities for Amazon Bedrock and biggest expansion of models to date 

During his re:Invent keynote, Swami Sivasubramanian, VP of AI and Data at AWS, introduced new innovations for Amazon Bedrock, a fully managed service for building and scaling generative AI applications. These updates offer customers greater flexibility to deploy production-ready AI faster. 

Key innovations include: 

  • Access to 100+ models from leading AI companies via the Amazon Bedrock Marketplace
  • New capabilities to better manage prompts at scale. 
  • Enhanced Amazon Bedrock Knowledge Bases, with structured data retrieval and GraphRAG. 
  • Amazon Bedrock Data Automation for efficient unstructured data processing. 

The Amazon Bedrock Marketplace is available today, with other features like inference management and data automation in preview. Models from Luma AI, Poolside, and Stability AI are coming soon. 

New Amazon Bedrock Data Automation 

AWS announced Amazon Bedrock Data Automation, a capability to extract, transform, and generate data from unstructured content using a single API. It processes documents, images, audio, and videos, transforming them into structured formats for use cases like intelligent document processing, video analysis, and retrieval-augmented generation (RAG). Customers can generate outputs using predefined defaults, such as video scene descriptions or audio transcripts, or customize outputs to match their data schema for seamless integration into databases. Integrated with Knowledge Bases, it enhances RAG applications by parsing embedded text and images, improving accuracy and relevance. 

With confidence scores and responses grounded in original content, Amazon Bedrock Data Automation mitigates hallucinations and increases transparency. Now available in preview, it streamlines unstructured data handling at scale. 

New capabilities for Amazon Bedrock Knowledge Bases 

Amazon Bedrock Knowledge Bases simplifies foundation model customization with contextual data using retrieval-augmented generation (RAG). Expanding beyond data sources like Amazon OpenSearch Serverless and Amazon Aurora, AWS introduces two powerful features. 

Structured Data Retrieval enables querying structured data from sources like Amazon S3 and Redshift using natural language prompts, which are translated into SQL queries. This reduces generative AI app development time from months to days, breaking down data silos and enhancing response accuracy. 

GraphRAG uses Amazon Neptune to create and traverse knowledge graphs, linking relationships between data for precise, relevant responses. Without requiring graph expertise, it reveals connections and improves transparency in response generation. 

Structured data retrieval and GraphRAG in Amazon Bedrock Knowledge Bases are available in preview.  

New Amazon Bedrock capabilities to help customers more effectively manage prompts at scale 

Developers often face trade-offs between accuracy, cost, and latency when selecting models. To simplify this process, AWS introduces two new features for Amazon Bedrock to optimize prompt management: 

Prompt Caching reduces latency and costs by securely caching frequently used prompts. This minimizes repeated processing, cutting costs by up to 90% and latency by up to 85%. For instance, a law firm’s AI chat app can cache sections of a document queried multiple times, processing them once and reusing the results, significantly lowering costs. 

Intelligent Prompt Routing dynamically routes prompts to the most suitable model within a family based on response quality and cost predictions. This feature reduces costs by up to 30% while maintaining accuracy, ensuring an optimal balance for customers. 

Access to more than 100 popular, emerging, and specialized models with Amazon Bedrock Marketplace 

Through the new Amazon Bedrock Marketplace capability, AWS is giving access to more than 100 popular, emerging, and specialized models, so customers can find the right set of models for their use case. This includes popular models such as Mistral AI’s Mistral NeMo Instruct 2407, Technology Innovation Institute’s Falcon RW 1B, and NVIDIA NIM microservices, along with a wide array of specialized models, including Writer’s Palmyra-Fin for the financial industry, Upstage’s Solar Pro for translation, Camb.ai’s text-to-audio MARS6, and EvolutionaryScale’s ESM3 generative model for biology. 

Once a customer finds a model they want, they select the appropriate infrastructure for their scaling needs and easily deploy on AWS through fully managed endpoints. Customers can then securely integrate the model with Amazon Bedrock’s unified application programming interfaces (APIs), leverage tools like Guardrails and Agents, and benefit from built-in security and privacy features. 

Amazon Bedrock Marketplace is available today. 

New in Amazon Bedrock: The broadest selection of models from leading AI companies 

Luma AI’s multimodal models, including the Luma Ray 2, are transforming video content creation with generative AI. AWS will be the first cloud provider to offer Ray 2, enabling customers to generate high-quality, realistic videos from text and images with cinematic quality. Users can experiment with camera angles and create videos for industries like architecture, fashion, film, and music. AWS will also provide access to Poolside’s Malibu and Point models, specializing in code generation, testing, and real-time code completion, as well as Poolside’s Assistant for integrated development environments (IDEs). Additionally, AWS will offer Stable Diffusion 3.5 Large from Stability AI, a model for generating high-quality images from text. Amazon Bedrock will also soon feature Amazon Nova models, offering industry-leading intelligence and performance across a range of tasks. 

New Amazon SageMaker AI capabilities reimagine how customers build and scale generative AI and machine learning models 

AWS announced four new innovations for Amazon SageMaker AI to help customers accelerate generative AI model development. These include three updates for SageMaker HyperPod, which scales generative AI model training across thousands of AI accelerators, cutting training time by up to 40%. New training recipes help customers quickly get started with popular models like Llama and Mistral, simplifying the optimization process. Flexible training plans allow customers to manage compute capacity, meeting timelines and budgets. With the new SageMaker HyperPod task governance innovation, customers can maximize accelerator utilization for model training, fine-tuning, and inference, reducing model development costs by up to 40%. Additionally, a new integration with partner applications like Comet and Fiddler simplifies deploying generative AI and ML tools within SageMaker, enhancing flexibility and reducing onboarding time. All innovations are now generally available. 

All of the new SageMaker innovations are generally available to customers today. Learn more about Amazon SageMaker

AWS re:Invent 2024 – Day 2 Recap

AWS Trainium2 instances now generally available

AWS announced the general availability of AWS Trainium2-powered Amazon EC2 instances, designed for high-performance deep learning and generative AI workloads. These instances deliver 30-40% better price performance than GPU-based EC2 instances. With 16 Trainium2 chips and ultra-fast NeuronLink, they provide 20.8 petaflops of compute, ideal for training large models. 

For larger models, Trn2 UltraServers connect four Trn2 instances, enabling scale across 64 Trainium2 chips. These servers accelerate training, improve inference performance, and power trillion-parameter models. Project Rainier, a collaboration with Anthropic, will create the world’s largest AI compute cluster using these UltraServers. 

Trn2 instances are generally available today in the US East (Ohio) AWS Region, with availability in additional regions coming soon. Trn2 UltraServers are available in preview. 

Trainium3 chips—designed for high-performance needs of next frontier of generative AI workloads

AWS announced Trainium3, its next generation AI chip, that will allow customers to build bigger models faster and deliver superior real-time performance when deploying them. It will be the first AWS chip made with a 3-nanometer process node, setting a new standard for performance, power efficiency, and density. Trainium3-powered UltraServers are expected to be four times more performant than Trn2 UltraServers, allowing customers to iterate even faster when building models and deliver superior real-time performance when deploying them. The first Trainium3-based instances are expected to be available in late 2025. 

New database capabilities announced including Amazon Aurora DSQL—the fastest distributed SQL database 

AWS announced enhancements to Amazon Aurora and Amazon DynamoDB, offering strong consistency, low latency, and global scalability. 

  • Amazon Aurora DSQL: This serverless, distributed SQL database supports 99.999% multi-Region availability and delivers four times faster reads and writes compared to other distributed SQL databases. It eliminates trade-offs between low latency and SQL while providing microsecond sync accuracy. 
  • DynamoDB Enhancements: DynamoDB global tables now offer strong consistency, ensuring real-time access to the latest data without code changes. 

Both Aurora DSQL and DynamoDB enhancements are in preview. 

Introducing Amazon Nova 

Amazon Nova introduces a new generation of foundation models (FMs) capable of processing text, images, and videos. The Amazon Nova models available in Amazon Bedrock include: Amazon Nova Micro, Amazon Nova Lite, Amazon Nova Pro, Amazon Nova Premier, Amazon Nova Canvas, and Amazon Nova Reel – enable applications for multimedia understanding and generation. These models, available through Amazon Bedrock, are designed for speed, cost-efficiency, and ease of integration with customers’ systems. 

Amazon Q Developer reimagines how developers build and operate software with generative AI 

Amazon Q Developer enhancements leverage generative AI to improve software development and operations: 

  • Automated Unit Tests: Amazon Q Developer automates the creation of unit tests, reducing the burden on developers and ensuring complete test coverage with less effort. This helps developers deliver reliable code faster and avoid costly rollbacks. 
  • Documentation Updates: Automates the creation and updating of project documentation, ensuring accuracy and reducing the time developers spend on understanding code. This enables quicker onboarding and more meaningful contributions from team members. 
  • Code Reviews: Amazon Q Developer automates code reviews, providing quick feedback to help developers maintain quality, style, and security standards. This speeds up the review process, saving time and allowing developers to resolve issues earlier. 
  • Operational Issue Resolution: Operational teams can quickly identify and resolve issues across AWS environments with Amazon Q Developer by analyzing vast data points to uncover service relationships and anomalies. It provides actionable hypotheses and guides users through fixes, streamlining issue resolution and reducing downtime. 

These capabilities are now available in IDEs, AWS Management Console, and through GitLab integration. 

Next generation of Amazon SageMaker to deliver unified platform for data, analytics, and AI 

AWS CEO Matt Garman unveiled the next generation of Amazon SageMaker. The revamped Amazon SageMaker integrates analytics, machine learning, and generative AI into a unified platform: 

  • SageMaker Unified Studio: The new unified studio provides a single environment for accessing and acting on data, integrating AWS analytics, ML, and AI tools. Amazon Q Developer helps customers tackle various data use cases with the best tools for the job. 
  • SageMaker Catalog: Amazon SageMaker Catalog provides secure access to data, models, and artifacts, ensuring compliance and enterprise security. Built on Amazon DataZone, it offers governance tools like data classification and toxicity detection to safeguard AI applications. 
  • SageMaker Lakehouse: Amazon SageMaker Lakehouse unifies data across S3, data lakes, Redshift, and federated sources, simplifying analytics and ML tool usage. It supports Apache Iceberg for seamless data processing and offers fine-grained access controls for secure data sharing.  
  • Zero-ETL Integrations: AWS’s zero-ETL integrations with SaaS applications like Zendesk and SAP simplify data access for analytics and AI in SageMaker Lakehouse and Redshift. This eliminates the need for complex data pipelines, speeding up insights and reducing costs. 

The new SageMaker platform enhances collaboration, security, and efficiency for data and AI projects.

AWS strengthens Amazon Bedrock with industry-first AI safeguard, new agent capability, and model customization 

AWS CEO Matt Garman unveiled new Amazon Bedrock capabilities to address key challenges in deploying generative AI. These features tackle hallucination-induced errors, enable orchestration of AI agents for complex tasks, and support smaller, cost-efficient models that rival large models in performance: 

  • Prevent factual errors due to hallucinations: Generative AI models can produce “hallucinations,” limiting trust in critical industries. Amazon Bedrock’s Automated Reasoning checks prevent errors using logical reasoning, ensuring accurate, auditable, and policy-aligned responses via Bedrock Guardrails. 
  • Easily build and coordinate multiple agents to execute complex workflows: Amazon Bedrock Agents enable applications to execute tasks by leveraging AI-powered agents. AWS now supports multi-agent collaboration, allowing customers to coordinate specialized agents for complex workflows, such as financial analysis, across systems and data sources with ease. 
  • Create smaller, faster, more cost-effective models: Amazon Bedrock Model Distillation lets customers create smaller, efficient models by transferring knowledge from larger models, balancing performance, cost, and latency—ideal for real-time applications. It works with models from Anthropic, Meta, and Amazon Nova Models. 

Automated Reasoning checks, multi-agent collaboration, and Model Distillation are all available in preview. 

AWS re:Invent 2024 – Day 1 Recap

New generative AI enhancements for Amazon Connect 

New generative AI enhancements for Amazon Connect, AWS’s cloud contact center solution. Serving over 10 million daily interactions, Amazon Connect now offers: 

  • Automated segmentation for proactive, personalized communications. 
  • Amazon Q in Connect, a generative AI-powered assistant for dynamic self-service experiences. 
  • Customizable AI guardrails to ensure safe, policy-compliant AI deployments. 
  • Generative AI-driven insights like intelligent contact categorization and agent evaluations for better training and service quality. 

Leading organizations like Frontdoor, Fujitsu, and Priceline are already leveraging these innovations for enhanced customer service at reduced costs. 

These features are now generally available. Learn more about the AWS News Blog and AWS Contact Center Blog

AWS announces new data center components to support AI and improve energy efficiency 

AWS has unveiled advanced data center components to power the next generation of AI, enhance energy efficiency, and drive customer innovation. These upgrades address growing generative AI demands while improving sustainability. Key features include: 

  • Simplified designs to lower energy use and reduce failure risks. 
  • Cooling and control innovations, enabling 12% more compute power per site, reducing the number of data centers required. 
  • Sustainability upgrades, such as a cooling system cutting energy use by 46%, concrete with 35% lower embodied carbon, and backup generators running on renewable diesel, which reduces greenhouse gas emissions by up to 90%. 

These components, already in some AWS data centers, will be fully implemented in new U.S. facilities starting early 2025. Watch this video and read the press release to learn about AWS’s new data center components. 

Peter DeSantis shows how AWS is innovating across the entire technology stack 

At AWS re:Invent’s Monday Night Live, Peter DeSantis, SVP of AWS Utility Computing, explored the engineering behind AWS services and its role in advancing AI workloads. Joined by Dave Brown, VP of AWS Compute & Networking Services, and Tom Brown, co-founder of Anthropic, DeSantis showcased how AWS delivers performance, reliability, and cost-efficiency for AI. 

Highlights included innovations like the AWS Trainium2 chip, purpose-built for machine learning, and the Firefly Optic Plug, which speeds AI cluster deployment by pre-testing wiring. DeSantis emphasized AWS’s commitment to deep customer insights and fast, impactful decisions—like its pioneering investment in custom silicon 12 years ago. 

Calling this “the next chapter,” he detailed how AWS innovates across the tech stack to deliver differentiated solutions for the most demanding workloads. 

Schedule a meeting with our AWS cloud solution experts and accelerate your cloud journey with Idexcel. 

What to Expect from AWS re:Invent 2024

AWS re:Invent 2024, hosted by Amazon Web Services (AWS), is one of the most highly anticipated events in cloud computing. Taking place from December 2–6, 2024, in Las Vegas, this event is perfect for anyone looking to stay ahead of the latest AWS advancements and leverage transformative innovations such as generative AI, analytics, and cloud operations. Whether you’re a developer, business decision-maker, IT leader, or technical project manager, AWS re:Invent offers something for everyone. 

This premier annual learning event empowers attendees to enhance their expertise in AWS technologies through a wide range of experiences. From keynote presentations by AWS leaders to smaller innovation talks and hands-on working sessions led by AWS experts, the event delivers invaluable knowledge-sharing opportunities. Additionally, re:Invent includes engaging activities like live music and games at the re:Play celebration. Attendees can select learning tracks tailored to their interests and skill levels. 

Keynotes and innovation talks will be available for livestream viewing during the event and on-demand afterward. However, the highly interactive breakout sessions are exclusively designed for in-person participation. 

As an AWS Advanced Tier Services Partner and Managed Service Provider (MSP), Idexcel team is attending this flagship event and looks forward to engaging with the AWS community. 

Keynote Session 1: Monday Night Live with Peter DeSantis | Monday, December 2 | 7:30 PM – 9:00 PM (PST) 

Join Peter DeSantis, Senior Vice President of AWS Utility Computing, for the Monday Night Live tradition, where he will dive deep into the engineering that powers AWS services. Explore how AWS’s unique approach and innovative culture drive cutting-edge solutions across the spectrum—from silicon to services—while maintaining a strong focus on performance and cost efficiency. 

Keynote Session 2: CEO Keynote with Matt Garman | Tuesday, December 3 | 8:00 AM – 10:30 AM (PST) 

Join AWS CEO Matt Garman as he shares how AWS is driving innovation across every dimension of the world’s leading cloud platform. Learn how AWS is reimagining core technologies and introducing new experiences to empower customers and partners in building a better future. 

Keynote Session 3: Dr. Swami Sivasubramanian Keynote | Wednesday, December 4 | 8:30 AM – 10:30 AM (PST) 

Join Dr. Swami Sivasubramanian, VP of AI and Data at AWS, to explore how building a robust data foundation can drive innovation and deliver unique solutions for your customers. Gain insights from customer speakers as they share real-world examples of leveraging data, including generative AI, to create exceptional customer experiences. 

Keynote Session 4: AWS Partner Keynote with Dr. Ruba Borno | Wednesday, December 4 | 3:00 PM – 4:30 PM (PST) 

Join the AWS Partner Keynote with Dr. Ruba Borno, Vice President of Global Specialists and Partners, as she explores the transformative impact of strategic partnerships. Learn how AWS and its partners are enabling innovative solutions and driving meaningful outcomes for customers. Hear directly from partners and customers about reimagined business models and success stories powered by AWS’s groundbreaking services. Discover how AWS is enhancing the digital experience for partners, creating high-value opportunities, and fostering long-term success. 

Keynote Session 5: Dr. Werner Vogels Keynote | Thursday, December 5 | 8:30 AM – 10:30 AM (PST) 

Join Dr. Werner Vogels, VP and CTO of Amazon.com, as he shares invaluable lessons and strategies for navigating the challenges of managing complex systems. In this keynote, he delves into the core principles of embracing complexity, drawing from Amazon’s expertise in building and scaling distributed systems at a massive scale. 

Allolankandy Anand(Vice President | Digital Transformation & Strategy at Idexcel) and his team will be attending this event to meet with customers and partners. Schedule a meeting with Allolankandy Anand to discover how Idexcel can deliver strategic and innovative cloud solutions to achieve your organization’s business goals.