Scale generative AI use cases, Part 1: Multi-tenant hub and spoke architecture using AWS Transit Gateway

Generative AI continues to reshape how businesses approach innovation and problem-solving. Customers are moving from experimentation to scaling generative AI use cases across their organizations, with more businesses fully integrating these technologies into their core processes. This evolution spans across lines of business (LOBs), teams, and software as a service (SaaS) providers. Although many AWS…

Read More

AWS AI infrastructure with NVIDIA Blackwell: Two powerful compute solutions for the next frontier of AI

Imagine a system that can explore multiple approaches to complex problems, drawing on its understanding of vast amounts of data, from scientific datasets to source code to business documents, and reasoning through the possibilities in real time. This lightning-fast reasoning isn’t waiting on the horizon. It’s happening today in our customers’ AI production environments. The…

Read More

A New Kind of AI Model Lets Data Owners Take Control

A new kind of large language model, developed by researchers at the Allen Institute for AI (Ai2), makes it possible to control how training data is used even after a model has been built. The new model, called FlexOlmo, could challenge the current industry paradigm of big artificial intelligence companies slurping up data from the…

Read More

Unlock retail intelligence by transforming data into actionable insights using generative AI with Amazon Q Business

Businesses often face challenges in managing and deriving value from their data. According to McKinsey, 78% of organizations now use AI in at least one business function (as of 2024), showing the growing importance of AI solutions in business. Additionally, 21% of organizations using generative AI have fundamentally redesigned their workflows, showing how AI is…

Read More

Improve conversational AI response times for enterprise applications with the Amazon Bedrock streaming API and AWS AppSync

Many enterprises are using large language models (LLMs) in Amazon Bedrock to gain insights from their internal data sources. Amazon Bedrock is a fully managed service that offers a choice of high-performing foundation models (FMs) from leading AI companies like AI21 Labs, Anthropic, Cohere, Meta, Mistral AI, Stability AI, and Amazon through a single API,…

Read More

Configure fine-grained access to Amazon Bedrock models using Amazon SageMaker Unified Studio

Enterprises adopting advanced AI solutions recognize that robust security and precise access control are essential for protecting valuable data, maintaining compliance, and preserving user trust. As organizations expand AI usage across teams and applications, they require granular permissions to safeguard sensitive information and manage who can access powerful models. Amazon SageMaker Unified Studio addresses these…

Read More

Query Amazon Aurora PostgreSQL using Amazon Bedrock Knowledge Bases structured data

Amazon Bedrock Knowledge Bases offers a fully managed Retrieval Augmented Generation (RAG) feature that connects large language models (LLMs) to internal data sources. This feature enhances foundation model (FM) outputs with contextual information from private data, making responses more relevant and accurate. At AWS re:Invent 2024, we announced Amazon Bedrock Knowledge Bases support for natural…

Read More