Gen AI News Talk
Evaluating RAG applications with Amazon Bedrock knowledge base evaluation
Organizations building and deploying AI applications, particularly those using large language models (LLMs) with Retrieval Augmented Generation (RAG) systems, face a significant challenge: how...
How GoDaddy built a category generation system at scale with batch inference for Amazon...
This post was co-written with Vishal Singh, Data Engineering Leader at Data & Analytics team of GoDaddy
Generative AI solutions have the potential to transform...
Revolutionizing customer service: MaestroQA’s integration with Amazon Bedrock for actionable insight
This post is cowritten with Harrison Hunter is the CTO and co-founder of MaestroQA.
MaestroQA augments call center operations by empowering the quality assurance (QA)...
How to run Qwen 2.5 on AWS AI chips using Hugging Face libraries
The Qwen 2.5 multilingual large language models (LLMs) are a collection of pre-trained and instruction tuned generative models in 0.5B, 1.5B, 3B, 7B, 14B, 32B,...
Creating asynchronous AI agents with Amazon Bedrock
The integration of generative AI agents into business processes is poised to accelerate as organizations recognize the untapped potential of these technologies. Advancements in...
Benchmarking customized models on Amazon Bedrock using LLMPerf and LiteLLM
Open foundation models (FMs) allow organizations to build customized AI applications by fine-tuning for their specific domains or tasks, while retaining control over costs...
Exploring creative possibilities: A visual guide to Amazon Nova Canvas
Compelling AI-generated images start with well-crafted prompts. In this follow-up to our Amazon Nova Canvas Prompt Engineering Guide, we showcase a curated gallery of...
From fridge to table: Use Amazon Rekognition and Amazon Bedrock to generate recipes and...
In today’s fast-paced world, time is of the essence and even basic tasks like grocery shopping can feel rushed and challenging. Despite our best...
Deploy DeepSeek-R1 distilled models on Amazon SageMaker using a Large Model Inference container
DeepSeek-R1 is a large language model (LLM) developed by DeepSeek AI that uses reinforcement learning to enhance reasoning capabilities through a multi-stage training process...
Benchmarking Amazon Nova and GPT-4o models with FloTorch
Based on original post by Dr. Hemant Joshi, CTO, FloTorch.ai
A recent evaluation conducted by FloTorch compared the performance of Amazon Nova models with OpenAI’s...