MULTIMODAL RETRIEVAL AUGMENTED GENERATION SYSTEM

Loading...
Thumbnail Image

Date

2025-01-23

Authors

Yuvzhenko, Denys

Journal Title

Journal ISSN

Volume Title

Publisher

Manuscript

Abstract

This study presents an asynchronous, web-based Retrieval-Augmented Generation (RAG) system that integrates multimodal inputs (text, images, tables) to enhance information retrieval and generation in various contexts. The system is developed in Python and hosted on AWS, combining Chroma DB for vector storage and Anthropic’s Claude-3-Haiku LLM model, which is accessible via AWS Bedrock. By leveraging modern cloud capabilities, the solution scales efficiently and handles diverse data modalities in real time. Through systematic experiments, this project highlights the effectiveness of multimodal embedding techniques for refining retrieval accuracy and providing context-aware responses. The architecture’s modular design supports seamless feature integration, making it adaptable for different use cases such as customer support, educational tools, and content creation. Key findings emphasize the role of vector databases in dynamic information updates and confirm that large language models, when appropriately curated and grounded, can produce high-quality, relevant outputs

Description

Keywords

Retrieval-Augmented Generation, Multimodal, Asynchronous Architecture, AWS, LLM, RAG

Citation

Yuvzhenko, Denys. (2025). MULTIMODAL RETRIEVAL AUGMENTED GENERATION SYSTEM. Kyiv: American University Kyiv. URI: https://er.auk.edu.ua/handle/234907866/164