Intellema
Back to Case Studies

Case Study

Multimodal Conversational AI for Qi-Card

Multimodal Conversational AI for Qi-Card

Category:

Conversational AI, Financial Technology, & RAG Systems

Impact:

2 Months | Hybrid | Contract

Background

Financial institutions often struggle with managing millions of daily customer interactions efficiently while maintaining context awareness and accuracy. Qi-Card, a leading Iraqi financial services provider, required an AI-driven chatbot capable of handling 2M+ daily requests, interpreting text and image inputs, and ensuring uninterrupted service reliability. A scalable, multimodal conversational system was needed—one that could integrate LLMs, RAG pipelines, and fallback intelligence to deliver a seamless customer experience.

Project Goals

  • Build and scale a customer-facing chatbot capable of managing 2M+ user requests daily
  • Implement a fallback system using AWS Bedrock and large language models (LLMs) for high availability
  • Develop a multimodal ingestion pipeline for screenshots and receipt processing
  • Architect a Retrieval-Augmented Generation (RAG) system for contextual response generation
  • Ensure code reliability and performance through comprehensive testing

Our Approach

Fallback Intelligence & LLM Integration

Developed a robust fallback mechanism leveraging AWS Bedrock and LLMs to maintain smooth user interaction during API failures and edge cases. Optimized request routing for resilience and reduced latency under peak load.

Multimodal Input Processing

Created an advanced image ingestion feature enabling the chatbot to understand screenshots, receipts, and document images. Integrated computer vision and OCR modules to extract contextual data for more accurate and personalized responses.

RAG Pipeline & Orchestration

Architected a dedicated Retrieval-Augmented Generation (RAG) pipeline for contextual data retrieval and response enrichment. Employed Apache Airflow for task orchestration, monitoring, and optimization of RAG execution cycles.

Testing & Reliability

Developed comprehensive unit and integration tests to ensure code reliability, minimize regression, and maintain performance across continuous deployments.

Key Results

  • Scaled chatbot infrastructure to handle 2M+ daily requests seamlessly
  • Achieved 99.9% uptime with intelligent fallback routing via AWS Bedrock
  • Enabled multimodal input handling, improving service accuracy by 20%
  • Deployed RAG pipelines orchestrated through Airflow for faster, context-aware retrieval
  • Strengthened codebase reliability through extensive automated testing

Technologies Used

AWS Bedrock & LLMs
AWS Bedrock & LLMs
Apache Airflow
Python
Python
OpenCV & OCR Tools
OpenCV & OCR Tools
PyTest / UnitTest
AWS Infrastructure (EC2, S3, Lambda)
AWS Infrastructure (EC2, S3, Lambda)

Connect with Intellema

Contact Us
Intellema – Intelligence Beyond Hype