GitXplorerGitXplorer
a

multimodal-rag-on-slide-decks

public
34 stars
6 forks
7 issues

Commits

List of commits on branch main.
Unverified
ea8586b78aa6064e883b742f4a32fe822f41b653

updated blog3

aainapudi committed 3 months ago
Unverified
c5783520c82297a645de0eb7e3da6580052ab40a

updated blog3

aainapudi committed 3 months ago
Unverified
edbb50c661f1f718d30af9e36ecc5d0ac82deeea

updated blog3

aainapudi committed 3 months ago
Unverified
98538a18edde8df0717567394fc8e5469056fabd

updated blog3

aainapudi committed 3 months ago
Verified
e16ac7a1b468ccbf77635f3ed38864ab31eff3ba

update the multimodal flowcharts

mmadhurprash committed 5 months ago
Verified
cd65cc2688dfbf14fcc0c9afb6cbbc4719e7e994

Delete Blog4-PDF-TitanEmbeddings/images/multimodal.drawio.xml

mmadhurprash committed 5 months ago

README

The README file for this repository.

About this repository

This is a 3-part series on the topic 'talk to your slide deck' using multimodal foundation models hosted on Amazon Bedrock and Amazon SageMaker.

  • Part 1 of the series implements a solution design that first converts slides into embeddings and stores these embeddings in a vector store (Amazon OpenSearch Serverless). When a user queries this data, LLaVA 1.5 is invoked and inference returned to user.
  • Part 2 of the series follows a different solution design. This approach will generate and store LLaVA 1.5 inferences in a vector store (Amazon OpenSearch Serverless) and use those stored inferences to respond to user queries.
  • Part 3 of the series will compare the two approaches.