Demo / README.md
HanLee's picture
feat: 02_06e
36ba8c8
|
raw
history blame
1.72 kB

Hands-On AI: Building and Deploying LLM-Powered Apps

This is the repository for the LinkedIn Learning course Hands-On AI: Building and Deploying LLM-Powered Apps. The full course is available from [LinkedIn Learning][lil-course-url].

See the readme file in the main branch for updated instructions and information.

Lab4: Indexing Documents into Vector Database

In the previous lab, we enabled document loading and chunking them into smaller sub documents. Now, we will need to index them into our search engine vector databse in order for us to build our Chat with PDF application using the RAG (Retrieval Augmented Generation) pattern.

In this lab, we will implement adding OpenAI's embedding model and index the documents we chunked in the previous section into a Vector Database. We will be using Chroma as the vector database of choice. Chroma is a lightweight embedding database that can live in memory, similar to SQLite.

Exercises

We will build on top of our existing chainlit app code in app/app.py in the app folder. As in our previous app, we added some template code and instructions in app/app.py

  1. Please go through the exercises in app/app.py.

  2. Please lanuch the application by running the following command on the Terminal:

chainlit run app/app.py -w

Solution

Please see app/app.py.

Alternatively, to launch the application, please run the following command on the Terminal:

chainlit run app/app.py -w

References