Demo / README.md
HanLee's picture
feat: 03_04
ce5d5d0
|
raw
history blame
1.15 kB

Hands-On AI: Building and Deploying LLM-Powered Apps

This is the repository for the LinkedIn Learning course Hands-On AI: Building and Deploying LLM-Powered Apps. The full course is available from [LinkedIn Learning][lil-course-url].

See the readme file in the main branch for updated instructions and information.

Lab6: Setup Prompting

Now our Chat with PDF application is up and running, but we run into a very slight problem: one of the key questions is not working despite that there's ample information in the PDF documents!!!

We can "fix" this is via prompt engineering. Prompt Engineering refers to methods for how to communicate with LLM to steer its behavior for desired outcomes without updating the model weights.

Before we can do that, we need to extract the prompt template out of the code.

Exercises

Please extract Langchain's prompt template out of the code base into an independent prompt.py in the app directory. Control (or Command for Mac) click will help you navigate this very quickly!

References