Spaces:
Runtime error
Runtime error
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
---
|
| 2 |
title: Financial Bot
|
| 3 |
-
emoji:
|
| 4 |
colorFrom: red
|
| 5 |
-
colorTo:
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 4.16.0
|
| 8 |
app_file: app.py
|
|
@@ -10,4 +10,16 @@ pinned: false
|
|
| 10 |
license: mit
|
| 11 |
---
|
| 12 |
|
| 13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
title: Financial Bot
|
| 3 |
+
emoji: π
|
| 4 |
colorFrom: red
|
| 5 |
+
colorTo: green
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 4.16.0
|
| 8 |
app_file: app.py
|
|
|
|
| 10 |
license: mit
|
| 11 |
---
|
| 12 |
|
| 13 |
+
This is the Inference module of a 3-part FTI feature-training-inference RAG-framework LLMOps course. \
|
| 14 |
+
In this iteration, I've replaced Falcon 7B Instruct with the currently-SoTa (Jan '24) Mistral-7B-Instruct-v0.2, \
|
| 15 |
+
fine-tuned using Unsloth on financial questions and answers generated with the help of GPT-4, quantized \
|
| 16 |
+
and augmented with a 4bit QLoRa. \
|
| 17 |
+
\
|
| 18 |
+
Prompt analysis and model registry is handled by Comet LLM, and finance news is pulled via an Alpaca API, processed \
|
| 19 |
+
by Bytewax, and then sent as a vector embedding to Qdrant's serverless vector store. LangChain chains the prompt and \
|
| 20 |
+
most relevant news article to provide answers with real-time finance information embedded within the output. \
|
| 21 |
+
\
|
| 22 |
+
#TODO: Add citations to output to show end-user which article has been used to generate the output.
|
| 23 |
+
|
| 24 |
+
I have contributed to the original MIT licensed (ka-ching!) course which can be found here:
|
| 25 |
+
https://medium.com/decoding-ml/the-llms-kit-build-a-production-ready-real-time-financial-advisor-system-using-streaming-ffdcb2b50714
|