Akhil-Theerthala commited on
Commit
1ce6116
·
verified ·
1 Parent(s): d86e138

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -4
README.md CHANGED
@@ -3,17 +3,90 @@ library_name: transformers
3
  tags:
4
  - finance
5
  - personal_finance
 
 
6
  license: mit
7
  datasets:
8
  - Akhil-Theerthala/PersonalFinance_v2
9
  base_model:
10
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
 
11
  ---
12
 
 
 
 
 
13
  ## Model Details
14
 
15
- ### Model Description
16
- This model is trained upon the personal finance dataset accesible here.
17
- This is a LoRA adapted version of the model.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
 
19
- {Further details will be added soon}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  tags:
4
  - finance
5
  - personal_finance
6
+ - lora
7
+ - Deepseek-Distill
8
  license: mit
9
  datasets:
10
  - Akhil-Theerthala/PersonalFinance_v2
11
  base_model:
12
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
13
+ pipeline_tag: text-generation
14
  ---
15
 
16
+ # Personal Finance Reasoning LoRA Model
17
+
18
+ This model is a LoRA-adapted version of `deepseek-ai/DeepSeek-R1-Distill-Qwen-14b`, specifically fine-tuned to enhance reasoning capabilities for complex queries within the personal finance domain.
19
+
20
  ## Model Details
21
 
22
+ - **Developed by:** Akhil-Theerthala
23
+ - **Model type:** Language model fine-tuned for causal language modeling and reasoning.
24
+ - **Language(s) (NLP):** English
25
+ - **License:** MIT
26
+ - **Finetuned from model:** `deepseek-ai/DeepSeek-R1-Distill-Qwen-14b`
27
+ - **Model Description:** Built upon the `deepseek-ai/DeepSeek-R1-Distill-Qwen-14b` base model, this adaptation is specifically engineered to address nuanced personal finance questions. By fine-tuning with the `Akhil-Theerthala/PersonalFinance_v2` dataset, which comprises diverse scenarios requiring contextual understanding and multi-step reasoning, the model aims to generate more relevant and coherent responses to financial inquiries. The use of LoRA facilitates this specialization efficiently, targeting key aspects of financial reasoning such as query analysis and context processing inherent in the training data.
28
+ - **Resources for more information:**
29
+ - Dataset: [Akhil-Theerthala/PersonalFinance_v2](https://huggingface.co/datasets/Akhil-Theerthala/PersonalFinance_v2)
30
+ - Base Model: [deepseek-ai/DeepSeek-R1-Distill-Qwen-14b](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14b)
31
+
32
+ ## Training Data
33
+ The model was fine-tuned on the `Akhil-Theerthala/PersonalFinance_v2` dataset.
34
+
35
+ - **Dataset Description:** This dataset contains approximately ~7.04k samples designed for personal finance reasoning. It includes diverse scenarios and queries related to financial advice, planning, and problem-solving. The data is structured to facilitate tasks like question answering and text generation requiring financial reasoning, often involving analysis of context and complex query structures.
36
+ - **Data Size:** ~7,000 training instances.
37
+ - **Format:** JSON
38
+ - **Source:** [Akhil-Theerthala/PersonalFinance_v2 on Hugging Face](https://huggingface.co/datasets/Akhil-Theerthala/PersonalFinance_v2)
39
+
40
+ ## Training Procedure
41
+
42
+ ### Adaptation Method:
43
+ The base model `DeepSeek-R1-Distill-Qwen-14b` (14B parameters) was adapted using Low-Rank Adaptation (LoRA). This method was chosen due to the dataset size (~7k samples), making it suitable for parameter-efficient fine-tuning rather than full fine-tuning.
44
+
45
+ ### Hyperparameters:
46
+ - **LoRA r (rank):** 64
47
+ - **LoRA alpha:** 128
48
+ - **LoRA dropout:** 0.1
49
+ - **Target modules:** all-linear
50
+ - **Optimizer:** Adam
51
+ - **Learning rate:** 1e-4
52
+ - **Batch size:** 40
53
+ - **Number of epochs:** 8
54
 
55
+
56
+ ## Evaluation
57
+ - **The Evaluation is currently under process, with the major method of evaluation using LLMs-as-Judge framework, preferably Gemini-2.5-Flash as the judge. The Benchmarks used and the comparisons done will be shared shortly.**
58
+ ### Intended Uses:
59
+ - Assisting users with personal finance queries by providing reasoned explanations and advice.
60
+ - Educational tool for understanding financial concepts.
61
+ - Generating text related to financial scenarios.
62
+ - Research in domain-specific reasoning and parameter-efficient fine-tuning.
63
+
64
+ ### Limitations:
65
+
66
+ - **Knowledge Cutoff:** The model's knowledge is limited to the information present in its training data (both the base model's pre-training and the `PersonalFinance_v2` fine-tuning dataset). It may not be aware of very recent financial events, regulations, or products.
67
+ - **Potential Biases:** The model may reflect biases present in the training data.
68
+ - **Hallucinations:** Like all large language models, it may occasionally generate plausible-sounding but incorrect information (hallucinations).
69
+ - **Domain Specificity:** While specialized for personal finance, its reasoning capabilities might be less robust outside this domain or for highly niche financial topics not well-represented in the training data.
70
+ - **Dataset Limitations:** The current version of the dataset is only ~7k which is barely enough to get quality results from LoRA adaptation. Further refinement and scaling up of the dataset is necessary.
71
+
72
+ **Cost Incurred:** As an independent data scientist trying to develop this project, the entire cost of development till now totals up to ~$80. The cost was mostly reduced by the usage of free/open-source methods whenever available. (Though those had resource constraints with my 16GB M2 Mac Mini). Further development of this process, definitely needs some better, more cost-efficient approaches.
73
+
74
+ ## Further Information & Collaboration
75
+ - **Contact:** akhiltvsn@gmail.com
76
+ - **Future Work:**
77
+ - Refining and expanding the `PersonalFinance_v2` dataset. From 7k to at least 50k samples.
78
+ - Exploring Mixture of Experts (MoE) methods for further model development.
79
+ - **Call for Collaboration:** I am a solo dude just randomly working on this project during my free time. If you are interested in this project, and want to expand the scope, then do ping me here, on Linkedin or just send me a mail.
80
+
81
+ ---
82
+ ## Citation
83
+ ```BibTex
84
+ @misc{akhil_theerthala_2025,
85
+ author = { Akhil Theerthala },
86
+ title = { Kuvera-14B-v0.1.0 (Revision 3fb04b9) },
87
+ year = 2025,
88
+ url = { https://huggingface.co/Akhil-Theerthala/Kuvera-14B-v0.1.0 },
89
+ doi = { 10.57967/hf/5707 },
90
+ publisher = { Hugging Face }
91
+ }
92
+ ```