delight2004 commited on
Commit
da268f7
·
verified ·
1 Parent(s): 932928b

Revert: remove 'How to use' section from README.md per user request.

Browse files
Files changed (1) hide show
  1. README.md +42 -107
README.md CHANGED
@@ -10,114 +10,49 @@ language:
10
  base_model: LiquidAI/LFM2-1.2B
11
  pipeline_tag: text-generation
12
  ---
13
- #
14
- lfm2-1.2b-sermon-instruct-qlora
15
- **
16
- Author
17
- **
18
- : Delight Aheebwa
19
-
20
- **
21
- Contact
22
- **
23
- : Please contact via Hugging Face or GitHub profile (delight2004)
24
- ##
25
- Model Overview
26
- -
27
-
28
- **
29
- Type
30
- **
31
- : Causal Language Model (LM)
32
- -
33
-
34
- **
35
- Base Model
36
- **
37
- : LiquidAI/LFM2-1.2B
38
- -
39
-
40
- **
41
- Fine-tuning technique
42
- **
43
- : QLoRA with PEFT (LoRA adapters)
44
- -
45
-
46
- **
47
- Language(s)
48
- **
49
- : English only
50
- -
51
-
52
- **
53
- Intended Use
54
- **
55
- : Research, educational, and sermon content generation on Christian and theological topics (especially inspired by John Piper's teachings).
56
- -
57
-
58
- **
59
- Tags
60
- **
61
- : Uganda, theology, Christianity
62
- -
63
-
64
- **
65
- License
66
- **
67
- : OpenRAIL Non-Commercial Variant
68
- ##
69
- Dataset & Training
70
- -
71
-
72
- **
73
- Data source
74
- **
75
- : Transcripts of YouTube sermons by John Piper (excluding "Ask Pastor John" podcast transcripts)
76
- -
77
-
78
- **
79
- Filtered dataset size
80
- **
81
- : 165 entries after filtering (~10% set aside for validation)
82
- -
83
-
84
- **
85
- Preprocessing
86
- **
87
- : Splitting and curation as
88
- [
89
- detailed in the training notebook
90
- ]
91
- (
92
- https://colab.research.google.com/drive/1pU7TQfneS48j8XxWjM1DPesUao5GEtIs#scrollTo=Wj07AZOhCnxl
93
- )
94
- -
95
-
96
- **
97
- Training details
98
- **
99
- :
100
-
101
- -
102
- Hardware: Google Colab free tier (T4 GPU)
103
-
104
- -
105
- epochs: 4
106
-
107
- -
108
- batch size: 1 (gradient_accumulation_steps=4)
109
-
110
- -
111
- learning rate: 2e-5
112
 
113
- ## How to use
 
114
 
115
- ```python
116
- from transformers import pipeline
 
 
 
 
 
 
117
 
118
- # Anyone can now load your model by its name!
119
- pipe = pipeline("text-generation", model="delight2004/lfm2-1.2b-sermon-instruct-qlora")
 
 
 
 
 
 
 
 
 
 
 
120
 
121
- result = pipe("Preach a short sermon in the style of John Piper on the topic of: 'Grace'")
122
- print(result[0]['generated_text'])
123
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  base_model: LiquidAI/LFM2-1.2B
11
  pipeline_tag: text-generation
12
  ---
13
+ # lfm2-1.2b-sermon-instruct-qlora
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
+ **Author**: Delight Aheebwa
16
+ **Contact**: Please contact via Hugging Face or GitHub profile (delight2004)
17
 
18
+ ## Model Overview
19
+ - **Type**: Causal Language Model (LM)
20
+ - **Base Model**: LiquidAI/LFM2-1.2B
21
+ - **Fine-tuning technique**: QLoRA with PEFT (LoRA adapters)
22
+ - **Language(s)**: English only
23
+ - **Intended Use**: Research, educational, and sermon content generation on Christian and theological topics (especially inspired by John Piper's teachings).
24
+ - **Tags**: Uganda, theology, Christianity
25
+ - **License**: OpenRAIL Non-Commercial Variant
26
 
27
+ ## Dataset & Training
28
+ - **Data source**: Transcripts of YouTube sermons by John Piper (excluding "Ask Pastor John" podcast transcripts)
29
+ - **Filtered dataset size**: 165 entries after filtering (~10% set aside for validation)
30
+ - **Preprocessing**: Splitting and curation as [detailed in the training notebook](https://colab.research.google.com/drive/1pU7TQfneS48j8XxWjM1DPesUao5GEtIs#scrollTo=Wj07AZOhCnxl)
31
+ - **Training details**:
32
+ - Hardware: Google Colab free tier (T4 GPU)
33
+ - epochs: 4
34
+ - batch size: 1 (gradient_accumulation_steps=4)
35
+ - learning rate: 2e-5
36
+ - sequence length: 512
37
+ - quantization: 4-bit (bnb_4bit_quant_type="nf4", bnb_4bit_compute_dtype=torch.bfloat16)
38
+ - Only LoRA adapter params were trained (~0.05% of total)
39
+ - Full trainer/config code: see Colab notebook above
40
 
41
+ ## Evaluation
42
+ - No formal evaluation/benchmarking was conducted. Use at your own discretion – feedback and community tests are welcome.
43
+
44
+ ## Limitations & Disclaimer
45
+ - Not intended for production or commercial use.
46
+ - Outputs should not be treated as official theological advice.
47
+ - Possible biases and limitations inherited from the dataset/model base – may reflect the original preacher's views.
48
+ - Model may hallucinate or generate plausible but incorrect theological claims or references.
49
+
50
+ ## Technical
51
+ - **Architecture**: Causal Transformer (1.2B params, LiquidAI flavor)
52
+ - **Adapter config**: PEFT/QLoRA
53
+ - **Training framework**: Hugging Face Transformers, TRL, PEFT, bitsandbytes, PyTorch
54
+ - **Compute**: Google Colab T4 (free tier, single GPU)
55
+ - **Notebook**: [john_piper.ipynb](https://colab.research.google.com/drive/1pU7TQfneS48j8XxWjM1DPesUao5GEtIs#scrollTo=Wj07AZOhCnxl)
56
+
57
+ ## Citation
58
+ If you use this model, please cite it or reference its Hugging Face page, and acknowledge John Piper's YouTube sermons as the data source.