File size: 1,455 Bytes
f6704e5
 
 
 
 
 
 
 
 
3e45e0d
767eba1
f6704e5
 
9b34925
48e6c6e
507fec3
f6704e5
120d3c4
443d439
95d3514
045ecee
 
0744ad4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ff2ca93
43ebb92
 
f6704e5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f1baad3
f6704e5
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
---
license: mit
language:
- en
tags:
- text-generation-inference
pipeline_tag: text-generation
---

![if-your-ai-girlfriend-is-not-a-locally-running-fine-tuned-v0-04wo67pdnuvf1](https://cdn-uploads.huggingface.co/production/uploads/64b7618e2f5a966b972e9978/8x-Ef3kxYJMfhwZ1CPHvp.png)

## GPT-Fem
An 81-million parameter LLM using GPT-2 encodings.
Trained using 16GB of text relating to and made by women, along with 1GB of multilingual text. (5.2 billion tokens)

This model should be fine-tuned before use.

![Screenshot from 2026-01-18 22-57-13](https://cdn-uploads.huggingface.co/production/uploads/64b7618e2f5a966b972e9978/50XOSomGKO0FQRtPEOjUO.png)

## Languages:
English,
Turkish,
Swedish,
Serbian,
Portugese,
Norwegian,
Welsh,
Thai,
Polish,
French,
Finnish,
Dutch,
Arabic,
Korean,
Japanese,
Danish,
Croatian,
Spanish,
Russian,
Chinese


## Technical Information
|                                 |     |
|---------------------------------|----:|
|Layers                           |10|
|Heads                            |10|
|Embeddings                       |640|
|Context Window                   |4096 tokens|
|Tokenizer                        |GPT-2 BPE|


## Training Information
|                                 |     |
|---------------------------------|----:|
|Training Loss                    |3.0|
|Validation Loss                  |2.99|
|Device                           |Google Colab A100|
|Training Time                    |5 Hours|