vedaco commited on
Commit
d1ef24d
·
verified ·
1 Parent(s): 42ab624

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -9
README.md CHANGED
@@ -1,16 +1,50 @@
1
  ---
2
- title: Veda Programming
3
- emoji: 💬
4
- colorFrom: yellow
5
- colorTo: purple
6
  sdk: gradio
7
- sdk_version: 5.42.0
8
  app_file: app.py
9
  pinned: false
10
- hf_oauth: true
11
- hf_oauth_scopes:
12
- - inference-api
13
  license: mit
14
  ---
15
 
16
- An example chatbot using [Gradio](https://gradio.app), [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub/v0.22.2/en/index), and the [Hugging Face Inference API](https://huggingface.co/docs/api-inference/index).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: Veda Programming LLM
3
+ emoji: 🕉️
4
+ colorFrom: purple
5
+ colorTo: blue
6
  sdk: gradio
7
+ sdk_version: 3.50.0
8
  app_file: app.py
9
  pinned: false
 
 
 
10
  license: mit
11
  ---
12
 
13
+ # 🕉️ Veda Programming LLM
14
+
15
+ A TensorFlow-based Large Language Model for programming code generation.
16
+
17
+ ## Features
18
+
19
+ - **Code Generation**: Generate Python code from prompts
20
+ - **Custom Training**: Train on your own code samples
21
+ - **Transformer Architecture**: Uses modern transformer blocks
22
+ - **Interactive Interface**: Easy-to-use Gradio interface
23
+
24
+ ## Model Architecture
25
+
26
+ - Transformer-based decoder architecture
27
+ - Configurable model sizes (small/medium/large)
28
+ - Causal attention masking for autoregressive generation
29
+ - Custom tokenizer optimized for code
30
+
31
+ ## Usage
32
+
33
+ 1. **Generate Code**: Enter a code prompt and adjust generation parameters
34
+ 2. **Train Model**: Paste your code samples and train the model
35
+ 3. **View Model Info**: Check model architecture and parameters
36
+
37
+ ## Parameters
38
+
39
+ - **Temperature**: Controls randomness (lower = more deterministic)
40
+ - **Top-K**: Limits sampling to top K tokens
41
+ - **Top-P**: Nucleus sampling threshold
42
+ - **Max Tokens**: Maximum number of tokens to generate
43
+
44
+ ## Training Data
45
+
46
+ The model can be trained on `programming.txt` containing Python code samples.
47
+
48
+ ## License
49
+
50
+ MIT License