legolasyiu commited on
Commit
7736f66
·
verified ·
1 Parent(s): 465a1f1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -0
README.md CHANGED
@@ -10,6 +10,61 @@ language:
10
  - en
11
  ---
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  # Uploaded finetuned model
14
 
15
  - **Developed by:** EpistemeAI
 
10
  - en
11
  ---
12
 
13
+ ## This is a metatune-gpt20b model used prototype for self-improving ai training loop.
14
+ - Coding
15
+ - Generates new data for itself,
16
+ - Evaluates its performance, and
17
+ - Adjusts its own hyperparameters based on improvement metrics.
18
+
19
+ ## Use cases:
20
+ - genuinely demonstrate scientific and mathematical understanding at a postdoctoral level.
21
+ - - Topics: Euler–Lagrange equation, vector calculus, statistical mechanics
22
+
23
+ # Inference examples
24
+
25
+ ## Transformers
26
+
27
+ You can use `gpt-oss-120b` and `gpt-oss-20b` with Transformers. If you use the Transformers chat template, it will automatically apply the [harmony response format](https://github.com/openai/harmony). If you use `model.generate` directly, you need to apply the harmony format manually using the chat template or use our [openai-harmony](https://github.com/openai/harmony) package.
28
+
29
+ To get started, install the necessary dependencies to setup your environment:
30
+
31
+ ```
32
+ pip install -U transformers kernels torch
33
+ ```
34
+
35
+ For Google Colab (free/Pro)
36
+ ```
37
+ !pip install -q --upgrade torch
38
+
39
+ !pip install -q transformers triton==3.4 kernels
40
+
41
+ !pip uninstall -q torchvision torchaudio -y
42
+ ```
43
+
44
+ Once, setup you can proceed to run the model by running the snippet below:
45
+
46
+ ```py
47
+ from transformers import pipeline
48
+ import torch
49
+ model_id = "EpistemeAI/metatune-gpt20b"
50
+ pipe = pipeline(
51
+ "text-generation",
52
+ model=model_id,
53
+ torch_dtype="auto",
54
+ device_map="auto",
55
+ )
56
+ messages = [
57
+ {"role": "user", "content": "Derive the Euler–Lagrange equation from the principle of stationary action.""},
58
+ ]
59
+ outputs = pipe(
60
+ messages,
61
+ max_new_tokens=3000,
62
+ )
63
+ print(outputs[0]["generated_text"][-1])
64
+ ```
65
+
66
+ ## Benchmark[TBD]
67
+
68
  # Uploaded finetuned model
69
 
70
  - **Developed by:** EpistemeAI