arbyazra123
commited on
Commit
·
24390d1
0
Parent(s):
init
Browse files- .DS_Store +0 -0
- .gitattributes +61 -0
- dataset/.DS_Store +0 -0
- dataset/game_logs_bt_vs_fsm.csv +3 -0
- dataset/game_logs_bt_vs_fsm_2.csv +3 -0
- dataset/game_logs_fsm_vs_bt.csv +3 -0
- dataset/game_logs_fsm_vs_fsm.csv +3 -0
- dataset/game_logs_mcts_vs_bt.csv +3 -0
- dataset/game_logs_mcts_vs_mcts.csv +3 -0
- dataset/game_logs_mcts_vs_primitive.csv +3 -0
- dataset/game_logs_primitive_vs_ga.csv +3 -0
- model/.DS_Store +0 -0
- model/ml.onnx +3 -0
- requirements.txt +10 -0
- train_ml.ipynb +205 -0
.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
.gitattributes
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.csv filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
|
| 3 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.lz4 filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.mds filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
# Audio files - uncompressed
|
| 41 |
+
*.pcm filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
*.sam filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
*.raw filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
# Audio files - compressed
|
| 45 |
+
*.aac filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
*.flac filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
*.mp3 filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
*.ogg filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
*.wav filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
# Image files - uncompressed
|
| 51 |
+
*.bmp filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
*.gif filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
*.png filter=lfs diff=lfs merge=lfs -text
|
| 54 |
+
*.tiff filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
# Image files - compressed
|
| 56 |
+
*.jpg filter=lfs diff=lfs merge=lfs -text
|
| 57 |
+
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
| 58 |
+
*.webp filter=lfs diff=lfs merge=lfs -text
|
| 59 |
+
# Video files - compressed
|
| 60 |
+
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
*.webm filter=lfs diff=lfs merge=lfs -text
|
dataset/.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
dataset/game_logs_bt_vs_fsm.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:02dc38efddcc83eddb7b369c990ed6fbecec3f3d219ddb744448356e26d75d90
|
| 3 |
+
size 130938719
|
dataset/game_logs_bt_vs_fsm_2.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8706953ef9f170c6cd5de1097927725fa92ee4c84862c029f1997dae3f7bf63e
|
| 3 |
+
size 26450015
|
dataset/game_logs_fsm_vs_bt.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a14eb7533b41b19b86379e4cfd74c715f2ff799b6a402169896311c75e79e180
|
| 3 |
+
size 47885979
|
dataset/game_logs_fsm_vs_fsm.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7558bc58030ced19da8c9c07ca7e91f53d83cd63ed16325eac2cbfdaec930559
|
| 3 |
+
size 397518069
|
dataset/game_logs_mcts_vs_bt.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:86401881fd5de369921c19094d1ad72b14a7f8cc6d0760225dfb2b062c61a820
|
| 3 |
+
size 12009518
|
dataset/game_logs_mcts_vs_mcts.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c32880015f11269cc1b89c4ae41c49852da615056e8baea56c0010be6d653a2
|
| 3 |
+
size 15892108
|
dataset/game_logs_mcts_vs_primitive.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f99ec3f450701af08f4c65221310a2575ea11f915dcd17a2afb3d948a39258d9
|
| 3 |
+
size 60463183
|
dataset/game_logs_primitive_vs_ga.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:96d270439aa504c06ee4991beb69bd54617df075ce78e237561f05d8bfb5eeff
|
| 3 |
+
size 39322613
|
model/.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
model/ml.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a86e7289da2b3236eab767a00dec412defb76e9ca0c2c64284ae14ab59c51be7
|
| 3 |
+
size 185930
|
requirements.txt
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
joblib==1.5.1
|
| 2 |
+
keras==2.13.1
|
| 3 |
+
matplotlib==3.10.5
|
| 4 |
+
numpy==1.24.3
|
| 5 |
+
onnx==1.17.0
|
| 6 |
+
pandas==2.3.1
|
| 7 |
+
scikit_learn==1.7.1
|
| 8 |
+
tensorflow==2.13.0
|
| 9 |
+
tensorflow_macos==2.13.0
|
| 10 |
+
tf2onnx==1.16.1
|
train_ml.ipynb
ADDED
|
@@ -0,0 +1,205 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cells": [
|
| 3 |
+
{
|
| 4 |
+
"cell_type": "code",
|
| 5 |
+
"execution_count": null,
|
| 6 |
+
"id": "8bb0209c-63c4-4601-894c-0ded8f4db2e6",
|
| 7 |
+
"metadata": {},
|
| 8 |
+
"outputs": [],
|
| 9 |
+
"source": [
|
| 10 |
+
"# PREPARE DATASET\n",
|
| 11 |
+
"\n",
|
| 12 |
+
"import numpy as np\n",
|
| 13 |
+
"import pandas as pd\n",
|
| 14 |
+
"import glob\n",
|
| 15 |
+
"\n",
|
| 16 |
+
"def filter_inside_arena(df, radius=4.73485, margin=0.95, apply=True):\n",
|
| 17 |
+
" if not apply:\n",
|
| 18 |
+
" return df.copy()\n",
|
| 19 |
+
" \n",
|
| 20 |
+
" bot_dist = np.sqrt(df[\"BotPosX\"]**2 + df[\"BotPosY\"]**2)\n",
|
| 21 |
+
" df[\"IsOutOfArena\"] = bot_dist > (radius * margin)\n",
|
| 22 |
+
" return df[~df[\"IsOutOfArena\"]].copy()\n",
|
| 23 |
+
"\n",
|
| 24 |
+
"\n",
|
| 25 |
+
"# Gather all battle logs from multiple CSV files\n",
|
| 26 |
+
"csv_files = glob.glob(\"dataset/game_logs_*.csv\")\n",
|
| 27 |
+
"\n",
|
| 28 |
+
"dfs = []\n",
|
| 29 |
+
"for file in csv_files:\n",
|
| 30 |
+
" df = pd.read_csv(file)\n",
|
| 31 |
+
" df = df.dropna(subset=[\"Name\", \"Duration\"])\n",
|
| 32 |
+
"\n",
|
| 33 |
+
" # Optional: Filter for winner actions only\n",
|
| 34 |
+
" # df = df[df[\"Actor\"] == \"RoundWinner\"]\n",
|
| 35 |
+
"\n",
|
| 36 |
+
" dfs.append(df)\n",
|
| 37 |
+
"\n",
|
| 38 |
+
"df_combined = pd.concat(dfs, ignore_index=True).drop_duplicates()\n",
|
| 39 |
+
"\n",
|
| 40 |
+
"# pply arena filter\n",
|
| 41 |
+
"df_safe = filter_inside_arena(df_combined, apply=True)\n",
|
| 42 |
+
"\n",
|
| 43 |
+
"# Save result\n",
|
| 44 |
+
"df_safe.to_csv(\"dataset/cleaned_log.csv\", index=False)\n"
|
| 45 |
+
]
|
| 46 |
+
},
|
| 47 |
+
{
|
| 48 |
+
"cell_type": "code",
|
| 49 |
+
"execution_count": null,
|
| 50 |
+
"id": "817107a1-86f4-49de-a366-f1e80536ecef",
|
| 51 |
+
"metadata": {},
|
| 52 |
+
"outputs": [],
|
| 53 |
+
"source": [
|
| 54 |
+
"# TRAIN SETUP\n",
|
| 55 |
+
"\n",
|
| 56 |
+
"import json\n",
|
| 57 |
+
"import pandas as pd\n",
|
| 58 |
+
"import numpy as np\n",
|
| 59 |
+
"import tf2onnx\n",
|
| 60 |
+
"\n",
|
| 61 |
+
"from sklearn.preprocessing import LabelEncoder\n",
|
| 62 |
+
"from sklearn.model_selection import train_test_split\n",
|
| 63 |
+
"from sklearn.impute import SimpleImputer\n",
|
| 64 |
+
"from sklearn.metrics import classification_report\n",
|
| 65 |
+
"\n",
|
| 66 |
+
"import tensorflow as tf\n",
|
| 67 |
+
"from tensorflow.keras.models import Model\n",
|
| 68 |
+
"from tensorflow.keras.layers import Input, Dense, BatchNormalization\n",
|
| 69 |
+
"from tensorflow.keras.utils import to_categorical\n",
|
| 70 |
+
"from tensorflow.keras.optimizers import Adam\n",
|
| 71 |
+
"from tensorflow.keras.callbacks import EarlyStopping\n",
|
| 72 |
+
"from tensorflow.keras.losses import CategoricalCrossentropy\n",
|
| 73 |
+
"\n",
|
| 74 |
+
"\n",
|
| 75 |
+
"# Load & proses data\n",
|
| 76 |
+
"df = pd.read_csv(\"dataset/cleaned_log.csv\")\n",
|
| 77 |
+
"\n",
|
| 78 |
+
"features = [\n",
|
| 79 |
+
" \"BotPosX\", \n",
|
| 80 |
+
" \"BotPosY\", \n",
|
| 81 |
+
" \"BotRot\", \n",
|
| 82 |
+
" # \"BotLinv\",\n",
|
| 83 |
+
" # \"BotAngv\", \n",
|
| 84 |
+
" # \"BotIsDashActive\",\n",
|
| 85 |
+
" # \"BotIsSkillActive\", \n",
|
| 86 |
+
" # \"BotIsOutFromArena\",\n",
|
| 87 |
+
" # enemy\n",
|
| 88 |
+
" \"EnemyBotPosX\", \n",
|
| 89 |
+
" \"EnemyBotPosY\", \n",
|
| 90 |
+
" \"EnemyBotRot\",\n",
|
| 91 |
+
" # \"EnemyBotLinv\",\n",
|
| 92 |
+
" # \"EnemyBotAngv\", \n",
|
| 93 |
+
" # \"EnemyBotIsDashActive\",\n",
|
| 94 |
+
" # \"EnemyBotIsSkillActive\", \n",
|
| 95 |
+
" # \"EnemyBotIsOutFromArena\",\n",
|
| 96 |
+
"]\n",
|
| 97 |
+
"\n",
|
| 98 |
+
"X = df[features]\n",
|
| 99 |
+
"imputer = SimpleImputer(strategy=\"mean\")\n",
|
| 100 |
+
"X = imputer.fit_transform(X)\n",
|
| 101 |
+
"\n",
|
| 102 |
+
"# Encode label\n",
|
| 103 |
+
"le = LabelEncoder()\n",
|
| 104 |
+
"y_action = le.fit_transform(df[\"Name\"])\n",
|
| 105 |
+
"y_duration = df[\"Duration\"].values.astype(\"float32\")\n",
|
| 106 |
+
"\n",
|
| 107 |
+
"# One-hot encoding for action\n",
|
| 108 |
+
"y_action_cat = to_categorical(y_action)\n",
|
| 109 |
+
"\n",
|
| 110 |
+
"# Split\n",
|
| 111 |
+
"X_train, X_test, y_action_train, y_action_test, y_duration_train, y_duration_test, df_train, df_val = train_test_split(\n",
|
| 112 |
+
" X, y_action_cat, y_duration, df, test_size=0.2, random_state=42\n",
|
| 113 |
+
")\n",
|
| 114 |
+
"\n",
|
| 115 |
+
"# Build model\n",
|
| 116 |
+
"inputs = Input(shape=(X.shape[1], ))\n",
|
| 117 |
+
"x = Dense(256, activation='relu')(inputs)\n",
|
| 118 |
+
"x = BatchNormalization()(x)\n",
|
| 119 |
+
"x = Dense(128, activation='relu')(x)\n",
|
| 120 |
+
"x = Dense(64, activation='relu')(x)\n",
|
| 121 |
+
"x = Dense(32, activation='relu')(x)\n",
|
| 122 |
+
"\n",
|
| 123 |
+
"output_action = Dense(y_action_cat.shape[1], activation='softmax', name=\"action\")(x)\n",
|
| 124 |
+
"output_duration = Dense(1, activation='linear', name=\"duration\")(x)\n",
|
| 125 |
+
"\n",
|
| 126 |
+
"loss_action = CategoricalCrossentropy(label_smoothing=0.1)\n",
|
| 127 |
+
"\n",
|
| 128 |
+
"# Compile model\n",
|
| 129 |
+
"model = Model(inputs=inputs, outputs=[output_action, output_duration])\n",
|
| 130 |
+
"model.compile(\n",
|
| 131 |
+
" optimizer=Adam(learning_rate=0.0001),\n",
|
| 132 |
+
" loss={\"action\": loss_action, \"duration\": \"mae\"},\n",
|
| 133 |
+
" metrics={'action': 'accuracy', 'duration': 'mae'},\n",
|
| 134 |
+
" weighted_metrics={'action': 'accuracy', 'duration': 'mae'}\n",
|
| 135 |
+
")\n",
|
| 136 |
+
"\n",
|
| 137 |
+
"# Early stopping\n",
|
| 138 |
+
"early_stop = EarlyStopping(\n",
|
| 139 |
+
" monitor='val_loss',\n",
|
| 140 |
+
" patience=10,\n",
|
| 141 |
+
" min_delta=0.001,\n",
|
| 142 |
+
" mode='min',\n",
|
| 143 |
+
" restore_best_weights=True,\n",
|
| 144 |
+
" verbose=1\n",
|
| 145 |
+
")\n",
|
| 146 |
+
"\n",
|
| 147 |
+
"# Train\n",
|
| 148 |
+
"model.fit(X_train, {\"action\": y_action_train, \"duration\": y_duration_train},\n",
|
| 149 |
+
" validation_data=(X_test, {'action': y_action_test, 'duration': y_duration_test}),\n",
|
| 150 |
+
" epochs=100,\n",
|
| 151 |
+
" batch_size=512,\n",
|
| 152 |
+
" callbacks=[early_stop],\n",
|
| 153 |
+
" )\n",
|
| 154 |
+
"\n",
|
| 155 |
+
"# Predict\n",
|
| 156 |
+
"pred_action_prob, pred_duration = model.predict(X_test)\n",
|
| 157 |
+
"pred_action = np.argmax(pred_action_prob, axis=1)\n",
|
| 158 |
+
"true_action = np.argmax(y_action_test, axis=1)\n",
|
| 159 |
+
"\n",
|
| 160 |
+
"# Evaluation\n",
|
| 161 |
+
"print(classification_report(true_action, pred_action, target_names=le.classes_))\n",
|
| 162 |
+
"\n",
|
| 163 |
+
"# Convert the model\n",
|
| 164 |
+
"spec = (tf.TensorSpec((None, X.shape[1]), tf.float32, name=\"input\"),)\n",
|
| 165 |
+
"onnx_model, _ = tf2onnx.convert.from_keras(model, input_signature=spec, opset=13)\n",
|
| 166 |
+
"\n",
|
| 167 |
+
"# Save to file\n",
|
| 168 |
+
"with open(\"model/ml.onnx\", \"wb\") as f:\n",
|
| 169 |
+
" f.write(onnx_model.SerializeToString())\n",
|
| 170 |
+
"\n",
|
| 171 |
+
"print(\"Model saved to model/ml.onnx\")\n",
|
| 172 |
+
"\n",
|
| 173 |
+
"class_labels = le.classes_.tolist()\n",
|
| 174 |
+
"\n",
|
| 175 |
+
"# Optional: Save labels to JSON\n",
|
| 176 |
+
"with open(\"model/action_labels.json\", \"w\") as f:\n",
|
| 177 |
+
" json.dump(class_labels, f)\n",
|
| 178 |
+
"\n",
|
| 179 |
+
"print(\"Exported label encoder classes to action_labels.json:\")\n",
|
| 180 |
+
"print(class_labels)"
|
| 181 |
+
]
|
| 182 |
+
}
|
| 183 |
+
],
|
| 184 |
+
"metadata": {
|
| 185 |
+
"kernelspec": {
|
| 186 |
+
"display_name": "Python 3 (ipykernel)",
|
| 187 |
+
"language": "python",
|
| 188 |
+
"name": "python3"
|
| 189 |
+
},
|
| 190 |
+
"language_info": {
|
| 191 |
+
"codemirror_mode": {
|
| 192 |
+
"name": "ipython",
|
| 193 |
+
"version": 3
|
| 194 |
+
},
|
| 195 |
+
"file_extension": ".py",
|
| 196 |
+
"mimetype": "text/x-python",
|
| 197 |
+
"name": "python",
|
| 198 |
+
"nbconvert_exporter": "python",
|
| 199 |
+
"pygments_lexer": "ipython3",
|
| 200 |
+
"version": "3.10.16"
|
| 201 |
+
}
|
| 202 |
+
},
|
| 203 |
+
"nbformat": 4,
|
| 204 |
+
"nbformat_minor": 5
|
| 205 |
+
}
|