Tasfiya025 commited on
Commit
0ce0e40
·
verified ·
1 Parent(s): c33b596

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +81 -0
README.md ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - time-series-classification
4
+ - human-activity-recognition
5
+ - multimodal
6
+ - cnn-lstm
7
+ - sensor-data
8
+ datasets:
9
+ - MultiModal_HumanActivity_SensorStream
10
+ license: apache-2.0
11
+ model-index:
12
+ - name: HAR_MultiModal_Classifier
13
+ results:
14
+ - task:
15
+ name: Time Series Classification
16
+ type: time-series-classification
17
+ metrics:
18
+ - type: accuracy
19
+ value: 0.931
20
+ name: Sequence Accuracy
21
+ - type: weighted_f1
22
+ value: 0.925
23
+ name: Weighted F1 Score
24
+ ---
25
+
26
+ # HAR_MultiModal_Classifier
27
+
28
+ ## 🏃 Overview
29
+
30
+ The **HAR_MultiModal_Classifier** is a deep learning model designed for **Human Activity Recognition (HAR)**. It classifies complex human activities from raw time-series sensor streams, utilizing data from accelerometers (Acc\_X, Y, Z), gyroscopes (Gyro\_X, Y, Z), and contextual physiological metrics (Heart\_Rate\_BPM, Calories\_Burned\_kJ, Device\_Location) simultaneously.
31
+
32
+ ## 🧠 Model Architecture
33
+
34
+ The architecture is a specialized **Convolutional Neural Network (CNN) combined with a Long Short-Term Memory (LSTM)** network, optimized for processing sequential, high-frequency sensor data.
35
+
36
+ * **Input:** Sequences of 50 timesteps, containing 9 features per step (6 sensor, 3 contextual/physiological).
37
+ * **CNN Layer:** Extracts spatial features and localized patterns from the sensor data windows.
38
+ * **LSTM Layer:** Captures the temporal dependencies and long-range sequential dynamics inherent in human motion (e.g., the cyclical pattern of "Walking").
39
+ * **Classification Head:** A dense layer with Softmax activation outputs the probability distribution over the 6 activity classes.
40
+ * **Target Classes:** Walking, Sitting, Running, Lifting\_Heavy, Typing, Climbing\_Stairs.
41
+
42
+ ## 🎯 Intended Use
43
+
44
+ This model is ideal for applications requiring continuous, precise activity monitoring:
45
+
46
+ 1. **Smart Wearable Devices:** Powering real-time activity tracking and fitness coaching.
47
+ 2. **Health Monitoring:** Detecting falls, anomalous activity, or adherence to prescribed exercise routines.
48
+ 3. **Contextual Computing:** Providing accurate context for mobile applications and ambient intelligence systems.
49
+ 4. **Robotics and Automation:** Training robots to understand human motion and collaboration.
50
+
51
+ ## ⚠️ Limitations
52
+
53
+ 1. **Device Dependence:** Performance is highly dependent on sensor quality, sampling rate, and device placement (Wrist, Chest, Back, etc.). Deviations from the `Device_Location` in the training set may reduce accuracy.
54
+ 2. **Activity Overlap:** The model may confuse activities with similar movement signatures (e.g., fast walking vs. slow jogging), despite multimodal input.
55
+ 3. **Subject Variance:** The model's accuracy may vary across new subjects due to differences in gait, body mass, and movement style, necessitating fine-tuning for personalized deployment.
56
+
57
+ ---
58
+
59
+ ### MODEL 2: **AspectScorer_ReviewBERT**
60
+
61
+ This model is a multi-output regression model based on BERT, trained to predict multiple numerical aspect scores from a single raw text review.
62
+
63
+ #### config.json
64
+
65
+ ```json
66
+ {
67
+ "_name_or_path": "bert-base-uncased",
68
+ "architectures": [
69
+ "BertForMultipleRegression"
70
+ ],
71
+ "hidden_size": 768,
72
+ "model_type": "bert",
73
+ "num_hidden_layers": 12,
74
+ "vocab_size": 30522,
75
+ "problem_type": "multi_output_regression",
76
+ "num_labels": 3,
77
+ "output_labels": ["Aspect_Performance", "Aspect_Price_Value", "Aspect_Aesthetics"],
78
+ "min_rating": 1.0,
79
+ "max_rating": 5.0,
80
+ "transformers_version": "4.35.2"
81
+ }