File size: 4,542 Bytes
297f800 f219945 dea5a71 7f8afa9 f219945 dea5a71 297f800 dea5a71 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 297f800 f219945 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 | ---
library_name: transformers
tags:
- finance
- crypto
- text-classification
- bert
- turkish
- BTC
- ETH
- XRP
license: mit
base_model:
- dbmdz/bert-base-turkish-cased
pipeline_tag: text-classification
language:
- tr
---
# SkyWalkertT1/crypto_bert_sentiment
## 📌 Model Details
### Model Description
This is a BERT-based sentiment classification model fine-tuned on Turkish-language cryptocurrency-related comments. It predicts one of three sentiment classes: positive, neutral, or negative. This model was built using the Hugging Face 🤗 Transformers library and is suitable for analyzing sentiment in crypto communities, forums, or financial social media texts in Turkish.
- **Developed by:** [SkyWalkertT1 - Furkan Fatih Çiftçi]
- **Funded by:** Personal / Community Open Source
- **Shared by:** SkyWalkertT1
- **Model type:** BERT-based Sequence Classification
- **Language(s) (NLP):** Turkish
- **License:** Apache 2.0
- **Finetuned from model:** `dbmdz/bert-base-turkish-cased`
## 📚 Training Details
### Training Data
Dataset consists of labeled Turkish-language comments related to cryptocurrency, manually tagged with 3 sentiment labels.
**The dataset used for training this model is proprietary and was created and labeled by the author.**
The dataset shape is approximately `(1171, 2)` — indicating 1171 samples with 2 columns (text and label).
### Model Sources
- **Repository:** https://huggingface.co/SkyWalkertT1/my_crypto_comment_model
## 🔍 Uses
### Direct Use
- Turkish sentiment analysis on crypto/financial text
- Educational / experimental use for NLP in Turkish
### Downstream Use
- Integration into crypto sentiment bots
- Turkish language feedback systems
- Sentiment dashboards for crypto forums
### Out-of-Scope Use
- Use on non-Turkish text
- Medical, legal, or other high-risk domain sentiment prediction
## ⚠️ Bias, Risks, and Limitations
The model was trained on data specific to cryptocurrency sentiment in Turkish. It may not generalize to other domains. Model performance may vary depending on the writing style and slang usage.
### Recommendations
- Do not use this model for critical decision-making.
- Human validation should accompany any automated output.
## 🚀 How to Get Started with the Model
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model_path = "SkyWalkertT1/my_crypto_comment_model"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForSequenceClassification.from_pretrained(model_path)
text = "Bugün piyasada büyük bir düşüş bekliyorum."
inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True)
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
inputs = {k: v.to(device) for k, v in inputs.items()}
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits
predicted_class = torch.argmax(logits, dim=1).item()
labels = ['negative', 'neutral', 'positive']
print(f"Prediction: {labels[predicted_class]}")
```
## 📚 Training Details
### Training Data
Dataset consists of labeled Turkish-language comments related to cryptocurrency, manually tagged with 3 sentiment labels.
### Training Procedure
Model was fine-tuned using Hugging Face's `Trainer` API.
#### Training Hyperparameters
- Epochs: 4
- Batch size: 16
- Optimizer: AdamW
- Learning rate: 2e-5
- Precision: fp32
## 📈 Evaluation
### Testing Data, Factors & Metrics
Model evaluated on a 20% validation split from the same dataset.
#### Metrics
- Accuracy
- F1-score (macro average)
### Results
- Accuracy: ~85%
- F1-macro: ~84%
## 🌍 Environmental Impact
Carbon emissions are minimal due to fine-tuning only (~4 hours on a single NVIDIA T4 GPU).
- **Hardware Type:** NVIDIA T4 (Google Colab)
- **Hours used:** ~4
- **Cloud Provider:** Google Colab
- **Carbon Emitted:** Approx. ~1 kg CO2eq
## 🧠 Technical Specifications
### Model Architecture and Objective
BERT transformer architecture with a classification head on top for sequence classification into 3 sentiment classes.
### Compute Infrastructure
- Google Colab
- PyTorch + Transformers
## 📣 Citation
**BibTeX:**
```bibtex
@misc{SkyWalkertT1_crypto_bert,
author = {Furkan Fatih Çiftçi},
title = {Turkish Crypto Sentiment Model},
year = {2025.08.03},
howpublished = {\url{https://huggingface.co/SkyWalkertT1/my_crypto_comment_model}},
}
```
## 📬 Contact
For feedback or collaboration:
- Email: furkan-fatih-ciftci@hotmail.com |