Update README.md
Browse files
README.md
CHANGED
|
@@ -240,7 +240,7 @@ print(result)
|
|
| 240 |
### 4. Türkçe Sentiment Analizi
|
| 241 |
```python
|
| 242 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 243 |
-
"
|
| 244 |
)
|
| 245 |
|
| 246 |
texts = [
|
|
@@ -313,7 +313,7 @@ Torch: 1.9+
|
|
| 313 |
from transformers import AutoTokenizer
|
| 314 |
|
| 315 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 316 |
-
"
|
| 317 |
)
|
| 318 |
|
| 319 |
# Basit cümle
|
|
@@ -410,7 +410,7 @@ pip install --upgrade transformers
|
|
| 410 |
# Tokenizer'ı güncelle
|
| 411 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 412 |
"your-username/mini-turkish-tokenizer",
|
| 413 |
-
revision="main"
|
| 414 |
)
|
| 415 |
```
|
| 416 |
|
|
|
|
| 240 |
### 4. Türkçe Sentiment Analizi
|
| 241 |
```python
|
| 242 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 243 |
+
"kaanilker/mini-turkish-tokenizer"
|
| 244 |
)
|
| 245 |
|
| 246 |
texts = [
|
|
|
|
| 313 |
from transformers import AutoTokenizer
|
| 314 |
|
| 315 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 316 |
+
"kaanilker/mini-turkish-tokenizer"
|
| 317 |
)
|
| 318 |
|
| 319 |
# Basit cümle
|
|
|
|
| 410 |
# Tokenizer'ı güncelle
|
| 411 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 412 |
"your-username/mini-turkish-tokenizer",
|
| 413 |
+
revision="main"
|
| 414 |
)
|
| 415 |
```
|
| 416 |
|