Update file tokenizer.py
Browse files- tokenizer.py +2 -2
tokenizer.py
CHANGED
|
@@ -163,7 +163,7 @@ class Tokenizer:
|
|
| 163 |
subprocess.run(['./a.out', c_data, c_vocab], cwd=c_dir)
|
| 164 |
|
| 165 |
|
| 166 |
-
def c_encode(self, text):
|
| 167 |
script_dir = os.path.dirname(__file__)
|
| 168 |
c_dir = os.path.join(script_dir, 'c_tokenizer/')
|
| 169 |
|
|
@@ -174,7 +174,7 @@ class Tokenizer:
|
|
| 174 |
f.write(text)
|
| 175 |
|
| 176 |
|
| 177 |
-
self.to_file(
|
| 178 |
self.c_compile(c_dir)
|
| 179 |
self.c_run(c_dir, c_data, c_vocab)
|
| 180 |
|
|
|
|
| 163 |
subprocess.run(['./a.out', c_data, c_vocab], cwd=c_dir)
|
| 164 |
|
| 165 |
|
| 166 |
+
def c_encode(self, text):
|
| 167 |
script_dir = os.path.dirname(__file__)
|
| 168 |
c_dir = os.path.join(script_dir, 'c_tokenizer/')
|
| 169 |
|
|
|
|
| 174 |
f.write(text)
|
| 175 |
|
| 176 |
|
| 177 |
+
self.to_file(c_vocab)
|
| 178 |
self.c_compile(c_dir)
|
| 179 |
self.c_run(c_dir, c_data, c_vocab)
|
| 180 |
|