Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
KevinKibe
/
working
like
0
TensorBoard
Safetensors
whisper
Model card
Files
Files and versions
xet
Metrics
Training metrics
Community
main
working
/
checkpoint-50-epoch-3
4.42 GB
Ctrl+K
Ctrl+K
1 contributor
History:
4 commits
KevinKibe
Saving train state of step 50
45cd265
verified
over 1 year ago
added_tokens.json
Safe
34.6 kB
Saving train state of step 50
over 1 year ago
config.json
Safe
2.27 kB
Saving train state of step 50
over 1 year ago
generation_config.json
Safe
3.65 kB
Saving train state of step 50
over 1 year ago
merges.txt
Safe
494 kB
Saving train state of step 50
over 1 year ago
model.safetensors
1.58 GB
xet
Saving train state of step 50
over 1 year ago
model_1.safetensors
2.14 GB
xet
Saving train state of step 50
over 1 year ago
normalizer.json
Safe
52.7 kB
Saving train state of step 50
over 1 year ago
optimizer.bin
pickle
Detected Pickle imports (3)
"torch.FloatStorage"
,
"torch._utils._rebuild_tensor_v2"
,
"collections.OrderedDict"
What is a pickle import?
694 MB
xet
Saving train state of step 50
over 1 year ago
preprocessor_config.json
Safe
339 Bytes
Saving train state of step 50
over 1 year ago
random_states_0.pkl
pickle
Detected Pickle imports (7)
"torch.ByteStorage"
,
"collections.OrderedDict"
,
"torch._utils._rebuild_tensor_v2"
,
"numpy.dtype"
,
"_codecs.encode"
,
"numpy.ndarray"
,
"numpy.core.multiarray._reconstruct"
How to fix it?
14.6 kB
xet
Saving train state of step 50
over 1 year ago
random_states_1.pkl
pickle
Detected Pickle imports (7)
"torch.ByteStorage"
,
"collections.OrderedDict"
,
"torch._utils._rebuild_tensor_v2"
,
"numpy.dtype"
,
"_codecs.encode"
,
"numpy.ndarray"
,
"numpy.core.multiarray._reconstruct"
How to fix it?
14.6 kB
xet
Saving train state of step 50
over 1 year ago
scheduler.bin
Safe
pickle
Pickle imports
No problematic imports detected
What is a pickle import?
1.06 kB
xet
Saving train state of step 50
over 1 year ago
special_tokens_map.json
Safe
2.19 kB
Saving train state of step 50
over 1 year ago
tokenizer.json
Safe
2.48 MB
Saving train state of step 50
over 1 year ago
tokenizer_config.json
Safe
283 kB
Saving train state of step 50
over 1 year ago
vocab.json
Safe
836 kB
Saving train state of step 50
over 1 year ago