Image-to-Text
Transformers
PyTorch
Safetensors
vision-encoder-decoder
image-text-to-text
image-captioning
Instructions to use bipin/image-caption-generator with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use bipin/image-caption-generator with Transformers:
# Use a pipeline as a high-level helper # Warning: Pipeline type "image-to-text" is no longer supported in transformers v5. # You must load the model directly (see below) or downgrade to v4.x with: # 'pip install "transformers<5.0.0' from transformers import pipeline pipe = pipeline("image-to-text", model="bipin/image-caption-generator")# Load model directly from transformers import AutoTokenizer, AutoModelForImageTextToText tokenizer = AutoTokenizer.from_pretrained("bipin/image-caption-generator") model = AutoModelForImageTextToText.from_pretrained("bipin/image-caption-generator") - Notebooks
- Google Colab
- Kaggle
Training in progress, step 500
Browse files- pytorch_model.bin +1 -1
- training_args.bin +1 -1
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 982141993
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fbbbe6b1ce6f441043e14a629b9cd517691e8e0d6afd531e16ec960ccabdf13
|
| 3 |
size 982141993
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3119
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:12d536302a5b33241213cb53564774bfeed1d79f2c40be57f2580edda4e52f1d
|
| 3 |
size 3119
|