| | --- |
| | title: Action Detection In Images |
| | emoji: 😻 |
| | colorFrom: blue |
| | colorTo: pink |
| | sdk: gradio |
| | sdk_version: 4.31.5 |
| | app_file: app.py |
| | pinned: false |
| | license: mit |
| | --- |
| | |
| | Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference |
| |
|
| | # General Action Classifier |
| |
|
| | This is a Gradio interface that allows users to upload an image and specify candidate labels to check if a certain action is present in the image. The app uses a CLIP-ViT model to classify the image based on the provided labels. |
| |
|
| | ## How to Use |
| |
|
| | 1. Upload an image. |
| | 2. Enter candidate labels separated by commas. |
| | 3. The app will classify the image and display the results. |
| |
|
| | ## Example |
| |
|
| | For instance, to check if a person is holding a beverage, you can enter labels like: "human with beverage, human, beverage". |
| |
|