Danny Yin commited on
Commit ·
db1bb53
1
Parent(s): 2d11f21
release
Browse files- LICENSE.md +35 -0
- README.md +132 -0
- config.json +289 -0
- model.safetensors +3 -0
- preprocessor_config.json +27 -0
LICENSE.md
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
NVIDIA License
|
| 2 |
+
|
| 3 |
+
1. Definitions
|
| 4 |
+
|
| 5 |
+
“Licensor” means any person or entity that distributes its Work.
|
| 6 |
+
“Work” means (a) the original work of authorship made available under this license, which may include software, documentation, or other files, and (b) any additions to or derivative works thereof that are made available under this license.
|
| 7 |
+
The terms “reproduce,” “reproduction,” “derivative works,” and “distribution” have the meaning as provided under U.S. copyright law; provided, however, that for the purposes of this license, derivative works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work.
|
| 8 |
+
Works are “made available” under this license by including in or with the Work either (a) a copyright notice referencing the applicability of this license to the Work, or (b) a copy of this license.
|
| 9 |
+
|
| 10 |
+
2. License Grant
|
| 11 |
+
|
| 12 |
+
2.1 Copyright Grant. Subject to the terms and conditions of this license, each Licensor grants to you a perpetual, worldwide, non-exclusive, royalty-free, copyright license to use, reproduce, prepare derivative works of, publicly display, publicly perform, sublicense and distribute its Work and any resulting derivative works in any form.
|
| 13 |
+
|
| 14 |
+
3. Limitations
|
| 15 |
+
|
| 16 |
+
3.1 Redistribution. You may reproduce or distribute the Work only if (a) you do so under this license, (b) you include a complete copy of this license with your distribution, and (c) you retain without modification any copyright, patent, trademark, or attribution notices that are present in the Work.
|
| 17 |
+
|
| 18 |
+
3.2 Derivative Works. You may specify that additional or different terms apply to the use, reproduction, and distribution of your derivative works of the Work (“Your Terms”) only if (a) Your Terms provide that the use limitation in Section 3.3 applies to your derivative works, and (b) you identify the specific derivative works that are subject to Your Terms. Notwithstanding Your Terms, this license (including the redistribution requirements in Section 3.1) will continue to apply to the Work itself.
|
| 19 |
+
|
| 20 |
+
3.3 Use Limitation. The Work and any derivative works thereof only may be used or intended for use non-commercially. Notwithstanding the foregoing, NVIDIA Corporation and its affiliates may use the Work and any derivative works commercially. As used herein, “non-commercially” means for non-commercial research activities or non-commercial research publications only.
|
| 21 |
+
|
| 22 |
+
3.4 Patent Claims. If you bring or threaten to bring a patent claim against any Licensor (including any claim, cross-claim or counterclaim in a lawsuit) to enforce any patents that you allege are infringed by any Work, then your rights under this license from such Licensor (including the grant in Section 2.1) will terminate immediately.
|
| 23 |
+
|
| 24 |
+
3.5 Trademarks. This license does not grant any rights to use any Licensor’s or its affiliates’ names, logos, or trademarks, except as necessary to reproduce the notices described in this license.
|
| 25 |
+
|
| 26 |
+
3.6 Termination. If you violate any term of this license, then your rights under this license (including the grant in Section 2.1) will terminate immediately.
|
| 27 |
+
|
| 28 |
+
4. Disclaimer of Warranty.
|
| 29 |
+
|
| 30 |
+
THE WORK IS PROVIDED “AS IS” WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING WARRANTIES OR CONDITIONS OF
|
| 31 |
+
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE OR NON-INFRINGEMENT. YOU BEAR THE RISK OF UNDERTAKING ANY ACTIVITIES UNDER THIS LICENSE.
|
| 32 |
+
|
| 33 |
+
5. Limitation of Liability.
|
| 34 |
+
|
| 35 |
+
EXCEPT AS PROHIBITED BY APPLICABLE LAW, IN NO EVENT AND UNDER NO LEGAL THEORY, WHETHER IN TORT (INCLUDING NEGLIGENCE), CONTRACT, OR OTHERWISE SHALL ANY LICENSOR BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES ARISING OUT OF OR RELATED TO THIS LICENSE, THE USE OR INABILITY TO USE THE WORK (INCLUDING BUT NOT LIMITED TO LOSS OF GOODWILL, BUSINESS INTERRUPTION, LOST PROFITS OR DATA, COMPUTER FAILURE OR MALFUNCTION, OR ANY OTHER DAMAGES OR LOSSES), EVEN IF THE LICENSOR HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
|
README.md
ADDED
|
@@ -0,0 +1,132 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
license_name: nvidia
|
| 4 |
+
license_link: LICENSE
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
## Model Overview
|
| 9 |
+
|
| 10 |
+
### Description:
|
| 11 |
+
|
| 12 |
+
AutoGaze is a ultra light-weight model that automatically removes redundant patches in a video before passing it to any Vision Transformer (ViT) or Multi-modal Large Language Model (MLLM).
|
| 13 |
+
|
| 14 |
+
Speficially, AutoGaze perceives each frame and autoregressively selects ("gazing") a minimal set of patches that can reconstruct the original video (i.e., non-redundant patches) up to a reconstruction loss threhold provided by the user. AutoGaze can self-decide when to stop gazing for each frame based on user's request on the acceptable maximum reconstruction loss.<br>
|
| 15 |
+
|
| 16 |
+
Empircally, AutoGaze can reduce #tokens in ViTs/MLLMs by up to 100x, reducing their latency by up to 19x/10x. This enables efficiently scaling MLLMs to 4K-resolution, 1K-frame videos, improving performance on benchmarks such as VideoMME. Especially, it improves performance by 14% on HLVid, a high-resolution long-form video benchmark proposed in this work as well.
|
| 17 |
+
|
| 18 |
+
This model is for research and development only. <br>
|
| 19 |
+
|
| 20 |
+
### Quick Start:
|
| 21 |
+
|
| 22 |
+
See [our GitHub repo](https://github.com/NVlabs/AutoGaze/QUICK_START.md) for instructions on how to use AutoGaze.
|
| 23 |
+
|
| 24 |
+
### License/Terms of Use:
|
| 25 |
+
|
| 26 |
+
NVIDIA license (see LICENSE.md). The reference to the NVIDIA License means the attached custom NSCLv1 license, under which users may use for purposes of conducting non-commercial research activities and non-commercial research publications.
|
| 27 |
+
|
| 28 |
+
### Deployment Geography:
|
| 29 |
+
Global <br>
|
| 30 |
+
|
| 31 |
+
### Use Case: <br>
|
| 32 |
+
The model is used to remove redundancy in videos and accelerate video encoders and MLLMs. <br>
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
## References(s):
|
| 36 |
+
GitHub: https://github.com/NVlabs/AutoGaze <br>
|
| 37 |
+
|
| 38 |
+
## Model Architecture:
|
| 39 |
+
**Architecture Type:** CNN and Transformer. <br>
|
| 40 |
+
|
| 41 |
+
**Network Architecture:** Custom Architecture. <br>
|
| 42 |
+
|
| 43 |
+
**Number of model parameters:** 3M <br>
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
## Input(s): <br>
|
| 47 |
+
**Input Type(s):** Video <br>
|
| 48 |
+
|
| 49 |
+
**Input Format(s):** Video: .mp4/.webm/.mov./etc. <br>
|
| 50 |
+
|
| 51 |
+
**Input Parameters:** Video: Three-Dimensional (3D)
|
| 52 |
+
|
| 53 |
+
**Other Properties Related to Input:** Video with any resolution or duration.
|
| 54 |
+
|
| 55 |
+
|
| 56 |
+
## Output(s)
|
| 57 |
+
|
| 58 |
+
**Output Type(s):** Integers (representing patch indices) <br>
|
| 59 |
+
|
| 60 |
+
**Output Format(s):** Integers <br>
|
| 61 |
+
|
| 62 |
+
**Output Parameters:** One-Dimensional (1D)
|
| 63 |
+
|
| 64 |
+
**Other Properties Related to Outupt:** N/A
|
| 65 |
+
|
| 66 |
+
Our AI models are designed and/or optimized to run on NVIDIA GPU-accelerated systems. By leveraging NVIDIA’s hardware (e.g. GPU cores) and software frameworks (e.g., CUDA libraries), the model achieves faster training and inference times compared to CPU-only solutions. <br>
|
| 67 |
+
|
| 68 |
+
## Software Integration:
|
| 69 |
+
**Runtime Engine(s):** N/A
|
| 70 |
+
|
| 71 |
+
**Supported Hardware Microarchitecture Compatibility:** <br>
|
| 72 |
+
NVIDIA Ampere <br>
|
| 73 |
+
NVIDIA Blackwell <br>
|
| 74 |
+
NVIDIA Jetson <br>
|
| 75 |
+
NVIDIA Hopper <br>
|
| 76 |
+
|
| 77 |
+
**Preferred/Supported Operating System(s):**
|
| 78 |
+
Linux <br>
|
| 79 |
+
Linux 4 Tegra <br>
|
| 80 |
+
QNX <br>
|
| 81 |
+
Windows <br>
|
| 82 |
+
|
| 83 |
+
The integration of foundation and fine-tuned models into AI systems requires additional testing using use-case-specific data to ensure safe and effective deployment. Following the V-model methodology, iterative testing and validation at both unit and system levels are essential to mitigate risks, meet technical and functional requirements, and ensure compliance with safety and ethical standards before deployment. <br>
|
| 84 |
+
|
| 85 |
+
## Model Version(s):
|
| 86 |
+
v1.0 - Initial release <br>
|
| 87 |
+
|
| 88 |
+
## Training, Testing, and Evaluation Datasets:
|
| 89 |
+
|
| 90 |
+
### Dataset Overview
|
| 91 |
+
** Total Size: 800K <br>
|
| 92 |
+
** Total Number of Datasets: 1 <br>
|
| 93 |
+
|
| 94 |
+
** Dataset partition: Training [97%], Testing [3%] <br>
|
| 95 |
+
** Time period for training data collection 2025/5-2025/8 <br>
|
| 96 |
+
** Time period for testing data collection 2025/5-2025/8 <br>
|
| 97 |
+
|
| 98 |
+
The data is constructed by collecting raw videos from existing video datasets and labeling gazing sequences for a subset of it using a greedy-search algorithm. <br>
|
| 99 |
+
|
| 100 |
+
## Public Datasets
|
| 101 |
+
The raw videos are collected from public dataset including Ego4D, 100DoH, InternVid, SA-1B, and IDL. <br>
|
| 102 |
+
|
| 103 |
+
## Training Dataset [The dataset the model was trained on]:
|
| 104 |
+
|
| 105 |
+
**Data Modality:** Video <br>
|
| 106 |
+
|
| 107 |
+
**Video Training Data Size:** 800K videos <br>
|
| 108 |
+
|
| 109 |
+
**Data Collection Method by dataset:** Automated <br>
|
| 110 |
+
|
| 111 |
+
**Labeling Method by dataset:** Automated <br>
|
| 112 |
+
|
| 113 |
+
**Properties:** 800K videos with 224*224 resolution and 16 frames each video. <br>
|
| 114 |
+
|
| 115 |
+
## Testing Dataset:
|
| 116 |
+
|
| 117 |
+
**Data Collection Method by dataset:** Automated <br>
|
| 118 |
+
|
| 119 |
+
**Labeling Method by dataset:** Automated <br>
|
| 120 |
+
|
| 121 |
+
**Properties:** 25K videos with 224*224 resolution and 16 frames each video. <br>
|
| 122 |
+
|
| 123 |
+
|
| 124 |
+
## Inference:
|
| 125 |
+
**Acceleration Engine:** N/A <br>
|
| 126 |
+
**Test Hardware:** A100 <br>
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
## Ethical Considerations:
|
| 130 |
+
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. Please report model quality, risk, security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/). <br>
|
| 131 |
+
|
| 132 |
+
Please make sure you have proper rights and permissions for all input image and video content; if image or video includes people, personal health information, or intellectual property, the image or video generated will not blur or maintain proportions of image subjects included. <br>
|
config.json
ADDED
|
@@ -0,0 +1,289 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"attn_mode": "sdpa",
|
| 3 |
+
"gaze_model_config": {
|
| 4 |
+
"connector_config": {
|
| 5 |
+
"_attn_implementation_autoset": false,
|
| 6 |
+
"_name_or_path": "",
|
| 7 |
+
"add_cross_attention": false,
|
| 8 |
+
"architectures": null,
|
| 9 |
+
"bad_words_ids": null,
|
| 10 |
+
"begin_suppress_tokens": null,
|
| 11 |
+
"bos_token_id": null,
|
| 12 |
+
"chunk_size_feed_forward": 0,
|
| 13 |
+
"cross_attention_hidden_size": null,
|
| 14 |
+
"decoder_start_token_id": null,
|
| 15 |
+
"diversity_penalty": 0.0,
|
| 16 |
+
"do_sample": false,
|
| 17 |
+
"early_stopping": false,
|
| 18 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 19 |
+
"eos_token_id": null,
|
| 20 |
+
"exponential_decay_length_penalty": null,
|
| 21 |
+
"finetuning_task": null,
|
| 22 |
+
"forced_bos_token_id": null,
|
| 23 |
+
"forced_eos_token_id": null,
|
| 24 |
+
"hidden_dim": 192,
|
| 25 |
+
"id2label": {
|
| 26 |
+
"0": "LABEL_0",
|
| 27 |
+
"1": "LABEL_1"
|
| 28 |
+
},
|
| 29 |
+
"is_decoder": false,
|
| 30 |
+
"is_encoder_decoder": false,
|
| 31 |
+
"label2id": {
|
| 32 |
+
"LABEL_0": 0,
|
| 33 |
+
"LABEL_1": 1
|
| 34 |
+
},
|
| 35 |
+
"length_penalty": 1.0,
|
| 36 |
+
"max_length": 20,
|
| 37 |
+
"min_length": 0,
|
| 38 |
+
"model_type": "",
|
| 39 |
+
"no_repeat_ngram_size": 0,
|
| 40 |
+
"num_beam_groups": 1,
|
| 41 |
+
"num_beams": 1,
|
| 42 |
+
"num_return_sequences": 1,
|
| 43 |
+
"num_tokens": 196,
|
| 44 |
+
"output_attentions": false,
|
| 45 |
+
"output_hidden_states": false,
|
| 46 |
+
"output_scores": false,
|
| 47 |
+
"pad_token_id": null,
|
| 48 |
+
"prefix": null,
|
| 49 |
+
"problem_type": null,
|
| 50 |
+
"pruned_heads": {},
|
| 51 |
+
"remove_invalid_values": false,
|
| 52 |
+
"repetition_penalty": 1.0,
|
| 53 |
+
"return_dict": true,
|
| 54 |
+
"return_dict_in_generate": false,
|
| 55 |
+
"sep_token_id": null,
|
| 56 |
+
"suppress_tokens": null,
|
| 57 |
+
"task_specific_params": null,
|
| 58 |
+
"temperature": 1.0,
|
| 59 |
+
"tf_legacy_loss": false,
|
| 60 |
+
"tie_encoder_decoder": false,
|
| 61 |
+
"tie_word_embeddings": true,
|
| 62 |
+
"tokenizer_class": null,
|
| 63 |
+
"top_k": 50,
|
| 64 |
+
"top_p": 1.0,
|
| 65 |
+
"torch_dtype": null,
|
| 66 |
+
"torchscript": false,
|
| 67 |
+
"typical_p": 1.0,
|
| 68 |
+
"use_bfloat16": false
|
| 69 |
+
},
|
| 70 |
+
"gaze_decoder_config": {
|
| 71 |
+
"_attn_implementation_autoset": false,
|
| 72 |
+
"_name_or_path": "",
|
| 73 |
+
"add_cross_attention": false,
|
| 74 |
+
"architectures": null,
|
| 75 |
+
"attention_bias": false,
|
| 76 |
+
"attention_dropout": 0.0,
|
| 77 |
+
"bad_words_ids": null,
|
| 78 |
+
"begin_suppress_tokens": null,
|
| 79 |
+
"bos_token_id": 12800000,
|
| 80 |
+
"chunk_size_feed_forward": 0,
|
| 81 |
+
"cross_attention_hidden_size": null,
|
| 82 |
+
"decoder_start_token_id": null,
|
| 83 |
+
"diversity_penalty": 0.0,
|
| 84 |
+
"do_sample": false,
|
| 85 |
+
"early_stopping": false,
|
| 86 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 87 |
+
"eos_token_id": 265,
|
| 88 |
+
"exponential_decay_length_penalty": null,
|
| 89 |
+
"finetuning_task": null,
|
| 90 |
+
"forced_bos_token_id": null,
|
| 91 |
+
"forced_eos_token_id": null,
|
| 92 |
+
"head_dim": 32,
|
| 93 |
+
"hidden_act": "silu",
|
| 94 |
+
"hidden_size": 192,
|
| 95 |
+
"id2label": {
|
| 96 |
+
"0": "LABEL_0",
|
| 97 |
+
"1": "LABEL_1"
|
| 98 |
+
},
|
| 99 |
+
"initializer_range": 0.02,
|
| 100 |
+
"intermediate_size": 384,
|
| 101 |
+
"is_decoder": false,
|
| 102 |
+
"is_encoder_decoder": false,
|
| 103 |
+
"label2id": {
|
| 104 |
+
"LABEL_0": 0,
|
| 105 |
+
"LABEL_1": 1
|
| 106 |
+
},
|
| 107 |
+
"length_penalty": 1.0,
|
| 108 |
+
"max_length": 20,
|
| 109 |
+
"max_position_embeddings": 131072,
|
| 110 |
+
"min_length": 0,
|
| 111 |
+
"mlp_bias": false,
|
| 112 |
+
"model_type": "llama",
|
| 113 |
+
"no_repeat_ngram_size": 0,
|
| 114 |
+
"num_attention_heads": 6,
|
| 115 |
+
"num_beam_groups": 1,
|
| 116 |
+
"num_beams": 1,
|
| 117 |
+
"num_hidden_layers": 4,
|
| 118 |
+
"num_key_value_heads": 6,
|
| 119 |
+
"num_multi_token_pred": 10,
|
| 120 |
+
"num_return_sequences": 1,
|
| 121 |
+
"output_attentions": false,
|
| 122 |
+
"output_hidden_states": false,
|
| 123 |
+
"output_scores": false,
|
| 124 |
+
"pad_token_id": null,
|
| 125 |
+
"prefix": null,
|
| 126 |
+
"pretraining_tp": 1,
|
| 127 |
+
"problem_type": null,
|
| 128 |
+
"pruned_heads": {},
|
| 129 |
+
"remove_invalid_values": false,
|
| 130 |
+
"repetition_penalty": 1.0,
|
| 131 |
+
"return_dict": true,
|
| 132 |
+
"return_dict_in_generate": false,
|
| 133 |
+
"rms_norm_eps": 1e-05,
|
| 134 |
+
"rope_scaling": {
|
| 135 |
+
"factor": 8.0,
|
| 136 |
+
"high_freq_factor": 4.0,
|
| 137 |
+
"low_freq_factor": 1.0,
|
| 138 |
+
"original_max_position_embeddings": 8192,
|
| 139 |
+
"rope_type": "llama3"
|
| 140 |
+
},
|
| 141 |
+
"rope_theta": 500000.0,
|
| 142 |
+
"sep_token_id": null,
|
| 143 |
+
"suppress_tokens": null,
|
| 144 |
+
"task_specific_params": null,
|
| 145 |
+
"temperature": 1.0,
|
| 146 |
+
"tf_legacy_loss": false,
|
| 147 |
+
"tie_encoder_decoder": false,
|
| 148 |
+
"tie_word_embeddings": false,
|
| 149 |
+
"tokenizer_class": null,
|
| 150 |
+
"top_k": 50,
|
| 151 |
+
"top_p": 1.0,
|
| 152 |
+
"torch_dtype": null,
|
| 153 |
+
"torchscript": false,
|
| 154 |
+
"typical_p": 1.0,
|
| 155 |
+
"use_bfloat16": false,
|
| 156 |
+
"use_cache": true,
|
| 157 |
+
"vocab_size": 266
|
| 158 |
+
},
|
| 159 |
+
"input_img_size": 224,
|
| 160 |
+
"model_type": "",
|
| 161 |
+
"num_vision_tokens_each_frame": 265,
|
| 162 |
+
"vision_model_config": {
|
| 163 |
+
"_attn_implementation_autoset": false,
|
| 164 |
+
"_name_or_path": "",
|
| 165 |
+
"add_cross_attention": false,
|
| 166 |
+
"architectures": null,
|
| 167 |
+
"bad_words_ids": null,
|
| 168 |
+
"begin_suppress_tokens": null,
|
| 169 |
+
"bos_token_id": null,
|
| 170 |
+
"chunk_size_feed_forward": 0,
|
| 171 |
+
"cross_attention_hidden_size": null,
|
| 172 |
+
"decoder_start_token_id": null,
|
| 173 |
+
"depth": 1,
|
| 174 |
+
"diversity_penalty": 0.0,
|
| 175 |
+
"do_sample": false,
|
| 176 |
+
"early_stopping": false,
|
| 177 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 178 |
+
"eos_token_id": null,
|
| 179 |
+
"exponential_decay_length_penalty": null,
|
| 180 |
+
"finetuning_task": null,
|
| 181 |
+
"forced_bos_token_id": null,
|
| 182 |
+
"forced_eos_token_id": null,
|
| 183 |
+
"hidden_dim": 192,
|
| 184 |
+
"id2label": {
|
| 185 |
+
"0": "LABEL_0",
|
| 186 |
+
"1": "LABEL_1"
|
| 187 |
+
},
|
| 188 |
+
"is_decoder": false,
|
| 189 |
+
"is_encoder_decoder": false,
|
| 190 |
+
"kernel_size": 16,
|
| 191 |
+
"label2id": {
|
| 192 |
+
"LABEL_0": 0,
|
| 193 |
+
"LABEL_1": 1
|
| 194 |
+
},
|
| 195 |
+
"length_penalty": 1.0,
|
| 196 |
+
"max_length": 20,
|
| 197 |
+
"min_length": 0,
|
| 198 |
+
"model_type": "",
|
| 199 |
+
"no_repeat_ngram_size": 0,
|
| 200 |
+
"num_beam_groups": 1,
|
| 201 |
+
"num_beams": 1,
|
| 202 |
+
"num_return_sequences": 1,
|
| 203 |
+
"out_dim": 192,
|
| 204 |
+
"output_attentions": false,
|
| 205 |
+
"output_hidden_states": false,
|
| 206 |
+
"output_scores": false,
|
| 207 |
+
"pad_token_id": null,
|
| 208 |
+
"prefix": null,
|
| 209 |
+
"problem_type": null,
|
| 210 |
+
"pruned_heads": {},
|
| 211 |
+
"remove_invalid_values": false,
|
| 212 |
+
"repetition_penalty": 1.0,
|
| 213 |
+
"return_dict": true,
|
| 214 |
+
"return_dict_in_generate": false,
|
| 215 |
+
"sep_token_id": null,
|
| 216 |
+
"suppress_tokens": null,
|
| 217 |
+
"task_specific_params": null,
|
| 218 |
+
"temperature": 1.0,
|
| 219 |
+
"temporal_patch_size": 1,
|
| 220 |
+
"tf_legacy_loss": false,
|
| 221 |
+
"tie_encoder_decoder": false,
|
| 222 |
+
"tie_word_embeddings": true,
|
| 223 |
+
"tokenizer_class": null,
|
| 224 |
+
"top_k": 50,
|
| 225 |
+
"top_p": 1.0,
|
| 226 |
+
"torch_dtype": null,
|
| 227 |
+
"torchscript": false,
|
| 228 |
+
"trunk_spatial_kernel_size": 3,
|
| 229 |
+
"trunk_temporal_kernel_size": 3,
|
| 230 |
+
"typical_p": 1.0,
|
| 231 |
+
"use_bfloat16": false
|
| 232 |
+
}
|
| 233 |
+
},
|
| 234 |
+
"gazing_ratio_config": {
|
| 235 |
+
"exponential": {
|
| 236 |
+
"gazing_ratio_max": 1,
|
| 237 |
+
"gazing_ratio_min": 0,
|
| 238 |
+
"lambda": 10
|
| 239 |
+
},
|
| 240 |
+
"fixed": {
|
| 241 |
+
"gazing_ratio": 0.75
|
| 242 |
+
},
|
| 243 |
+
"sample_strategy_during_inference": "fixed",
|
| 244 |
+
"sample_strategy_during_training": "fixed",
|
| 245 |
+
"uniform": {
|
| 246 |
+
"gazing_ratio_max": 1,
|
| 247 |
+
"gazing_ratio_min": 0
|
| 248 |
+
}
|
| 249 |
+
},
|
| 250 |
+
"gazing_ratio_each_frame_config": {
|
| 251 |
+
"dirichlet": {
|
| 252 |
+
"alpha": 0.5
|
| 253 |
+
},
|
| 254 |
+
"sample_strategy_during_inference": "uniform",
|
| 255 |
+
"sample_strategy_during_training": "self",
|
| 256 |
+
"self": null,
|
| 257 |
+
"uniform": null
|
| 258 |
+
},
|
| 259 |
+
"has_task_loss_requirement_during_inference": true,
|
| 260 |
+
"has_task_loss_requirement_during_training": false,
|
| 261 |
+
"image_mean": [
|
| 262 |
+
0.485,
|
| 263 |
+
0.456,
|
| 264 |
+
0.406
|
| 265 |
+
],
|
| 266 |
+
"image_std": [
|
| 267 |
+
0.229,
|
| 268 |
+
0.224,
|
| 269 |
+
0.225
|
| 270 |
+
],
|
| 271 |
+
"max_batch_size": null,
|
| 272 |
+
"max_num_frames": 16,
|
| 273 |
+
"model_type": "autogaze",
|
| 274 |
+
"num_vision_tokens_each_frame": 265,
|
| 275 |
+
"scales": "32+64+112+224",
|
| 276 |
+
"task_loss_requirement_config": {
|
| 277 |
+
"fixed": {
|
| 278 |
+
"task_loss_requirement": 0.7
|
| 279 |
+
},
|
| 280 |
+
"sample_strategy_during_inference": "fixed",
|
| 281 |
+
"sample_strategy_during_training": "uniform",
|
| 282 |
+
"uniform": {
|
| 283 |
+
"task_loss_requirement_max": 1.0,
|
| 284 |
+
"task_loss_requirement_min": 0.5
|
| 285 |
+
}
|
| 286 |
+
},
|
| 287 |
+
"transformers_version": "4.51.3",
|
| 288 |
+
"use_flash_attn": false
|
| 289 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a48e6a83a198368e3798420ff5d5df42af7c0003c9230f85581c39a2ea64e9eb
|
| 3 |
+
size 13039216
|
preprocessor_config.json
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"crop_size": {
|
| 3 |
+
"height": 224,
|
| 4 |
+
"width": 224
|
| 5 |
+
},
|
| 6 |
+
"do_center_crop": false,
|
| 7 |
+
"do_normalize": true,
|
| 8 |
+
"do_rescale": true,
|
| 9 |
+
"do_resize": true,
|
| 10 |
+
"image_mean": [
|
| 11 |
+
0.485,
|
| 12 |
+
0.456,
|
| 13 |
+
0.406
|
| 14 |
+
],
|
| 15 |
+
"image_processor_type": "AutoGazeImageProcessor",
|
| 16 |
+
"image_std": [
|
| 17 |
+
0.229,
|
| 18 |
+
0.224,
|
| 19 |
+
0.225
|
| 20 |
+
],
|
| 21 |
+
"offset": true,
|
| 22 |
+
"resample": 2,
|
| 23 |
+
"rescale_factor": 0.00784313725490196,
|
| 24 |
+
"size": {
|
| 25 |
+
"shortest_edge": 224
|
| 26 |
+
}
|
| 27 |
+
}
|