| ## 2.24.0 | |
| * Fix missing space in error message | |
| * use model flag for normalizing embeddings | |
| * init logit_bias for non siglip pretrained models | |
| * Fix logit_bias load_checkpoint addition | |
| * Make CoCa model match CLIP models for logit scale/bias init | |
| * Fix missing return of "logit_bias" in CoCa.forward | |
| * Add NLLB-CLIP with SigLIP models | |
| * Add get_logits method and NLLB tokenizer | |
| * Remove the empty file src/open_clip/generation_utils.py | |
| * Update params.py: "BatchNorm" -> "LayerNorm" in the description string for "--lock-text-freeze-layer-norm" | |
| ## 2.23.0 | |
| * Add CLIPA-v2 models | |
| * Add SigLIP models | |
| * Add MetaCLIP models | |
| * Add NLLB-CLIP models | |
| * CLIPA train code | |
| * Minor changes/fixes | |
| * Remove protobuf version limit | |
| * Stop checking model name when loading CoCa models | |
| * Log native wandb step | |
| * Use bool instead of long masks | |
| ## 2.21.0 | |
| * Add SigLIP loss + training support | |
| * Add more DataComp models (B/16, B/32 and B/32@256) | |
| * Update default num workers | |
| * Update CoCa generation for `transformers>=4.31` | |
| * PyTorch 2.0 `state_dict()` compatibility fix for compiled models | |
| * Fix padding in `ResizeMaxSize` | |
| * Convert JIT model on state dict load for `pretrained='filename…'` | |
| * Other minor changes and fixes (typos, README, dependencies, CI) | |
| ## 2.20.0 | |
| * Add EVA models | |
| * Support serial worker training | |
| * Fix Python 3.7 compatibility | |
| ## 2.19.0 | |
| * Add DataComp models | |
| ## 2.18.0 | |
| * Enable int8 inference without `.weight` attribute | |
| ## 2.17.2 | |
| * Update push_to_hf_hub | |
| ## 2.17.0 | |
| * Add int8 support | |
| * Update notebook demo | |
| * Refactor zero-shot classification code | |
| ## 2.16.2 | |
| * Fixes for context_length and vocab_size attributes | |
| ## 2.16.1 | |
| * Fixes for context_length and vocab_size attributes | |
| * Fix --train-num-samples logic | |
| * Add HF BERT configs for PubMed CLIP model | |
| ## 2.16.0 | |
| * Add improved g-14 weights | |
| * Update protobuf version | |
| ## 2.15.0 | |
| * Add convnext_xxlarge weights | |
| * Fixed import in readme | |
| * Add samples per second per gpu logging | |
| * Fix slurm example | |
| ## 2.14.0 | |
| * Move dataset mixtures logic to shard level | |
| * Fix CoCa accum-grad training | |
| * Safer transformers import guard | |
| * get_labels refactoring | |
| ## 2.13.0 | |
| * Add support for dataset mixtures with different sampling weights | |
| * Make transformers optional again | |
| ## 2.12.0 | |
| * Updated convnext configs for consistency | |
| * Added input_patchnorm option | |
| * Clean and improve CoCa generation | |
| * Support model distillation | |
| * Add ConvNeXt-Large 320x320 fine-tune weights | |
| ## 2.11.1 | |
| * Make transformers optional | |
| * Add MSCOCO CoCa finetunes to pretrained models | |
| ## 2.11.0 | |
| * coca support and weights | |
| * ConvNeXt-Large weights | |
| ## 2.10.1 | |
| * `hf-hub:org/model_id` support for loading models w/ config and weights in Hugging Face Hub | |
| ## 2.10.0 | |
| * Added a ViT-bigG-14 model. | |
| * Added an up-to-date example slurm script for large training jobs. | |
| * Added a option to sync logs and checkpoints to S3 during training. | |
| * New options for LR schedulers, constant and constant with cooldown | |
| * Fix wandb autoresuming when resume is not set | |
| * ConvNeXt `base` & `base_w` pretrained models added | |
| * `timm-` model prefix removed from configs | |
| * `timm` augmentation + regularization (dropout / drop-path) supported | |
| ## 2.9.3 | |
| * Fix wandb collapsing multiple parallel runs into a single one | |
| ## 2.9.2 | |
| * Fix braceexpand memory explosion for complex webdataset urls | |
| ## 2.9.1 | |
| * Fix release | |
| ## 2.9.0 | |
| * Add training feature to auto-resume from the latest checkpoint on restart via `--resume latest` | |
| * Allow webp in webdataset | |
| * Fix logging for number of samples when using gradient accumulation | |
| * Add model configs for convnext xxlarge | |
| ## 2.8.2 | |
| * wrapped patchdropout in a torch.nn.Module | |
| ## 2.8.1 | |
| * relax protobuf dependency | |
| * override the default patch dropout value in 'vision_cfg' | |
| ## 2.8.0 | |
| * better support for HF models | |
| * add support for gradient accumulation | |
| * CI fixes | |
| * add support for patch dropout | |
| * add convnext configs | |
| ## 2.7.0 | |
| * add multilingual H/14 xlm roberta large | |
| ## 2.6.1 | |
| * fix setup.py _read_reqs | |
| ## 2.6.0 | |
| * Make openclip training usable from pypi. | |
| * Add xlm roberta large vit h 14 config. | |
| ## 2.5.0 | |
| * pretrained B/32 xlm roberta base: first multilingual clip trained on laion5B | |
| * pretrained B/32 roberta base: first clip trained using an HF text encoder | |
| ## 2.4.1 | |
| * Add missing hf_tokenizer_name in CLIPTextCfg. | |
| ## 2.4.0 | |
| * Fix #211, missing RN50x64 config. Fix type of dropout param for ResNet models | |
| * Bring back LayerNorm impl that casts to input for non bf16/fp16 | |
| * zero_shot.py: set correct tokenizer based on args | |
| * training/params.py: remove hf params and get them from model config | |
| ## 2.3.1 | |
| * Implement grad checkpointing for hf model. | |
| * custom_text: True if hf_model_name is set | |
| * Disable hf tokenizer parallelism | |
| ## 2.3.0 | |
| * Generalizable Text Transformer with HuggingFace Models (@iejMac) | |
| ## 2.2.0 | |
| * Support for custom text tower | |
| * Add checksum verification for pretrained model weights | |
| ## 2.1.0 | |
| * lot including sota models, bfloat16 option, better loading, better metrics | |
| ## 1.2.0 | |
| * ViT-B/32 trained on Laion2B-en | |
| * add missing openai RN50x64 model | |
| ## 1.1.1 | |
| * ViT-B/16+ | |
| * Add grad checkpointing support | |
| * more robust data loader | |