releases.shpreview

v0.8.2dev0

v0.8.2dev0 Release

$npx -y @buildinternet/releases show rel_ZZa3x6CToE1GK56I9rArg

Part way through the conversion of models to multi-weight support (model_arch.pretrain_tag), module reorg for future building, and lots of new weights and model additions as we go...

This is considered a development release. Please stick to 0.6.x if you need stability. Some of the model names, tags will shift a bit, some old names have already been deprecated and remapping support not added yet. For code 0.6.x branch is considered 'stable' https://github.com/rwightman/pytorch-image-models/tree/0.6.x

Dec 23, 2022 🎄☃

  • Add FlexiViT models and weights from https://github.com/google-research/big_vision (check out paper at https://arxiv.org/abs/2212.08013)
    • NOTE currently resizing is static on model creation, on-the-fly dynamic / train patch size sampling is a WIP
  • Many more models updated to multi-weight and downloadable via HF hub now (convnext, efficientnet, mobilenet, vision_transformer*, beit)
  • More model pretrained tag and adjustments, some model names changed (working on deprecation translations, consider main branch DEV branch right now, use 0.6.x for stable use)
  • More ImageNet-12k (subset of 22k) pretrain models popping up:
    • efficientnet_b5.in12k_ft_in1k - 85.9 @ 448x448
    • vit_medium_patch16_gap_384.in12k_ft_in1k - 85.5 @ 384x384
    • vit_medium_patch16_gap_256.in12k_ft_in1k - 84.5 @ 256x256
    • convnext_nano.in12k_ft_in1k - 82.9 @ 288x288

Dec 8, 2022

  • Add 'EVA l' to vision_transformer.py, MAE style ViT-L/14 MIM pretrain w/ EVA-CLIP targets, FT on ImageNet-1k (w/ ImageNet-22k intermediate for some)
modeltop1param_countgmacmactshub
eva_large_patch14_336.in22k_ft_in22k_in1k89.2304.5191.1270.2link
eva_large_patch14_336.in22k_ft_in1k88.7304.5191.1270.2link
eva_large_patch14_196.in22k_ft_in22k_in1k88.6304.161.663.5link
eva_large_patch14_196.in22k_ft_in1k87.9304.161.663.5link

Dec 6, 2022

modeltop1param_countgmacmactshub
eva_giant_patch14_560.m30m_ft_in22k_in1k89.81014.41906.82577.2link
eva_giant_patch14_336.m30m_ft_in22k_in1k89.61013620.6550.7link
eva_giant_patch14_336.clip_ft_in1k89.41013620.6550.7link
eva_giant_patch14_224.clip_ft_in1k89.11012.6267.2192.6link

Dec 5, 2022

  • Pre-release (0.8.0dev0) of multi-weight support (model_arch.pretrained_tag). Install with pip install --pre timm
    • vision_transformer, maxvit, convnext are the first three model impl w/ support
    • model names are changing with this (previous _21k, etc. fn will merge), still sorting out deprecation handling
    • bugs are likely, but I need feedback so please try it out
    • if stability is needed, please use 0.6.x pypi releases or clone from 0.6.x branch
  • Support for PyTorch 2.0 compile is added in train/validate/inference/benchmark, use --torchcompile argument
  • Inference script allows more control over output, select k for top-class index + prob json, csv or parquet output
  • Add a full set of fine-tuned CLIP image tower weights from both LAION-2B and original OpenAI CLIP models
modeltop1param_countgmacmactshub
vit_huge_patch14_clip_336.laion2b_ft_in12k_in1k88.6632.5391407.5link
vit_large_patch14_clip_336.openai_ft_in12k_in1k88.3304.5191.1270.2link
vit_huge_patch14_clip_224.laion2b_ft_in12k_in1k88.2632167.4139.4link
vit_large_patch14_clip_336.laion2b_ft_in12k_in1k88.2304.5191.1270.2link
vit_large_patch14_clip_224.openai_ft_in12k_in1k88.2304.281.188.8link
vit_large_patch14_clip_224.laion2b_ft_in12k_in1k87.9304.281.188.8link
vit_large_patch14_clip_224.openai_ft_in1k87.9304.281.188.8link
vit_large_patch14_clip_336.laion2b_ft_in1k87.9304.5191.1270.2link
vit_huge_patch14_clip_224.laion2b_ft_in1k87.6632167.4139.4link
vit_large_patch14_clip_224.laion2b_ft_in1k87.3304.281.188.8link
vit_base_patch16_clip_384.laion2b_ft_in12k_in1k87.286.955.5101.6link
vit_base_patch16_clip_384.openai_ft_in12k_in1k8786.955.5101.6link
vit_base_patch16_clip_384.laion2b_ft_in1k86.686.955.5101.6link
vit_base_patch16_clip_384.openai_ft_in1k86.286.955.5101.6link
vit_base_patch16_clip_224.laion2b_ft_in12k_in1k86.286.617.623.9link
vit_base_patch16_clip_224.openai_ft_in12k_in1k85.986.617.623.9link
vit_base_patch32_clip_448.laion2b_ft_in12k_in1k85.888.317.923.9link
vit_base_patch16_clip_224.laion2b_ft_in1k85.586.617.623.9link
vit_base_patch32_clip_384.laion2b_ft_in12k_in1k85.488.313.116.5link
vit_base_patch16_clip_224.openai_ft_in1k85.386.617.623.9link
vit_base_patch32_clip_384.openai_ft_in12k_in1k85.288.313.116.5link
vit_base_patch32_clip_224.laion2b_ft_in12k_in1k83.388.24.45link
vit_base_patch32_clip_224.laion2b_ft_in1k82.688.24.45link
vit_base_patch32_clip_224.openai_ft_in1k81.988.24.45link
  • Port of MaxViT Tensorflow Weights from official impl at https://github.com/google-research/maxvit
    • There was larger than expected drops for the upscaled 384/512 in21k fine-tune weights, possible detail missing, but the 21k FT did seem sensitive to small preprocessing
modeltop1param_countgmacmactshub
maxvit_xlarge_tf_512.in21k_ft_in1k88.5475.8534.11413.2link
maxvit_xlarge_tf_384.in21k_ft_in1k88.3475.3292.8668.8link
maxvit_base_tf_512.in21k_ft_in1k88.2119.9138704link
maxvit_large_tf_512.in21k_ft_in1k88212.3244.8942.2link
maxvit_large_tf_384.in21k_ft_in1k88212132.6445.8link
maxvit_base_tf_384.in21k_ft_in1k87.9119.673.8332.9link
maxvit_base_tf_512.in1k86.6119.9138704link
maxvit_large_tf_512.in1k86.5212.3244.8942.2link
maxvit_base_tf_384.in1k86.3119.673.8332.9link
maxvit_large_tf_384.in1k86.2212132.6445.8link
maxvit_small_tf_512.in1k86.169.167.3383.8link
maxvit_tiny_tf_512.in1k85.73133.5257.6link
maxvit_small_tf_384.in1k85.56935.9183.6link
maxvit_tiny_tf_384.in1k85.13117.5123.4link
maxvit_large_tf_224.in1k84.9211.843.7127.4link
maxvit_base_tf_224.in1k84.9119.52495link
maxvit_small_tf_224.in1k84.468.911.753.2link
maxvit_tiny_tf_224.in1k83.430.95.635.8link

Oct 15, 2022

  • Train and validation script enhancements
  • Non-GPU (ie CPU) device support
  • SLURM compatibility for train script
  • HF datasets support (via ReaderHfds)
  • TFDS/WDS dataloading improvements (sample padding/wrap for distributed use fixed wrt sample count estimate)
  • in_chans !=3 support for scripts / loader
  • Adan optimizer
  • Can enable per-step LR scheduling via args
  • Dataset 'parsers' renamed to 'readers', more descriptive of purpose
  • AMP args changed, APEX via --amp-impl apex, bfloat16 supportedf via --amp-dtype bfloat16
  • main branch switched to 0.7.x version, 0.6x forked for stable release of weight only adds
  • master -> main branch rename

Fetched April 7, 2026

v0.8.2dev0 — timm (pytorch-image-models) — releases.sh