Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
* fix: loading of in21k vit * fix: config for gh200 * fix: hostfile * fix: hostfile * fix: requirements * fix: multinode * fix: impor tos * fix: model name arg, dataset size, text query * chore: ignore ids to keep * feat: build index and save embeddings + rankings * fix: update for cp medium * fix: update build_map_index_filter * fix: save faster, remove expensive gather * fix: option to save * fix: reqs * fix: protobuf * fix: beam req * fix: beam req * fix: faster multi-gpu index filteirng * feat: filter uuids from json files * fix: print statement * fix: remove duplicates * fix: remove unused * feat: precompute text embeddings * fix: remove unused * fix: updates * fix: download and process * fix: working now * feat: cross attention/coca-esque model * fix: reorder tar so npy near others * feat: add back sampling with replacement (need to test) * feat: train with pretrained * fix: precomputed * feat: cpu offload faster than regular opt w grad checkp * fix: can load and resample now! * feat: sugarcrepe eval * fix: prefix for imagenet * fix: add prefix in trainer for eval * feat: datacomp evals for contrators models * fix: prefix + print after evals * fix: if empty set to none * fix: dataset-size pass via cli * fix: fairness eval * fix: allow null path for imagenet for testing * feat: mlm + contrastive loss * fix: imagenet fixes * fix: deepspeed config * fix: imagenet eval * feat: three towers image <-> text, text <-> frozen * fix: eval steps * fix: hf model updates * fix: vit pos embed * feat: three towers current * fix: multinode fixes * fix: global rank in multinode * fix: progbar only global rank 0 * feat: higher lr * fix: eval strategy epochs logging fix * feat: no clamp logits config * feat: 3 epoch training * feat: update hostfile * fix: 10 epochs * fix: update hostfile * feat: upload embs to atlas * feat: dino v1 * fix: grad check * fix: clip model * feat: 32k vit-l * fix: update hostfile * fix: workers * fix: more logging * fix: no wandb for now * fix: try smaller vit * fix: try more ds stuff * fix: try openclip loss * fix: remove unneeded print * fix: test clip loss * fix: 32k run * fix: are evals broken? * fix: 16k testing * fix: evals * fix: evals * chore: logging * fix: remove prints * feat: config * fix: remove rng, trust openclip * fix: idk? * fix: path * fix: rank * feat: ok now working L14 * feat: 32k higher lr exp * feat: fb vit mae * feat: mae train * fix: map mae * fix: sp * fix: batch size * feat: 10epoch 65k * feat: higher lr * feat: no wd * feat: long train * feat: 81k bs * feat: 3 epoch 65k * feat: 10 epoch * fix: large 3 epoch train * fix: workers * fix: model utils loading * fix: dataloader for datacomp1b * fix: remove pdb * fix: workers * feat: dfn 2b * fix: bs * fix: bs * fix: wandb * fix: imagenet workers * feat: try unidirectional * fix: path for old h100 * fix: map * fix: lets try this again * fix: try fusing * fix: bad code * fix: 32k map fix * fix: bs and default get for dataset * fix: fused * fix; dumb * fix: try this * feat: pos embed with swiglu gated * fix: patch size * fix: runs now * fix: back to mlp * fix: stage 3? * fix: try again * fix: remove pos embed * fix: wtf * feat: mean pool test again? * feat: augments * fix: try no checkpointing * feat: 3 epoch augmentation train * fix: no randaugment * fix: dataset size * feat: 65k run with augs * fix: imagenet path * feat: try resume training multinode * fix: hostfile * fix: no flip for this train * fix: imagenet * refactor: remove unused * refactor: rename text_encoder -> nomic_encoder * refactor: remove captioner * chore: bump pydantic >= 2.0.0 * feat: eval for clip models * feat: v1.5 config * fix: hf code * refactor: move hf tests to separate * chore: remove unused * refactor: remove * refactor: unused code * refactor: not used * fix: remove unused * refactor: remove xattn * refactor: remove xattn * fix: try to resume * fix: v1.5 * fix: remove unused import * fix: remove ema * fix: remove ema * fix: instructions * feat: tracing code * feat: add stacks * feat: export_stacks=True * fix: with_stack * fix: tensorboard profiling (kind of) working * fix: don't profile, test full thing * feat: moar batch * feat: train * refactor: clean up code * feat: download data * fix: pydantic, workers crashing * fix: prefix * chore: ignore data folder * feat: loadable hf model * fix: map pooling bug * fix: comment old pooling * feat: flickr eval running * feat: flickr to config * feat: flickr eval train * fix: flickr eval doesn't hang * feat: biencoder test * fix: enforce no dynamic ntk * feat: unidirectional * feat: base timm models * fix: simplify vit pos_embed * fix: cls token confusion * feat: timm dinov2 with registers * wip vit rotary * feat: yolo 65k scratch vit * fix: hostfile * fix: revert back to bidirectional * fix: spelling * fix: path * fix: wandb * fix: shards * fix: reqs * feat: eva-style models, timm vit-base * fix: timm vit-b 224 image * feat: timm vit-b-16 first experiment * fix: no flip * feat: eva02 vit base * feat: pooling heads from timm vit * feat: add augreg vits as option * fix: remove pooling heads * fix: dumb renaming of model so eva loads with autoconfig * feat: eva config for training * fix: model loading * feat: 65k eva 3 epoch train * feat: map no clamp * fix: hostfile * fix: reduce workers * fix: no clamp * fix: config * feat: v1.5 train * fix: hostfile + config * fix: config for lower lr * fix: hamming * fix: train * feat: hf vision model code * fix: hostfile * fix: path * refactor: clean up code base * refactor: rename * fix: remove hostfile * refactor: remove sugarcrepe * style: black and isort * docs: readme and config fixes * fix: trainers, come back later
- Loading branch information