Skip to content

Conversation

@rwightman
Copy link
Collaborator

  • remove dud attention, involution + my swin attention adaptation don't seem worth keeping
  • add or update several new 26/50 layer ResNe(X)t variants that were used in experiments
  • remove models associated with dead-end or uninteresting experiment results
  • weights coming soon...
* remove dud attention, involution + my swin attention adaptation don't seem worth keeping * add or update several new 26/50 layer ResNe(X)t variants that were used in experiments * remove models associated with dead-end or uninteresting experiment results * weights coming soon...
…p / fixes for byoanet. Rename resnet26ts to tfs to distinguish (extra fc).
* add polynomial decay 'poly' * cleanup cycle specific args for cosine, poly, and tanh sched, t_mul -> cycle_mul, decay -> cycle_decay, default cycle_limit to 1 in each opt * add k-decay for cosine and poly sched as per https://arxiv.org/abs/2004.05909 * change default tanh ub/lb to push inflection to later epochs
…g as an alternate to CE w/ smoothing. For training experiments.
…ll trying to improve reliability of sgd test.
@rwightman rwightman merged commit a6e8598 into master Sep 14, 2021
guoriyue pushed a commit to guoriyue/pytorch-image-models that referenced this pull request May 24, 2024
Update attention / self-attn based models from a series of experiments
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

2 participants