nielsr HF staff commited on
Commit
5a14841
1 Parent(s): c42fd79

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -24,13 +24,15 @@ Disclaimer: The team releasing Swin Transformer v2 did not write a model card fo
24
 
25
  The Swin Transformer is a type of Vision Transformer. It builds hierarchical feature maps by merging image patches (shown in gray) in deeper layers and has linear computation complexity to input image size due to computation of self-attention only within each local window (shown in red). It can thus serve as a general-purpose backbone for both image classification and dense recognition tasks. In contrast, previous vision Transformers produce feature maps of a single low resolution and have quadratic computation complexity to input image size due to computation of self-attention globally.
26
 
 
 
27
  ![model image](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/swin_transformer_architecture.png)
28
 
29
  [Source](https://paperswithcode.com/method/swin-transformer)
30
 
31
  ## Intended uses & limitations
32
 
33
- You can use the raw model for image classification. See the [model hub](https://huggingface.co/models?search=swin) to look for
34
  fine-tuned versions on a task that interests you.
35
 
36
  ### How to use
@@ -56,7 +58,7 @@ predicted_class_idx = logits.argmax(-1).item()
56
  print("Predicted class:", model.config.id2label[predicted_class_idx])
57
  ```
58
 
59
- For more code examples, we refer to the [documentation](https://huggingface.co/transformers/model_doc/swin.html#).
60
 
61
  ### BibTeX entry and citation info
62
 
 
24
 
25
  The Swin Transformer is a type of Vision Transformer. It builds hierarchical feature maps by merging image patches (shown in gray) in deeper layers and has linear computation complexity to input image size due to computation of self-attention only within each local window (shown in red). It can thus serve as a general-purpose backbone for both image classification and dense recognition tasks. In contrast, previous vision Transformers produce feature maps of a single low resolution and have quadratic computation complexity to input image size due to computation of self-attention globally.
26
 
27
+ Swin Transformer v2 adds 3 main improvements: 1) a residual-post-norm method combined with cosine attention to improve training stability; 2) a log-spaced continuous position bias method to effectively transfer models pre-trained using low-resolution images to downstream tasks with high-resolution inputs; 3) a self-supervised pre-training method, SimMIM, to reduce the needs of vast labeled images.
28
+
29
  ![model image](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/swin_transformer_architecture.png)
30
 
31
  [Source](https://paperswithcode.com/method/swin-transformer)
32
 
33
  ## Intended uses & limitations
34
 
35
+ You can use the raw model for image classification. See the [model hub](https://huggingface.co/models?search=swinv2) to look for
36
  fine-tuned versions on a task that interests you.
37
 
38
  ### How to use
 
58
  print("Predicted class:", model.config.id2label[predicted_class_idx])
59
  ```
60
 
61
+ For more code examples, we refer to the [documentation](https://huggingface.co/transformers/model_doc/swinv2.html#).
62
 
63
  ### BibTeX entry and citation info
64