티스토리 뷰

NeurIPS 2023, Arxiv Link

 

What they did

  • They propose "Battle-of-the-Backones"(BoB), benchmarking a diverse suite of (vision) pre-trained models, including classic ImageNet-trained CNN, vision-language models (perhaps CLIP). 

  • The target tasks are OOD-generalization, classification, object-detection and Image retrieval.

Key observations

  • with more than 1,500 training runs, they found that CNN pretrained in a supervised fashion still perfrom best on most tasks.

IN standing for the ImageNet, which represents the supervise-trained CNNs

  • (My opinion 1) it is not suprising that the supervisly-pre-trained models (IN-nK in the above figure) works better on "supervised-learning classification problem."
  • (My opinion 2) the recent attentions and success of the ViT architecture is not because its model architecture is well-suited for a vision task---in other words, it has a good inductive-bias for vision tasks)---, but it is because the transformer-like architecture has been well-optimized throughout the deep-learning community after the advent of the LLM.
  • ViTs are more sensitive to the amount of the data and number of parameters than CNNs
    • (My opinion 3) this may claim that the ViT architecture may scale better than CNN, at the some scale of data and model parameters. (In fact, the authors already discuss it in Sec 5.)
  • Monocular depth-estimation achieves performances competitive with top-conventional supervised and SSL backbones -> the author claims it implies the depth-estimation tasks may serve as a powerful and generalizable primary or auxiliary pre-training task.
    • (Q) Even pre-trained depth-estimation task can work well on the generative task such as the Stable-Diffusion?? (Fortunately, the input-output dimension is well-matched for both tasks unlike other tasks like classification or others, meaning no need to modify the model itself.)

And there are many other insightful observation in the paper! please check-out the original paper.

공지사항
최근에 올라온 글
최근에 달린 댓글
Total
Today
Yesterday
링크
«   2024/09   »
1 2 3 4 5 6 7
8 9 10 11 12 13 14
15 16 17 18 19 20 21
22 23 24 25 26 27 28
29 30
글 보관함