2020 한국인공지능학회 동계강좌 정리 – 8. UNIST 임성빈 교수님, Automated Machine Learning for Visual Domain

This entry is part 8 of 9 in the series 2020 한국인공지능학회 동계강좌

 

2020 인공지능학회 동계강좌를 신청하여 2020.1.8 ~ 1.10 3일 동안 다녀왔다. 총 9분의 연사가 나오셨는데, 프로그램 일정은 다음과 같다.

전체를 묶어서 하나의 포스트로 작성하려고 했는데, 주제마다 내용이 꽤 많을거 같아, 한 강좌씩 시리즈로 묶어서 작성하게 되었다. 여덟 번째 포스트에서는 UNIST 임성빈 교수님의 “Automated Machine Learning for Visual Domain” 강연 내용을 다룬다.

  1. Introduction
    1. kakao brain on Competitions
      • Data Science Bawl 2018 [Link]
      • Youtube 8m Challenge [Link]
      • GQA Challenge 2019 [Link]
      • BEA 2019 GEC Challenge [Link]
      • NeurIPS 2019 AutoCV Challenge [Link]
      • Medical Segmentation Decathlon [Link]
        • Multi-task learning
          1. Characteristics
            • 3D-image
            • small amount of data
            • unbalanced labels
            • large-range object scales (small object ~ large object)
            • multi-class labels
          2. Multi-task learning + knowledge transfer
            • From known tasks, ex) liver, brain, hippocampus, …
            • To unknown tasks, ex) colon cancer, hepatic vessels, …
        • Baseline : 3D U-Net
          1. Kayalibay et al, CNN-based Segmentation of Medical Imaging Data, 2017
  2. NAS
    1. Introduction
      •  Definition
        • The process of automating architecture engineering
        • NAS can be seen as subfield of AutoML
        • significant overlap with :
          1. Hyperparameter optimization
          2. Meta learning
      • Literature
        • AutoML.org Freiburg [Link]
      • Categorizing NAS
        • Elsken et al, Neural Architecture Search: A Survey, 2019
    2. Search Space
      • Soft Start
        • Chain-structured Neural Nets
        • Multi-branch Neural Nets
      • Network Structure Code
        • Zhong et al, BlockQNN: Efficient Block-wise Neural Network Architecture Generation, CVPR 2018
      • CNN Design
        • Pham et al, Efficient Neural Architecture Search via Parameter Sharing, ICML 2018
        • Perez-Rua et al, Efficient Progressive Neural Architecture Search, BMVC 2018
      • Cell Design
        • Zoph et al, Learning Transferable Architectures for Scalable Image Recognition, CVPR 2018
        • Liu et al, Progressive Neural Architecture Search, ECCV 2018
    3. Search Strategy
      • Reinforcement Learning
        • Model the layer selection process as a MDP
        • Baker et al, Designing Neural Network Architectures using Reinforcement Learning, ICLR 2017
        • Zoph, Neural Architecture Search with Reinforcement Learning, ICLR 2018
      • Evolution method
        • Zoph et al, Learning Transferable Architectures for Scalable Image Recognition, CVPR 2018
        • AmoebaNet
          1. Real et al, Regularized Evolution for Image Classifier Architecture Search, AAAI 2019
          2. SOTA at ImageNet classification
        • Liu et al, Progressive Neural Architecture Search, ECCV 2018
      • Gradient Based
        • Hyperparameter optimization
          1. Shin et al, DIFFERENTIABLE NEURAL NETWORK ARCHITECTURE SEARCH, ICLR 2018 Workshop
        • Architecture selection
          1. Saxena & Verbeek, Convolutional Neural Fabrics, NIPS 2016
        • Connectivity selection
          1. Ahmed & Torresani, MaskConnect: Connectivity Learning by Gradient Descent, ECCV 2018
        • Micro-search strategy
          1. Liu et al, DARTS: Differentiable Architecture Search, ICLR 2019
            • Optimization
              \min_{\alpha} L_{val} ( w^*(\alpha), \alpha )
              s.t. \quad w^*(\alpha) = argmin_w \quad L_{train} (w, \alpha)
            • => inner loop : minimize w*
              => outer loop : minimize \alpha^*
            • DARTS needs Fine-tuning
        • Softmax annealing
          1. Xie et al, SNAS: Stochastic Neural Architecture Search, ICLR 2019
            • Softmax annealing
              \bar{o}^{(i, j)}(x) = \sum_{o \in O} \dfrac{exp(\alpha^{(i,j)}_o)}{\sum_{o' \in O} exp(\alpha^{(i,j)}_{o'})}o(x) \rightarrow \sum_{o \in O} \dfrac{exp(\alpha^{(i,j)}_o / \tau)}{\sum_{o' \in O} exp(\alpha^{(i,j)}_{o'} / \tau)}o(x)
            • Gumbel noise
    4. Scalable NAS
      • Applications to 3D Medical Images Segmentation
      • Baseline : 3D U-Net
      • Problem
        • DARTS on 3D images
          1. Out of memory
            • brain MRI => Big dimension (155 * 240 * 240)
            • backpropagation is infeasible
      • Stochastic NAS + operation sampling
        • Kim et al, Scalable Neural Architecture Search for 3D Medical Image Segmentation, MICCAI 2019
        • Scalable NAS is transferrable
      • torchgpipe [Link]
        • A GPipe implementation in PyTorch.
        • GPipe is a scalable pipeline parallelism library published by Google Brain
    5. AutoAugment
      • Automated Data Augmentation Strategies Search
      • Problem
        • Incomplete Data
          1. Data with uncertainty (wrong labels)
          2. Inconsistent data (Doctor, Resident Doctor, Internship)
          3. Insufficient data ( lack of data )
      • AutoAug = automated augmentation search
        • Cubuk et al, AutoAugment: Learning Augmentation Strategies from Data, CVPR 2019
      • AutoAugment for Insufficient Data
        • Lim et al, Fast AutoAugment, NIPS 2019
          1. Find augmentation policies without human intervention
      • Fast Training with Time Constraint
        • Automatic Computationally LIght Network Transfer, 2020 [Link]
          1. 20min with 1GPU + Fast AutoAugment
    6. Further Reading
      • Liu et al, Auto-DeepLab: Hierarchical Neural Architecture Search for Semantic Image Segmentation, CVPR 2019
      • Chen et al, DetNAS: Backbone Search for Object Detection, NIPS 2019
      • Cai et al, ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware, ICLR 2019
      • Cubuk et al, RandAugment: Practical automated data augmentation with a reduced search space, arXiv 2019
      • Hataya et al, Faster AutoAugment: Learning Augmentation Strategies using Backpropagation, arXiv 2019
      • Zhang et al, Adversarial AutoAugment, ICLR 2020
Series Navigation<< 2020 한국인공지능학회 동계강좌 정리 – 7. AITrics 김세훈 박사님, Meta Learning for Few-shot Classification2020 한국인공지능학회 동계강좌 정리 – 9. 연세대 황승원 교수님, Knowledge in Neural NLP >>

Leave a Comment