Skip to content

Latest commit

 

History

History
71 lines (54 loc) · 4.13 KB

File metadata and controls

71 lines (54 loc) · 4.13 KB

Machine Learning for Computer Vision Papers Reading

Pre-requisite:

Week - 1 [Back Propagation, Gradient Descent]

  • Learning representations by back-propagating errors [David E. Rumelhart, Geoffrey E. Hinton, Ronald J. Williams] [Link]

Week - 2 [Haar Wavelets and Morlet Wavelets]

  • Invariant Scattering Convolution Networks [Joan Bruna, Stephane Mallat] [Link]

Week - 3 [Back Propagation, SGD, Chain Rule in Maths]

  • ImageNet Classification with Deep Convolutional Neural Networks [Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton] [Link]
  • Very Deep Convolutional Networks for Large-Scale Image Recognition [Karen Simonyan, Andrew Zisserman] [arXiv]
  • Max-Pooling
  • Overfitting, Saturation and Dropout

Week - 4 [Intro to ResNet and Implementation]

  • Going Deeper with Convolutions [Christian Szegedy, Wei Liu, Andrew Rabinvich] [arXiv]
  • Deep Residual Learning for Image Recognition [Kaiming He, Xiangyu Zhang, Jian Sun] [arXiv]
  • Aggregated Residual Transformations for Deep Neural Networks [Saining Xie, Kaiming He] [arXiv]

Week - 5

N/A

Week - 6 [Image Segmentation Pixel-Level Classification]

  • Fully Convolutional Networks for Semantic Segmentation [Jonathan Long] [Link]
  • Fast Approximate Energy Minimization via Graph Cuts [Yuri Boykov] [Link]
  • Exact optimization for Markov random fields with convex priors [Hiroshi Ishikawa] [Link]
  • “GrabCut” — Interactive Foreground Extraction using Iterated Graph Cuts [Carsten Rother] [Link]

Week - 7

  • Synergistic Face Detection and Pose Estimation with Energy-Based Models [Margarita Osadchy, Yann LeCun, Matthew L. Miller] [Link]
  • Rapid Object Detection using a Boosted Cascade of Simple Features [Paul Viola, Michael Jones] [Link]
  • Real-Time Continuous Pose Recovery of Human Hands Using Convolutional Networks [NYU Hand Pose Dataset]

Week - 8

  • Hand Photo Data Collection

Week - 9

Week - 10 [Variational Auto Encoders]

  • Tutorial on Variational Autoencoders(VAEs), Carl Doersch. August, 2016. https://arxiv.org/abs/1606.05908
    • Relevance: sampling from distributions inside deep networks, and can be trained with stochastic gradient descent. VAEs have already shown promise in generating many kinds of complicated data, including handwritten digits, faces, house numbers, CIFAR images, physical models of scenes, segmentation, and predicting the future from static images.
  • "Variational Convolutional Networks for Human-Centric Annotations," 13th Asian Conference on Computer Vision, 2016. Tsung-Wei Ke, Che-Wei Lin, Tyng-Luh Liu and Davi Geiger,
    • Relevance: Use of VAEs to annotate automatically images.