Skip to content

Kronopath/Machine-Learning-For-Computer-Vision

 
 

Repository files navigation

Machine Learning for Computer Vision Papers Reading

Pre-requisite:

Week - 1 [Back Propagation, Gradient Descent]

  • Learning representations by back-propagating errors [David E. Rumelhart, Geoffrey E. Hinton, Ronald J. Williams] [Link]

Week - 2 [Haar Wavelets and Morlet Wavelets]

  • Invariant Scattering Convolution Networks [Joan Bruna, Stephane Mallat] [Link]

Week - 3 [Back Propagation, SGD, Chain Rule in Maths]

  • ImageNet Classification with Deep Convolutional Neural Networks [Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton] [Link]
  • Very Deep Convolutional Networks for Large-Scale Image Recognition [Karen Simonyan, Andrew Zisserman] [arXiv]
  • Max-Pooling
  • Overfitting, Saturation and Dropout

Week - 4 [Intro to ResNet and Implementation]

  • Going Deeper with Convolutions [Christian Szegedy, Wei Liu, Andrew Rabinvich] [arXiv]
  • Deep Residual Learning for Image Recognition [Kaiming He, Xiangyu Zhang, Jian Sun] [arXiv]
  • Aggregated Residual Transformations for Deep Neural Networks [Saining Xie, Kaiming He] [arXiv]

Week - 5

N/A

Week - 6 [Image Segmentation Pixel-Level Classification]

  • Fully Convolutional Networks for Semantic Segmentation [Jonathan Long] [Link]
  • Fast Approximate Energy Minimization via Graph Cuts [Yuri Boykov] [Link]
  • Exact optimization for Markov random fields with convex priors [Hiroshi Ishikawa] [Link]
  • “GrabCut” — Interactive Foreground Extraction using Iterated Graph Cuts [Carsten Rother] [Link]

Week - 7

  • Synergistic Face Detection and Pose Estimation with Energy-Based Models [Margarita Osadchy, Yann LeCun, Matthew L. Miller] [Link]
  • Rapid Object Detection using a Boosted Cascade of Simple Features [Paul Viola, Michael Jones] [Link]
  • Real-Time Continuous Pose Recovery of Human Hands Using Convolutional Networks [NYU Hand Pose Dataset]

Week - 8

  • Hand Photo Data Collection

Week - 9

Week - 10 [Variational Auto Encoders]

  • Tutorial on Variational Autoencoders(VAEs), Carl Doersch. August, 2016. https://arxiv.org/abs/1606.05908
    • Relevance: sampling from distributions inside deep networks, and can be trained with stochastic gradient descent. VAEs have already shown promise in generating many kinds of complicated data, including handwritten digits, faces, house numbers, CIFAR images, physical models of scenes, segmentation, and predicting the future from static images.
  • "Variational Convolutional Networks for Human-Centric Annotations," 13th Asian Conference on Computer Vision, 2016. Tsung-Wei Ke, Che-Wei Lin, Tyng-Luh Liu and Davi Geiger,
    • Relevance: Use of VAEs to annotate automatically images.

About

Paper reading and takeaways in ML and CV fields. [Updating]

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 95.2%
  • Python 4.8%