[学习笔记] CS131 Computer Vision: Foundations and Applications:Lecture 9 深度学习2

深度学习

So far this week

  • Edge detection
  • RANSAC
  • SIFT
  • K-Means
  • Linear classifier
  • Mean-shift
  • PCA/Eigenfaces
  • Image features

    Current Research

  • Learning hierarchical representations from data
  • End-to-end learning: raw inputs to predictions
  • can use a small set of simple tools to solve many problems
  • has led to rapid progress on many problems
  • Inspired by the brain(very loosely!)

Deep learning for different problems

vision tasks

  • visual recognition

  • object detection: what and where
  • object segmentation
  • image caption
  • visual question answering
  • super resolution
  • image retrieval
  • style transfer

    outside vision tasks

  • Machine Translation
  • Text Synthesis
  • Speech Recognition
  • Speech Synthesis

Motivation

Data-driven approach:

  1. collect a dataset of images and labels
  2. use machine learning to train an image calssifier
  3. evaluate the classifier on a withheld set of test images

feature learning
what is feature learning?[^what is feature learning]

deep learning

Supervised learning

linear regression

neural network

neural networks with many layers

Gradient descent

how to find the best weights \(w^T\)

which way is down hill?

gradient descent

fancier rules:

  • Momentum
  • NAG
  • Adagrad
  • Adadelta
  • Rmsprop

这里以后可以再 看看!

Backpropagation

a two-layer neural network in 25 lines of code

import numpy as np
D,H,N = 8, 64,32
#randomly initialize weights
W1 = np.random.randn(D,H)
W2 = np.random.randn(H,D)
for t in xrange(10000):
    x = np.random.randn(N,D)
    y = np.sin(x)
    s = x.dot(W1)
    a = np.maxium(s,0)
    y_hat = a.dot(W2)
    loss = 0.5*np.sum((y_hat-y)**2.0)
    dy_hat = y_hat - y
    dW2 = a.T.dot(W2.T)
    da = dy_hat.dot(W2.T)
    ds = (s > 0)*da
    dW1 = x.T.dot(ds)
    W1 -= learning_rate*dW1
    W2 -= learning_rate*dW2

[^what is feature learning]:
In Machine Learning, feature learning or representation learningis a set of techniques that learn a feature: a transformation of raw data input to a representation that can be effectively exploited in machine learning tasks. This obviates manual feature engineering, which is otherwise necessary, and allows a machine to both learn at a specific task (using the features) and learn the features themselves.
Feature learning is motivated by the fact that machine learning tasks such as classification often require input that is mathematically and computationally convenient to process. However, real-world data such as images, video, and sensor measurement is usually complex, redundant, and highly variable. Thus, it is necessary to discover useful features or representations from raw data. Traditional hand-crafted features often require expensive human labor and often rely on expert knowledge. Also, they normally do not generalize well. This motivates the design of efficient feature learning techniques, to automate and generalize this.
Feature learning can be divided into two categories: supervised and unsupervised feature learning, analogous to these categories in machine learning generally.
In supervised feature learning, features are learned with labeled input data. Examples include Supervised Neural Networks, Multilayer Perceptron, and (supervised) dictionary Learning.
In unsupervised feature learning, features are learned with unlabeled input data. Examples include dictionary learning, independent component analysis, autoencoders, and various forms of clustering.

时间: 2024-08-03 21:00:11

[学习笔记] CS131 Computer Vision: Foundations and Applications:Lecture 9 深度学习2的相关文章

[学习笔记] CS131 Computer Vision: Foundations and Applications:Lecture 4 像素和滤波器

Background reading: Forsyth and Ponce, Computer Vision Chapter 7 Image sampling and quantization Types of images: binary, gray scale, color Resolution: DPI: dots per inch, spatial pixel density Image histograms: histogram of an image provides the fre

STM32学习笔记——USART串口(向原子哥和火哥学习)

一.USART简介 通用同步异步收发器(USART)提供了一种灵活的方法与使用工业标准NRZ异步串行数据格式的外部设备之间进行全双工数据交换.USART利用分数波特率发生器提供宽范围的波特率选择. STM32 的串口资源相当丰富的,功能也相当强劲.STM32F103ZET6 最多可提供 5 路串口,有分数波特率发生器,支持同步单向通信和半双工单线通信,支持LIN(局部互连网),智能卡协议和IrDA(红外数据组织)SIR ENDEC规范,以及调制解调器(CTS/RTS)操作.它还允许多处理器通信.

Android学习笔记——文件路径(/mnt/sdcard/...)、Uri(content://media/external/...)学习

一.URI 通用资源标志符(Universal Resource Identifier, 简称"URI"). Uri代表要操作的数据,Android上可用的每种资源 - 图像.视频片段等都可以用Uri来表示. URI一般由三部分组成: 访问资源的命名机制. 存放资源的主机名. 资源自身的名称,由路径表示. Android的Uri由以下三部分组成: "content://".数据的路径.标示ID(可选) 举些例子,如: 所有联系人的Uri: content://con

百度DMLC分布式深度机器学习开源项目(简称“深盟”)上线了如xgboost(速度快效果好的Boosting模型)、CXXNET(极致的C++深度学习库)、Minerva(高效灵活的并行深度学习引擎)以及Parameter Server(一小时训练600T数据)等产品,在语音识别、OCR识别、人脸识别以及计算效率提升上发布了多个成熟产品。

百度为何开源深度机器学习平台? 有一系列领先优势的百度却选择开源其深度机器学习平台,为何交底自己的核心技术?深思之下,却是在面对业界无奈时的远见之举. 5月20日,百度在github上开源了其深度机器学习平台.此番发布的深度机器学习开源平台属于"深盟"的开源组织,其核心开发者来自百度深度学习研究院(IDL),微软亚洲研究院.华盛顿大学.纽约大学.香港科技大学,卡耐基·梅陇大学等知名公司和高校. 通过这一开源平台,世界各地的开发者们可以免费获得更优质和更容易使用的分布式机器学习算法源码,

java学习笔记-0918班亓东*:尚硅谷班级的学习气氛浓厚

转眼来北京java培训一个月过去,在这一个月里我认识了很多老师和同学,让我有种重新回到学生时代上学时感觉,这的学习气氛很好,每个同学都很用功的学习,有些基础好的同学也很愿意帮助别人,同学之间的相互关心,相互帮助,解决了很多学习上的问题,说起来还要感谢我的同桌,没有他的帮助我想我也会学的很吃力,每次遇到问题他都能及时帮同学解决,同学之间的彼此帮助,让班级的学习气氛更加浓厚,营造一个良好的学习氛围,对于将来的学习会大有帮助,好的学习氛围是大家共同努力的成果. 在尚硅谷java培训每个阶段都会有阶段性

UFLDL教程笔记及练习答案四(建立分类用深度学习)

此次主要由自我学习过度到深度学习,简单记录如下: (1)深度学习比浅层网络学习对特征具有更优异的表达能力和紧密简洁的表达了比浅层网络大的多的函数集合. (2)将传统的浅层神经网络进行扩展会存在数据获取.局部最值和梯度弥散的缺点. (3)栈式自编码神经网络是由多层稀疏自编码器构成的神经网络(最后一层采用的softmax回归或者logistic回归分类),采用逐层贪婪的训练方法得到初始的参数,这样在数据获取方面就可以充分利用无标签的数据.通过逐层贪婪的训练方法又称为预训练,然后可以使用有标签的数据集

对比学习资料《深度学习入门:基于Python的理论与实现》+《深度学习原理与实践》+《深度学习理论与实战基础篇》电子资料

我认为<深度学习入门:基于Python的理论与实现>电子书代码基本原理,简洁清楚,所用数学基本就是大一水平,适合做第一本入门书. <深度学习入门:基于Python的理论与实现>书代码实现的神经网络全是用numpy,超级适合刚刚入门想了解神经网络,数学不好看公式看不懂的人,看完这本基本就懂深度学习是弄啥子了. 如果连这本都看不懂的话,可以去看<Python神经网络编程>.我个人认为这两本书是最简单直白的方式让人理解神经网络的了. <深度学习原理与实践>电子书代

Deep Learning 十_深度学习UFLDL教程:Convolution and Pooling_exercise(斯坦福大学深度学习教程)

前言 理论知识:UFLDL教程和http://www.cnblogs.com/tornadomeet/archive/2013/04/09/3009830.html 实验环境:win7, matlab2015b,16G内存,2T机械硬盘 实验内容:Exercise:Convolution and Pooling.从2000张64*64的RGB图片(它是the STL10 Dataset的一个子集)中提取特征作为训练数据集,训练softmax分类器,然后从3200张64*64的RGB图片(它是th

《深度学习与计算机视觉算法原理框架应用》PDF+《深度学习之PyTorch实战计算机视觉》PDF

下载:https://pan.baidu.com/s/12-s95JrHek82tLRk3UQO_w <深度学习与计算机视觉 算法原理.框架应用>PDF,带书签,347页.<大数据架构详解:从数据获取到深度学习>PDF,带书签,373页.配套源代码. 下载:https://pan.baidu.com/s/1P0-o29x0ZrXp8WotN7GzcA<深度学习之PyTorch实战计算机视觉> 更多分享:https://pan.baidu.com/s/1g4hv05UZ_