MobileNets: Open-Source Models for Efficient On-Device Vision

https://research.googleblog.com/2017/06/mobilenets-open-source-models-for.html

Wednesday, June 14, 2017

Posted by Andrew G. Howard, Senior Software Engineer and Menglong Zhu, Software Engineer

(Cross-posted on the Google Open Source Blog)

Deep learning has fueled tremendous progress in the field of computer
vision in recent years, with neural networks repeatedly pushing the frontier of visual recognition technology.
While many of those technologies such as object, landmark, logo and
text recognition are provided for internet-connected devices through the
Cloud Vision API, we
believe that the ever-increasing computational power of mobile devices
can enable the delivery of these technologies into the hands of our
users, anytime, anywhere, regardless of internet connection. However,
visual recognition for on device and embedded applications poses many
challenges — models must run quickly with high accuracy in a
resource-constrained environment making use of limited computation,
power and space.

Today we are pleased to announce the release of MobileNets, a family of mobile-first computer vision models for TensorFlow,
designed to effectively maximize accuracy while being mindful of the
restricted resources for an on-device or embedded application.
MobileNets are small, low-latency, low-power models parameterized to
meet the resource constraints of a variety of use cases. They can be
built upon for classification, detection, embeddings and segmentation
similar to how other popular large scale models, such as Inception, are used.

Example use cases include detection, fine-grain classification, attributes and geo-localization.

This release contains the model definition for MobileNets in TensorFlow using TF-Slim, as well as 16 pre-trained ImageNet classification checkpoints for use in mobile projects of all sizes. The models can be run efficiently on mobile devices with TensorFlow Mobile.


Model Checkpoint

Million MACs

Million Parameters

Top-1 Accuracy

Top-5 Accuracy

MobileNet_v1_1.0_224

569

4.24

70.7

89.5

MobileNet_v1_1.0_192

418

4.24

69.3

88.9

MobileNet_v1_1.0_160

291

4.24

67.2

87.5

MobileNet_v1_1.0_128

186

4.24

64.1

85.3

MobileNet_v1_0.75_224

317

2.59

68.4

88.2

MobileNet_v1_0.75_192

233

2.59

67.4

87.3

MobileNet_v1_0.75_160

162

2.59

65.2

86.1

MobileNet_v1_0.75_128

104

2.59

61.8

83.6

MobileNet_v1_0.50_224

150

1.34

64.0

85.4

MobileNet_v1_0.50_192

110

1.34

62.1

84.0

MobileNet_v1_0.50_160

77

1.34

59.9

82.5

MobileNet_v1_0.50_128

49

1.34

56.2

79.6

MobileNet_v1_0.25_224

41

0.47

50.6

75.0

MobileNet_v1_0.25_192

34

0.47

49.0

73.6

MobileNet_v1_0.25_160

21

0.47

46.0

70.7

MobileNet_v1_0.25_128

14

0.47

41.3

66.2
Choose the right
MobileNet model to fit your latency and size budget. The size of the
network in memory and on disk is proportional to the number of
parameters. The latency and power usage of the network scales with the
number of Multiply-Accumulates (MACs) which measures the number of fused
Multiplication and Addition operations. Top-1 and Top-5 accuracies are
measured on the ILSVRC dataset.

We are excited to share MobileNets with the open-source community. Information for getting started can be found at the TensorFlow-Slim Image Classification Library. To learn how to run models on-device please go to TensorFlow Mobile. You can read more about the technical details of MobileNets in our paper, MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.

Acknowledgements

MobileNets were made possible with the hard work of many engineers and
researchers throughout Google. Specifically we would like to thank:

Core Contributors: Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, Hartwig Adam

Special thanks to: Benoit Jacob, Skirmantas Kligys, George
Papandreou, Liang-Chieh Chen, Derek Chow, Sergio Guadarrama, Jonathan
Huang, Andre Hentz, Pete Warden

时间: 2024-08-29 17:11:35

MobileNets: Open-Source Models for Efficient On-Device Vision的相关文章

Dynamic device virtualization

A system and method for providing dynamic device virtualization is herein disclosed. According to one embodiment, the computer-implemented method includes providing a hypervisor and one or more guest virtual machines (VMs). Each guest VM is disposed

EPPB also support BlackBerry device

各位看倌不是小弟要賣弄英文,實在是外國朋友希望知道上一篇"雲取證"中所用的工具Elcomsoft Phone Password Breaker支援黑莓機否?又要求非要看到截屏才算數..因此以下就是補充說明: 1.Elcomsoft Phone Password Breaker-> Click [Choose source] 2.It support BlackBerry device Wish you success~

models常用字段及参数,orm查询优化, choices参数及实际应用

一.自定义字段类型 from django.db import models # Create your models here. # 自定义CHAR类型 from django.db.models import Field class RealCharField(Field): # 这里是要继承Field类的 def __init__(self, max_length, *args, **kwargs): self.max_length = max_length # 拦截一个父类的方法 操作完

斯坦福CS课程列表

http://exploredegrees.stanford.edu/coursedescriptions/cs/ CS 101. Introduction to Computing Principles. 3-5 Units. Introduces the essential ideas of computing: data representation, algorithms, programming "code", computer hardware, networking, s

spice model parameter

===================BSIM4.3.0 Model Selectors/Controllers============ LEVEL SPICE3 model selector VERSION Model version BINUNIT Binning unitr PARAMCHK Switch for parameter value check MOBMOD Mobility model RDSMOD Bias-dependent source/drain resistance

Python之路,Day15 - Django适当进阶篇

Python之路,Day15 - Django适当进阶篇 本节内容 学员管理系统练习 Django ORM操作进阶 用户认证 Django练习小项目:学员管理系统设计开发 带着项目需求学习是最有趣和效率最高的,今天就来基于下面的需求来继续学习Django 项目需求: 1.分讲师\学员\课程顾问角色,2.学员可以属于多个班级,学员成绩按课程分别统计3.每个班级至少包含一个或多个讲师4.一个学员要有状态转化的过程 ,比如未报名前,报名后,毕业老学员5.客户要有咨询纪录, 后续的定期跟踪纪录也要保存6

卷积神经网络和CIFAR-10:Yann LeCun专访 Convolutional Nets and CIFAR-10: An Interview with Yann LeCun

Recently Kaggle hosted a competition on the CIFAR-10 dataset. The CIFAR-10 dataset consists of 60k 32x32 colour images in 10 classes. This dataset was collected by AlexKrizhevsky, Vinod Nair, and Geoffrey Hinton. Many contestants used convolutional n

数据集的使用

Participate in Reproducible Research General Image Processing OpenCV (C/C++ code, BSD lic) Image manipulation, matrix manipulation, transforms Torch3Vision (C/C++ code, BSD lic) Basic image processing, matrix manipulation and feature extraction algor

Computer Vision Algorithm Implementations

Participate in Reproducible Research General Image Processing OpenCV (C/C++ code, BSD lic) Image manipulation, matrix manipulation, transforms Torch3Vision (C/C++ code, BSD lic) Basic image processing, matrix manipulation and feature extraction algor