博客
关于我
强烈建议你试试无所不能的chatGPT,快点击我
(转) Deep learning architecture diagrams
阅读量:6303 次
发布时间:2019-06-22

本文共 3874 字,大约阅读时间需要 12 分钟。

Machine learning made easy

Deep learning architecture diagrams

2016-09-30

Like a wild stream after a wet season in African savanna diverges into many smaller streams forming lakes and puddles, deep learning has diverged into a myriad of specialized architectures. Each architecture has a diagram. Here are some of them.

 

 

 

Neural networks are conceptually simple, and that’s their beauty. A bunch of homogenous, uniform units, arranged in layers, weighted connections between them, and that’s all. At least in theory. Practice turned out to be a bit different. Instead of feature engineering, we now have , as described by Stephen Merrity:

The romanticized description of deep learning usually promises that the days of hand crafted feature engineering are gone - that the models are advanced enough to work this out themselves. Like most advertising, this is simultaneously true and misleading.

Whilst deep learning has simplified feature engineering in many cases, it certainly hasn’t removed it. As feature engineering has decreased, the architectures of the machine learning models themselves have become increasingly more complex. Most of the time, these model architectures are as specific to a given task as feature engineering used to be.

To clarify, this is still an important step. Architecture engineering is more general than feature engineering and provides many new opportunities. Having said that, however, we shouldn’t be oblivious to the fact that where we are is still far from where we intended to be.

Not quite as bad as doings of , but not too good either.

Kardash chair

An example of architecture specific to a given task

LSTM diagrams

How to explain those architectures? Naturally, with a diagram. A diagram will make it all crystal clear.

Let’s first inspect the two most popular types of networks these days, CNN and LSTM. You’ve already seen a , so turning to the iconic LSTM:

LSTM

It’s easy, just take a closer look:

LSTM

As they say, in mathematics you don’t understand things, you just get used to them.

Fortunately, there are good explanations, for example  and .

LSTM still too complex? Let’s try a simplified version, GRU (Gated Recurrent Unit). Trivial, really.

LSTM

Especially this one, called minimal GRU.

Minimal GRU

More diagrams

Various modifications of LSTM are now common. Here’s one, called deep bidirectional LSTM:

DB-LSTM

DB-LSTM, 

DB-LSTM

The rest are pretty self-explanatory, too. Let’s start with a combination of CNN and LSTM, since you have both under your belt now:

CRMN

Convolutional Residual Memory Network, 

DNTM

Dynamic NTM, 

ENTM

Evolvable Neural Turing Machines, 

Recurrent model of visual attention

Recurrent Model Of Visual Attention, 

Unsupervised domain adaptation by backpropagation

Unsupervised Domain Adaptation By Backpropagation, 

Deeply Recursive CNN for super resolution

Deeply Recursive CNN For Image Super-Resolution, 

This diagram of multilayer perceptron with synthetic gradients scores high on clarity:

Synthetic gradients

MLP with synthetic gradients, 

Drinking cat

Every day brings more. Here’s a fresh one, again from Google:

Google's Neural Machine Translation System

Google’s Neural Machine Translation System, 

And Now for Something Completely Different

Drawings from the  are pleasantly simple, but, unfortunately, serve mostly as eye candy. For example:

  

ESM, ESN and ELM

These look like not-fully-connected perceptrons, but are supposed to represent a Liquid State Machine, an Echo State Network, and an Extreme Learning Machine.

How does LSM differ from ESN? That’s easy, it has green neuron with triangles. But how does ESN differ from ELM? Both have blue neurons.

Seriously, while similar, ESN is a recursive network and ELM is not. And this kind of thing should probably be visible in an architecture diagram.

Comments

 

Recent Posts

Twitter

Follow  for notifications about new posts.

  • Status updating...
 
Also check out   for things related to machine learning and data science in general.

GitHub

Most articles come with some . We push it to Github.

Cubert

Visualize your data in interactive 3D, as described .

Copyright © 2016 - Zygmunt Z. - Powered by 

转载地址:http://qcfxa.baihongyu.com/

你可能感兴趣的文章
hdu 5471(状压DP or 容斥)
查看>>
oracle.jdbc.driver.OracleDriver和oracle.jdbc.OracleDriver这两个驱动的区别
查看>>
NSQ部署
查看>>
git常用命令记录
查看>>
IBM发布新一代云计算工具包MobileFirst Foundation
查看>>
唯品会HDFS性能挑战和优化实践
查看>>
大规模学习该如何权衡得失?解读NeurIPS 2018时间检验奖获奖论文
查看>>
大厂前端高频面试问题与答案精选
查看>>
我们用5分钟写了一个跨多端项目
查看>>
Visual Studio 15.4发布,新增多平台支持
查看>>
有赞透明多级缓存解决方案(TMC)设计思路
查看>>
如何设计高扩展的在线网页制作平台
查看>>
Git 2.5增加了工作树、改进了三角工作流、性能等诸多方面
查看>>
Swift 5将强制执行内存独占访问
查看>>
中台之上(二):为什么业务架构存在20多年,技术人员还觉得它有点虚?
查看>>
深度揭秘腾讯云低功耗广域物联网LPWAN 技术及应用
查看>>
与Jeff Sutherland谈敏捷领导力
查看>>
More than React(四)HTML也可以静态编译?
查看>>
React Native最佳学习模版- F8 App开源了
查看>>
云服务正在吞噬世界!
查看>>