anhinga_drafts: (Default)
[personal profile] anhinga_drafts
Stanford Colloquium on Computer Systems: a talk by Geoffrey Hinton, Apr. 27, 2016:

https://www.youtube.com/watch?v=VIRCybGgHts


Some of the highlights of this remarkable lecture:

1) Hinton argues for the networks with many orders of magnitude more parameters than data points and with strong regularization.

(Cf. recent paper on "Outrageously Large Neural Networks", https://arxiv.org/abs/1701.06538

This has 100 billion parameters, so only 3 orders of magnitude below the size Hinton suggests for the brain.)

2) He specifically explains how "dropout" works as regularization (one considers a huge ensemble of models, then for each step of the gradient descent, one samples from that ensemble by picking one model, regularization happens via sharing weights between models from the ensemble, the resulting models approximates the geometric mean of all the models in the ensemble, hence it is related to the "product of experts" scheme).

This is the most immediate practically applicable part of the talk.

He makes a strong case that spikes do something similar to this dropout regularization.

3) He conjectures that the derivative of the error with respect to the neuron input is coded by the derivative of the output signal of the neuron with respect to time (while the derivative of the value detected by that neuron with respect to time is coded by some other neuron, if that derivative is needed).

Then, strangely enough, backpropagation becomes implemented via Hebbian learning and moreover, the spike-time-dependent plasticity rule naturally emerges from looking at the derivative filter over the spike train.

This is the most revolutionary part of the talk.

4) The problem of the lack of symmetric reverse connections in the brain is solved via recent work on "feedback alignment" (where one can use random weights instead of actual weights while computing the derivative by the chain rule). He explains the intuition behind the "feedback alignment" and why it works, but at the same time he says that the initial hopes that it would be better than backprop due to capturing second-derivative information do not seem to be justified, and that the classic backprop seems to be slightly better, rather than the opposite. (The original article on "feedback alignment" contianing those initial hopes is https://arxiv.org/abs/1411.0247 ; it's still a nice method, even if we now don't think it's better than backprop.)

***

The talk itself is 65 min, then questions and answers. I found it useful to watch in some increments of a few minutes each, one thought at a time, rewinding often.

Date: 2017-02-05 10:33 am (UTC)
From: [identity profile] datjko.livejournal.com
Миша, я еще не смотрел, но спасибо, что кинул ссылку и спасибо за highlights (и спасибо за рейтинг "remarkable" - я все еще всему подряд ставлю 6 звезд из пяти и это приводит только к раздуванию числа незакрытых гештальтов и окон в браузере).

Date: 2017-03-27 04:12 am (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
Привет!

Я тут, наконец, понял, почему загадочное правило Хинтона работает:

http://www.cs.brandeis.edu/~bukatin/understanding-hinton-rule.html

Date: 2017-03-27 06:19 am (UTC)
From: [identity profile] datjko.livejournal.com
Миша, привет!
спасибо, что делишься. мне пока, увы, из моего забоя ничего не понятно, Хинтона я так и не досмотрел. wtf is "the derivative of a neuron output with respect to time" and Hebbian learning я когда-нибудь узнаю, я почти созрел, но тут приходится ударно низковисящие груши околачивать, а там тоже для меня много нового.

а ты видел https://arxiv.org/abs/1702.02181 "Deep Learning with Dynamic Computation Graphs"? Это может помочь твоей имплементации? Я очень надеюсь в скором времени с поиграть с этим фолдом.

Date: 2017-03-28 10:32 am (UTC)
From: [identity profile] datjko.livejournal.com
мой ответ помечен как спам. наверно за ссылку на фолд. тебе его видно?

Date: 2017-04-04 10:03 pm (UTC)
From: [identity profile] anhinga-anhinga.livejournal.com
Нет, я только вижу те же два, которые и тебе должны быть видны (Feb 5 and March 28).

(Я сейчас попробую послать мой е-майл в личку, на всякий случай, а то comment notifications практически перестали ходить в последние дней 10.)

Profile

anhinga_drafts: (Default)
anhinga_drafts

June 2022

S M T W T F S
   1234
5678 91011
12131415161718
19202122232425
2627282930  

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Jan. 1st, 2026 08:12 pm
Powered by Dreamwidth Studios