title
stringlengths 12
112
| published
stringlengths 19
23
| url
stringlengths 28
28
| video_id
stringlengths 11
11
| channel_id
stringclasses 5
values | id
stringlengths 16
31
| text
stringlengths 0
596
| start
float64 0
37.8k
| end
float64 2.18
37.8k
|
---|---|---|---|---|---|---|---|---|
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3005.7400000000002
|
any old information might come in, and we might collapse and or we might never reach
| 3,005.74 | 3,016.34 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3010.8
|
consensus because any old information might come in. However, if we introduce the attention
| 3,010.8 | 3,022.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3016.34
|
mechanism into this whole thing, and only draw in information from the selected neighbors
| 3,016.34 | 3,028.68 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3022.1200000000003
|
that already are in the same group in the same island as me, then this consensus algorithm
| 3,022.12 | 3,034.44 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3028.6800000000003
|
works. So if the network, the network is now forced kind of to learn to build these islands
| 3,028.68 | 3,042.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3034.44
|
of similar things in order to make this consensus work if we regularize this consensus. So I
| 3,034.44 | 3,049.92 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3042.6
|
believe he makes the case for the attention mechanism. I don't think he, in this case,
| 3,042.6 | 3,056.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3049.92
|
considers kind of the up the next up layer islands, what I would say is you need to consider
| 3,049.92 | 3,066.36 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3056.88
|
the island membership all the way up the columns in order to decide which things which locations,
| 3,056.88 | 3,073 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3066.36
|
right, it's free to choose which embeddings at other locations it should resemble. I think,
| 3,066.36 | 3,084.2 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3073.0
|
yeah, this is the case for the attention mechanism. Okay, I hope you're still half with me. If
| 3,073 | 3,090.66 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3084.2
|
not, I'm, I'm bit confused too. But I think what he's doing is he says, contrastive learning
| 3,084.2 | 3,096.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3090.66
|
would be good, you can use it, but you have to be careful at which layer you do it. Another
| 3,090.66 | 3,104.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3096.24
|
regularizer to form these islands would be this regularize the network to conform to
| 3,096.24 | 3,110.96 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3104.72
|
the consensus option, opinion. However, if you simply aggregate information from the
| 3,104.72 | 3,118.04 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3110.96
|
same layer, then that wouldn't work because, you know, the different things in the same
| 3,110.96 | 3,123.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3118.04
|
layer might correspond to completely different parts of the image. Drawing in information
| 3,118.04 | 3,128.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3123.6
|
from there would not help you. How do you solve this by introducing the very attention
| 3,123.6 | 3,134.98 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3128.52
|
mechanism that he introduced in order to only draw in information from parts of the same
| 3,128.52 | 3,144.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3134.98
|
layer that actually are related to you? Okay, the next thing, the next consideration he
| 3,134.98 | 3,151.06 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3144.8
|
does is representing coordinate transformations. How does this represent coordinate transformations,
| 3,144.8 | 3,158.14 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3151.06
|
there was a capsule net paper where he explicitly represents coordinate transformations in kind
| 3,151.06 | 3,166.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3158.14
|
of four dimension quaternion space. And he says, that is probably not needed, because
| 3,158.14 | 3,176.92 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3166.3199999999997
|
you don't want to hear says you could represent this by a by four by four matrices. However,
| 3,166.32 | 3,182.68 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3176.92
|
if you simply allocate 16 numbers in each embedding vector, in order to represent the
| 3,176.92 | 3,187.04 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3182.68
|
part whole coordinate transformation, like the transformation that relates the part to
| 3,182.68 | 3,192.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3187.04
|
the whole, that does not make it easy to represent uncertainty about the aspects of pose and
| 3,187.04 | 3,198.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3192.52
|
certainty about others. So the problem here is that we know that humans, when they watch
| 3,192.52 | 3,206.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3198.8
|
something right here, when they watch a scene, like this is a chair, and there is a person,
| 3,198.8 | 3,212.44 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3206.4
|
a very tiny person on the chair, we don't see necessarily the coordinate frame of the
| 3,206.4 | 3,217.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3212.44
|
world, what we see is we see the coordinate frame of the chair, like maybe this is the
| 3,212.44 | 3,225.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3217.52
|
center, and we see the person in relation to the chair, our brain seems to do this intuitively,
| 3,217.52 | 3,229.98 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3225.16
|
and hinting things that a system like this should also do it intuitively. So somehow,
| 3,225.16 | 3,234.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3229.98
|
the coordinate transformations involved going from the eye to the reference through the
| 3,229.98 | 3,241.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3234.84
|
frame of the chair, and then from the chair to the person, they should be somehow in encoded
| 3,234.84 | 3,248.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3241.72
|
in this network. However, he also says that it's probably not necessary to encode them
| 3,241.72 | 3,252.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3248.3199999999997
|
explicitly as you know, explicit coordinate transformations, because not only does that
| 3,248.32 | 3,259.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3252.8399999999997
|
make it harder, probably to learn, but also, you can't represent uncertainty. In fact,
| 3,252.84 | 3,264.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3259.7999999999997
|
you can represent uncertainty, that's the next thing right here, much better by having
| 3,259.8 | 3,272.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3264.88
|
a higher dimensional thing that you're trying to guess, right? If you are trying to guess
| 3,264.88 | 3,277.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3272.2400000000002
|
a distribution with three components, and you simply have a three dimensional vector,
| 3,272.24 | 3,283.5 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3277.76
|
you have no way of representing uncertainty. However, if you have a nine dimensional vector,
| 3,277.76 | 3,290.62 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3283.5
|
you can have three opinions about the distribution. So this is an opinion, this is an opinion,
| 3,283.5 | 3,295.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3290.62
|
and then this is an opinion. And then you can sort of aggregate and you can say, Well,
| 3,290.62 | 3,301.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3295.88
|
I'm pretty sure about these two things, because all my opinions are pretty close. But this
| 3,295.88 | 3,309.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3301.12
|
one here, I'm not so sure because my individual things say different things, things say things.
| 3,301.12 | 3,315.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3309.3199999999997
|
All right, I've this video is too long. So that's his argument right here, we don't need
| 3,309.32 | 3,322.78 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3315.48
|
explicit representing of uncertainty, because by simply over parameterizing, we can already
| 3,315.48 | 3,331.98 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3322.78
|
represent uncertainty well. And we also don't need disentangled position information and,
| 3,322.78 | 3,341.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3331.98
|
and so on. Sorry, we don't need different position informations, because, again, the
| 3,331.98 | 3,346.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3341.52
|
work can take care of that. And he gives a good example, like why would you have disentangled
| 3,341.52 | 3,357.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3346.88
|
coordinate frame if you have an image? And in the image, the picture in it is this. How
| 3,346.88 | 3,367.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3357.48
|
do you know if that is a rhomboid shape? Or if it is a rec, if it is a rectangular piece
| 3,357.48 | 3,373.64 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3367.12
|
of paper viewed from the side, I should probably draw it way closer, something like something
| 3,367.12 | 3,382.08 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3373.64
|
like this. I suck at this. You get probably get what I mean. Like, if it is a different
| 3,373.64 | 3,389.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3382.08
|
object, it has a like the object and the coordinate transformation are dependent upon each other.
| 3,382.08 | 3,394.9 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3389.4
|
And so it makes sense for the neural network to actually entangle the two, because the
| 3,389.4 | 3,401.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3394.9
|
two things depend on each other. In essence, he's just saying, don't worry about explicitly
| 3,394.9 | 3,407.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3401.52
|
representing all of the different things. We got it like the neural network can do all
| 3,401.52 | 3,415.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3407.7200000000003
|
of these things, like uncertainty or position, and post transformations. So here he compares
| 3,407.72 | 3,425.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3415.12
|
it to different other architectures. comparison to CNN comparison to transformers comparison
| 3,415.12 | 3,430.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3425.4
|
to capsule models. And at the end, it goes into video at the very beginning, he says,
| 3,425.4 | 3,437.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3430.88
|
the paper is about actually a video system. And you can kind of see that because we go
| 3,430.88 | 3,443.98 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3437.48
|
through this algorithm in multiple time steps, right? You have, it's like you analyze an
| 3,437.48 | 3,451.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3443.98
|
image with these columns, which gives you sort of a 3d 3d tensor with the image at the
| 3,443.98 | 3,458.04 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3451.16
|
bottom. And you go in the next time step, you have a new 3d tensor, right, you pass
| 3,451.16 | 3,464.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3458.04
|
this whole information around with the image at the bottom. And it says, well, why does
| 3,458.04 | 3,469.3 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3464.76
|
that need to be the same image that could also be different images. So you could use
| 3,464.76 | 3,476.26 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3469.3
|
the system to analyze video. So what he does is he says, at the same time, you do this
| 3,469.3 | 3,482.58 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3476.26
|
time step to find agreement, you could actually swap out the video frame, the X, you can swap
| 3,476.26 | 3,487.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3482.5800000000004
|
out the video frame, and produce a slightly different video frame. And you could actually
| 3,482.58 | 3,493.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3487.1200000000003
|
have a kind of an ensemble regularizing effect. So as the whole columns here, the whole system
| 3,487.12 | 3,499.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3493.24
|
comes to a consensus over time, you feed in different information at the bottom. And what
| 3,493.24 | 3,507.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3499.72
|
he says is that, you know, if this is a slow enough video, then the top layers here would
| 3,499.72 | 3,513.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3507.2799999999997
|
probably could still reach an agreement, while the bottom layers would change rapidly. But
| 3,507.28 | 3,521 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3513.2799999999997
|
that could be sort of an ensemble or a regularizer, regularizing effect that it even has. So he
| 3,513.28 | 3,526.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3521.0
|
intrinsically connects these two time dimensions, because they would be separate, right, you
| 3,521 | 3,533.64 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3526.48
|
could input a video. And then in, you know, in each frame, you could do this consensus
| 3,526.48 | 3,539.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3533.64
|
finding algorithm. But he says, No, it's actually cool to consider them together to do the consensus
| 3,533.64 | 3,545.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3539.32
|
finding while you sort of watch the video, it's just not clear that you always need the
| 3,539.32 | 3,550.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3545.24
|
same amount of consensus finding steps as you need as you have video frames. So maybe,
| 3,545.24 | 3,556 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3550.7999999999997
|
you want to, maybe you want to take like five consensus steps per video frame, or the other
| 3,550.8 | 3,564.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3556.0
|
way around? Not sure. In any case, I think that's a pretty cool idea. And he says things
| 3,556 | 3,569.64 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3564.16
|
like, if the changes are rapid, there is no time available to iteratively settle on a
| 3,564.16 | 3,574.64 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3569.64
|
good set of embedding vectors for interpreting a specific frame. This means that the glom
| 3,569.64 | 3,580.56 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3574.64
|
architecture cannot correctly interpret complicated shapes. If the images are changing rapidly,
| 3,574.64 | 3,585.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3580.56
|
try taking an irregularly shaped potato and throwing it up in the air such a way that
| 3,580.56 | 3,590.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3585.24
|
it rotates at one or two cycles per second. Even if you smoothly track the potato, you
| 3,585.24 | 3,596.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3590.7599999999998
|
cannot see what shape it is. Now I don't have a potato, but I can give you an avocado. So
| 3,590.76 | 3,613.2 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3596.88
|
if you give me a second, how's that? Could you track the shape? I don't know. Probably
| 3,596.88 | 3,621.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3613.2000000000003
|
in his correct. All right, he talks about is this biologically plausible? And I don't
| 3,613.2 | 3,626.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3621.8
|
want to go too much into this. He discusses some restrictions like yeah, we still use
| 3,621.8 | 3,632.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3626.6
|
backprop and is backprop plausible and so on. I love this sentence. In the long run,
| 3,626.6 | 3,638.68 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3632.4
|
however, we are all dead. And then the footnote saying there are alternative facts. But yeah,
| 3,632.4 | 3,645.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3638.68
|
he discusses whether it's biological plausible. How could you modify it to make it more plausible?
| 3,638.68 | 3,652.64 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3645.48
|
For example, when you want to do contrastive learning, there is evidence that dreams during
| 3,645.48 | 3,657.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3652.64
|
so during sleep, you do contrastive learning, like you produce the negative examples during
| 3,652.64 | 3,665 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3657.3199999999997
|
sleep, and then during the day, you collect the positive examples and so on. So I think
| 3,657.32 | 3,673.08 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3665.0
|
this is a more speculative part of the paper, but it's pretty cool to it's pretty cool to
| 3,665 | 3,680.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t3673.08
|
read it. And lastly, he goes into discussion. He also says that this paper is too long already.
| 3,673.08 | 3,686.44 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.