Тёмный

'Eigen-view' on Grid Cells - September 21, 2020 

Numenta
Подписаться 15 тыс.
Просмотров 1 тыс.
50% 1

In this week’s research meeting, Marcus Lewis presents his ‘Eigen-view’ on grid cells and connects ideas in 3 underlying papers to Numenta’s research. He discusses the mapping of grid cells in terms of eigenvectors, and evaluates eigenvectors in terms of the Fourier transform (space) and the non-Fourier transform, called “spectral graph theory” (2D graph).
Marcus’s whiteboard presentation: miro.com/app/board/o9J_klMK6P0=/
Papers referenced:
► "Prediction with directed transitions: complex eigen structure, grid cells and phase coding" by Changmin Yu, Timothy E.J. Behrens, Neil Burgess - arxiv.org/abs/2006.03355
► “The hippocampus as predictive map” by Kimberly L. Stachenfeld, Matthew M. Botvinick, Samuel J. Gershman - www.biorxiv.org/content/10.11...
► “Grid Cells Encode Local Positional Information” by Revekka Ismakov, Omri Barak, Kate Jeffery, Dori Derdikman - www.cell.com/current-biology/...
- - - - -
Numenta is leading the new era of machine intelligence. Our deep experience in theoretical neuroscience research has led to tremendous discoveries on how the brain works. We have developed a cohesive theory, core software technology, and numerous software applications all based on principles of the neocortex. Our innovative work delivers breakthrough capabilities and demonstrates that a computing approach based on biological learning principles can do things that today’s programmed computers cannot do.
Subscribe to our News Digest for the latest news about neuroscience and artificial intelligence:
tinyurl.com/NumentaNewsDigest
Subscribe to our Newsletter for the latest Numenta updates:
tinyurl.com/NumentaNewsletter
Our Social Media:
/ numenta
/ officialnumenta
/ numenta
Our Open Source Resources:
github.com/numenta
discourse.numenta.org/
Our Website:
numenta.com/

Наука

Опубликовано:

 

5 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 8   
@ChrisMcAce
@ChrisMcAce 3 года назад
26:00 ...it should be noted that this happens because the random walk is isotropic, aka rotationally symmetric; the same in all directions. That seems reasonable for the allocentric description of the local (x,y) environment of a terrestrial animal. One wouldn't want to make this assumption for the vertical direction, where gravity plays a role.
@drdca8263
@drdca8263 Месяц назад
If we had a random walk where at small times $\Delta t$ the motion was described by a multivariate normal distribution associated with the matrix $(\Delta t) \Sigma(x)$ (where x is the current location and \Sigma(x) is a positive definite matrix that is a function of position), Then, I think we could replace the Laplacian with $\tr(\Sigma(x) H)$ where H is the Hessian (or maybe that should be $\tr(\Sigma(x)^{-1} H)$? I may have gotten mixed up.) (The usual Laplacian is obtained when \Sigma(x) is the identity matrix for all x) I don’t know how the random walk should be described in the case of flight. It seems to me that with flight, momentum would be a bigger deal than it is when walking around on the ground. Maybe one would have to move to a phase space of position and momentum together? Or... idk. I don’t know what bats actually do when exploring. But, if moving vertically were just more costly, both up and down (... which doesn’t seem right...), or more generally if there’s a bias to move along some axiis over others, and where which axiis are preferred can vary in space, then I think the differential operator I described above should work to describe the diffusion of a probability distribution over position, and, seems to me like it could *maybe* result in some nice distorted versions of the usual eigenfunctions of the Laplacian? Like, if you apply some coordinate transform, and have the image of the uniform random walk under this transform, and if the resulting random walk in these new coordinates be described in terms of such a \Sigma(x) , then... uhh... I think the transformation might sorta end up relating the different eigenfunctions? ... but maybe not quite?
@ChrisMcAce
@ChrisMcAce Месяц назад
@@drdca8263 As I remember it, my comment was about the observation that the eigenvalues (!) are the same for all eigenfunctions (cosines; well, technically we should be talking about complex exponentials, but I'll be sloppy) with the same frequency, no matter the direction of the cosine. This requires isotropy. I believe the sinusoids will always be eigenfunctions, no matter what differential operator / diffusion / convolution you choose. At least so long as we're talking about Euclidean space and not something more general like a graph. When looking at their discretized versions, both difference operators and diffusions ARE convolutions. Since all convolutions commute, they must share eigenvectors. In continuous spaces the rigorous math might be more interesting; a quick websearch turns up the book "Convolution-like Structures, Differential Operators and Diffusion Processes" by Sousa et al. I think you are right in focusing on the _actions_ animals take. And on what the relevant state space is. At the end of the day, the actions determine the operator that would benefit from diagonalization. E.g. a human might model the effect of taking a step forward (a common action for a human) as a convolution of their current position estimate with a Gaussian centered around a point that's in the direction they are facing and a "mean step-size" away from the origin. Since it's a convolution, sinusoids are again eigenfunctions, but the set of eigenvalues representing this "step-action" is different from those that represent, say, a uniform random walk. If the idea is that the entorhinal cortex and the hippocampus implement a kind of SLAM algorithm ( en.wikipedia.org/wiki/Simultaneous_localization_and_mapping ), then by measuring the grid cells' activity we're likely taking a peek at the odometry. If the brain uses a spectral representation for the animal's location-estimate, updates due to an action are just a per-grid-cell multiplication of its activity, rather than e.g. a convolution. The interesting question then becomes: what part of this is biologically hardcoded and what part can be learned? Perhaps the cortical neurons do run something like en.wikipedia.org/wiki/Generalized_Hebbian_algorithm to do PCA and when they happen to look at covariances caused by convolutional actions, the principal components they learn are these "rotated sinusoids" we observe as grid cells.
@jabowery
@jabowery 3 года назад
Replace Fourier with Laplace as conjectured mathematical framework for neuron function and see how that fits. Also something rather significant about simplex representation is it may lead to the application of linear programming optimization heuristics. Speaking of linear systems how is it that nonlinear warps arise from linear combinations of linear systems?
@karthikrajkatipally1497
@karthikrajkatipally1497 3 года назад
Laplace transform is good idea too. I suggested DMD in the thread. There is a paper called DYAN on arxiv. She is a colleague of mine. She used Laplace transform. It has same benefits as DMD. They are pretty equivalent. I use them interchangeably for my projects whereever spatiotemporal coupling is involved.
@karthikrajkatipally1497
@karthikrajkatipally1497 3 года назад
Regarding warping: They are linear locally. May be that's why. I guess we have to think of it like a manifold because of local linear interactions.
@karthikrajkatipally1497
@karthikrajkatipally1497 3 года назад
I really would like marcus to look into Dynamic mode decomposition. I always avoided Fourier for its instability and found DMD more generalized version. DMD perfectly fits this predictive coding or next state prediction like HTM does while doing Fourier kind of decomposition. The output from DMD is spatiotemporal decomposition; preserving space time coupling. objects and motion are separable based on flow. just like Jeff Hawkins was thinking about it Flow is all you need kind of idea from 1D modules: The complex Eigen values from decomposition cover frequency ranges starting from x + i0; (representative of static flow) and other complex Eigen values values representing other frequency components. HTM can overcome disadvantage DMD has. DMD needs stack of previous States and predicts the stack of next States. HTM doesn't need that since state is captured. DMD is used for uncovering physics model. HTM will be a continuous equivalent of it those class of models. ( Though temporal graph neural networks are norm for such tasks now).
Далее
ПОМОГЛА НАЗЫВАЕТСЯ😂
00:20
Просмотров 1,4 млн
The Clever Way to Count Tanks - Numberphile
16:45
Просмотров 696 тыс.
6.3 - Hippocampus and Place Cells
10:40
Просмотров 16 тыс.
The moment we stopped understanding AI [AlexNet]
17:38
Просмотров 854 тыс.
How to Soldering wire in Factory ?
0:10
Просмотров 12 млн
Проверил, как вам?
0:58
Просмотров 352 тыс.