Neural manifolds - The Geometry of Behaviour

Поділитися
Вставка
  • Опубліковано 21 тра 2024
  • This video is my take on 3B1B's Summer of Math Exposition (SoME) competition
    It explains in pretty intuitive terms how ideas from topology (or "rubber geometry") can be used in neuroscience, to help us understand the way information is embedded in high-dimensional representations inside neural circuits
    OUTLINE:
    00:00 Introduction
    01:34 - Brief neuroscience background
    06:23 - Topology and the notion of a manifold
    11:48 - Dimension of a manifold
    15:06 - Number of holes (genus)
    18:49 - Putting it all together
    ____________
    Main paper:
    Chaudhuri, R., Gerçek, B., Pandey, B., Peyrache, A. & Fiete, I. The intrinsic attractor manifold and population dynamics of a canonical cognitive circuit across waking and sleep. Nat Neurosci 22, 1512-1520 (2019).
    _________________________
    Other relevant references:
    1.Jazayeri, M. & Ostojic, S. Interpreting neural computations by examining intrinsic and embedding dimensionality of neural activity. arXiv:2107.04084 [q-bio] (2021).
    2.Gallego, J. A., Perich, M. G., Chowdhury, R. H., Solla, S. A. & Miller, L. E. Long-term stability of cortical population dynamics underlying consistent behavior. Nat Neurosci 23, 260-270 (2020).
    3.Bernardi, S. et al. The Geometry of Abstraction in the Hippocampus and Prefrontal Cortex. Cell 183, 954-967.e21 (2020).
    4.Shine, J. M. et al. Human cognition involves the dynamic integration of neural activity and neuromodulatory systems. Nat Neurosci 22, 289-296 (2019).
    5.Remington, E. D., Narain, D., Hosseini, E. A. & Jazayeri, M. Flexible Sensorimotor Computations through Rapid Reconfiguration of Cortical Dynamics. Neuron 98, 1005-1019.e5 (2018).
    6.Low, R. J., Lewallen, S., Aronov, D., Nevers, R. & Tank, D. W. Probing variability in a cognitive map using manifold inference from neural dynamics. biorxiv.org/lookup/doi/10.1101... (2018) doi:10.1101/418939.
    7.Elsayed, G. F., Lara, A. H., Kaufman, M. T., Churchland, M. M. & Cunningham, J. P. Reorganization between preparatory and movement population responses in motor cortex. Nat Commun 7, 13239 (2016).
    8.Peyrache, A., Lacroix, M. M., Petersen, P. C. & Buzsáki, G. Internally organized mechanisms of the head direction sense. Nat Neurosci 18, 569-575 (2015).
    9.Dabaghian, Y., Mémoli, F., Frank, L. & Carlsson, G. A Topological Paradigm for Hippocampal Spatial Map Formation Using Persistent Homology. PLoS Comput Biol 8, e1002581 (2012).
    10.Yu, B. M. et al. Gaussian-Process Factor Analysis for Low-Dimensional Single-Trial Analysis of Neural Population Activity. Journal of Neurophysiology 102, 614-635 (2009).
    11.Singh, G. et al. Topological analysis of population activity in visual cortex. Journal of Vision 8, 11-11 (2008).
    The majority of animations in this video were made using Manim - an open source python library (github.com/ManimCommunity/manim) and brainrender (github.com/brainglobe/brainrender)
    Icons by flaticon.com and biorender.com
    Relevant resources:
    Definition of a hole:
    mathworld.wolfram.com/Hole.html
    Cup-doughnut transformation:
    • Coffee Cup Donut
    Voltage traces and spike trains were obtained from atrificially simulated neurons using BRIAN2 python package:
    Stimberg, M, Brette, R, Goodman, DFM. “Brian 2, an Intuitive and Efficient Neural Simulator.” eLife 8 (2019): e47314. doi: 10.7554/eLife.47314.
    Generated spike trains were later analysed to obtain rate curves with the help of ELEPHANT:
    Denker M, Yegenoglu A, Grün S (2018) Collaborative HPC-enabled workflows on the HBP Collaboratory using the Elephant framework. Neuroinformatics 2018, P19. doi: 10.12751/incf.ni2018.0019
    ________
    Socials:
    VK: atpsynthase
    #SoME1

КОМЕНТАРІ • 510

  • @gyorgybuzsaki271
    @gyorgybuzsaki271 2 роки тому +216

    Super Artem

    • @generaltheory
      @generaltheory 2 роки тому +1

      A good start. Don't get too excited. You'll have too many actual insights if you get deeper.

    • @ardic97sokak
      @ardic97sokak Рік тому +1

      It is great, except the last comment about impressing girls!

    • @phasor50
      @phasor50 2 місяці тому

      what's wrong with that?@@ardic97sokak

  • @mbgdemon
    @mbgdemon 2 роки тому +303

    “So next time you try to impress girls by talking about topology you won’t be limited by coffee mugs and donuts”
    Fuck you got me, brilliant

  • @samizdat_
    @samizdat_ 2 роки тому +605

    If a layman like me can understand this then this guy is winning this competition for sure. Respect from India!

    • @ArtemKirsanov
      @ArtemKirsanov  2 роки тому +59

      Thanks!! I hope so 😅
      (I’m sorry for the late reply - my semester has just begun and biochemistry lab reports are being merciless)

    • @samizdat_
      @samizdat_ 2 роки тому +11

      @@ArtemKirsanov good luck with your semester!

    • @leif1075
      @leif1075 2 роки тому +2

      @@ArtemKirsanov is biochemistry mostly fun and enjoyable? Can you share how you stay motivated? And do you have to study 24/7 or just a few hours a day is enough? Thanks for sharing.

    • @aaaaaaaaidan
      @aaaaaaaaidan 2 роки тому +28

      An Indian guy says your UA-cam tutorial is good. You can relax now, you have officially won the internet.

    • @timothymaharaj5861
      @timothymaharaj5861 2 роки тому +2

      🇮🇳🇮🇳🇮🇳

  • @YoussefEIdrissi
    @YoussefEIdrissi 2 роки тому +83

    I am a researcher who investigates how architectural space impacts our psyche, and topology and Lacanian's psychoanlysis (based on topology) impressed me. Your video is just brillantly presented, the informations, the flow, the few puns, the animations. It's perfect. Keep on rocking more spaces!

    • @disectormusic
      @disectormusic 11 місяців тому

      will look this up! i keep eye matching multiple objects , play with perception, overall profiles of buildings, things, its amazing how things line up

    • @thepurpleman119
      @thepurpleman119 10 місяців тому +3

      The most pointless researcher job

    • @narcissesmith9466
      @narcissesmith9466 8 місяців тому

      If it gets proven mathematically that lacan models are somehow accurate I will literally shit myself

    • @UnoAluminio
      @UnoAluminio Місяць тому

      This is about neuroscience. Psychoanalysis has nothing to do with science. Good luck!

  • @gustavomezzovilla7248
    @gustavomezzovilla7248 2 роки тому +128

    This is INSANILY GOOD. Never saw so good exposition about topological spaces and manifolds without beeing pedant.

  • @ChechoColombia1
    @ChechoColombia1 2 роки тому +112

    This is why i'm studying mathematics. Thank you for this amazing video.

    • @DavidCaveperson
      @DavidCaveperson 2 роки тому +6

      As someone who has a BS in math and is struggling to land an actuarial job, study neural engineering

    • @fenrisunchained1926
      @fenrisunchained1926 2 роки тому

      The problem with math is teaching children they can have -5 apples. On paper it works in reality your teaching them debt.

  • @Mutual_Information
    @Mutual_Information 2 роки тому +314

    Dude, super well done. I also entered the competition and I think this is one of the best entries I’ve seen (I’ve seen a lot!). Good luck, well deserved!

    • @ArtemKirsanov
      @ArtemKirsanov  2 роки тому +17

      Thanks so much, man!! I appreciate it

    • @Mutual_Information
      @Mutual_Information 2 роки тому +2

      @@ArtemKirsanov This vid is blowing up!

    • @Mutual_Information
      @Mutual_Information 2 роки тому +10

      @@ArtemKirsanov Dude you were in the top 20! Congrats man I was hoping for this one!

  • @theloganator13
    @theloganator13 2 роки тому +140

    Awesome presentation! Interestingly, in my lab we use nearly identical methods to describe turbulent fluids. Resolve the fluid velocity vector field on a grid, stack the velocity components to make an N-dimensional vector, and you can describe the evolution of the fluid as a 1D trajectory through ND space. Once again, not all configurations of the fluid are possible because of physical constraints, so the trajectory is bound to the "inertial manifold".
    We can use this description to compute all sorts of interesting things about the fluid, but I should be writing that in my thesis and not on UA-cam right now 😅

    • @n119511m
      @n119511m Рік тому +7

      we should ALL be writing our thesis and not on youtube right now

    • @jessevos3986
      @jessevos3986 Рік тому +1

      That sounds super cool, where/with which keywords could I find out more about that?

    • @theloganator13
      @theloganator13 Рік тому +5

      @@jessevos3986 search for dynamical systems and state space, which will take you to articles about the general method. Add "fluid" to your search terms to see it applied like I described

    • @larryjoubert9992
      @larryjoubert9992 Рік тому

      Awesome

  • @jeromesullivan4th
    @jeromesullivan4th 2 роки тому +72

    'Thought-space' has long fascinated me, both neurologically and philosophically. I've often wondered if the mathematical and philosophical world could be connected by understanding thought-space on a neurological level.
    Excellent work by all those involved in this presentation!!

    • @davidarvingumazon5024
      @davidarvingumazon5024 2 роки тому +3

      Thought Space?

    • @ivoryas1696
      @ivoryas1696 Рік тому +1

      @@davidarvingumazon5024
      I think it's also called headspace, at times.

    • @user-tk7iu8vq5o
      @user-tk7iu8vq5o 10 місяців тому +1

      Idk if he talked about this. But I saw this concept in a (psych)trip once represented as a Möbius strip

    • @robinbelanger4294
      @robinbelanger4294 Місяць тому

      Yuuuuup! For years I've believed that every thought, idea, piece of knowledge/information has a kind of 'geometry' to it. So when I have that visceral recognition: I've seen this shape of thing before; I take that very seriously. It really shows how silly the siloing of fields are. Yes, you really did see that same "shape" in literature, and physics, and art, and 15 other places. We really need to be exploring and learning about the incredible creatures and ecologies that make up the human-zoo of thought space.

  • @ep9196
    @ep9196 2 роки тому +67

    Fantastic! been waiting for some 3B1B grade classes on Brain-Information processing. Please keep the subject going. Thanks

  • @DevashishGuptaOfficial
    @DevashishGuptaOfficial 2 роки тому +25

    I'm crying this is so insightful and beautiful 😢

  • @benwinstanleymusic
    @benwinstanleymusic 2 роки тому +24

    Absolutely brilliant, you've managed to explain all of this so plainly and clearly, one of the best videos on topology I've seen on YT

  • @Ruktiet
    @Ruktiet 2 роки тому +5

    Subscribed! Amazing presentation with extremely well-made animations and I am glad that you don't assume the viewers to be only interested in a watered-down version of the more difficult concepts so to make them easier to understand, but with loss of generality. Also, the profoundness of the result in the end (that the measured group of neurons actually represent a 1D manifold, just as expected) is extremely interesting. Thanks for this.

  • @LeonGalindoStenutz
    @LeonGalindoStenutz 2 роки тому

    @Artem Kirsanov, this is absolutely brilliant in so many ways. Just wow!
    Though i am far from fluent in any of the languages you are speaking, on a more intuitive and perceptive level, all you share in this video is absolutely relevant to addressing fundamental problems in neurology, environmental science, philosophy, and myriad fields and disciplines --including several which i am intensely focused including tackling species-level, planetary-level, and whole-systems existential threats and transforming them into existential opportunities.
    I will review again, dive a little deeper into some of the concepts for a stronger grasp, and hopefully return.
    First time i see a video of yours -- i am blown away. I've been following 3B1B for quite some time and love his work and spirit too. You both represent a whole new world that is being birthed.
    Your video above is exciting, deep, uplifting, beautiful, and quite literally enlightening. Thank you.
    Look forward to seeing more, and hoping to meet someday, hopefully in a not too distant future. Best regards from Brazil.

  • @robindoermann
    @robindoermann 2 роки тому +1

    This is easily the best explanation of the basics of topology I've ever come across. I finally understand what a manifold is! Fascinating video.

  • @churrundog
    @churrundog 2 роки тому +2

    Also, that closing statement truly hit the nail on the real practicality of topology

  • @agatha__jacob
    @agatha__jacob 2 роки тому +9

    This is so awesome! And to see that you were featured in 3Blue1Brown's videos is so refreshing to see. Great luck to you and to your journeys in math and Neuroscience!

  • @zhangalex734
    @zhangalex734 2 роки тому

    You deserve much more views! Such clarity and fluidity between topics. Just subscribed, please keep up the great work!

  • @heyjamilton
    @heyjamilton Рік тому +1

    This is an amazing video. Thoroughly explained a complex topic, and I really loved your emphasis on developing an intuitive understanding. Great work!

  • @adrianoaxel1196
    @adrianoaxel1196 2 роки тому

    This video it not only a piece of art in itself... it made me feel a positive hope for many reasons... how young this guy is... and how the work started by 3B1B is isnpiring other great quality works........ congratulations for this great work and.... thanks!

  • @EmissaryOfSmeagol
    @EmissaryOfSmeagol 2 роки тому

    The video started off really well. I was pretty hooked by your energy and all the neuroscience jargon.

  • @fabiopakk
    @fabiopakk 2 роки тому

    Fantastic work, I'm impressed by the discovery itself and how this videos is explaining the concepts so well. I really like the moments to pause and think, before the answer is given. Good job!

  • @MusicEngineeer
    @MusicEngineeer 2 роки тому +67

    that's interesting. however, i have a question that is a bit glossed over: is it actually justified to smooth out the spikes with a lowpass and just look at the average spike rate? i mean, from the get go, it's totally not obvious, why something like this should be justifiable. it's entirely conceivable that the exact timing of each individual spike matters - maybe they are synced in some way and need to arrive in sync at their target neurons to work as intended? i'm sure, neuroscientists have figured that out and have determined that this is not the case. but can someone shed some more light on that?

    • @animowany111
      @animowany111 2 роки тому +32

      Nope, there's no real justification to that, except that it works in most cases. The brain most often encodes data in stochastic firing rates (as far as we can tell), but we know there are some parts of the brain (like the optic nerve) where timing has a much larger effect.
      This is mostly to make analysis tractable, as adding precise timing information would blow up your dimensionality to something unusable. There aren't really any scalable algorithms that target that sort of timing data either.

    • @aitor9185
      @aitor9185 2 роки тому +47

      This is actually a rather hot debate in neuroscience: precise time vs rate coding.
      On the one hand, some people argue that neurons record the precise time of spikes and therefore smoothing out with a low frequency destroys the information you are tying to look at. There is some evidence for this, where some neurons act like biological logic gates at super fast timescales, such as an AND gate firing only if two spikes are received at the same time. It has also been shown that some neurons will always fire at the same time when doing a particular task, something like 50ms after starting a movement. However, the analysis of precise spike times is very difficult due to the high uncertainty of the recordings (some spikes can be missed or wrongly attributed to a neuron) and the lack of mathematical tools to analyze event sequences.
      On the other hand, many neurons also act like integrators, that is they remember spikes for a while and will become active after a certain threshold is hit. For the rate coding theory smoothing over many spikes is fine, since every single spike is not that important, just the average over a slightly longer time. Rates are often used because they are also much easier to work with than precise times and have been shown to work quite well for some predictions, such as the head direction (mentioned in the video), hand movement direction, auditory/visual input and more.
      I think the truth will be some combination of the two neural coding theories, which still have to be unified properly.

    • @MusicEngineeer
      @MusicEngineeer 2 роки тому +16

      aha! thanks for the replies. well, i would say that "it works in most cases" actually *is* a justification. i mean, that's how science operates, right?: make a hypothesis, make a prediction based on that, collect data, see if data is compatible with prediction. and if it is, the hypothesis get bumped up in credibility. but yeah - it would also be a lot nicer, if one would have not only experimental evidence but also a plausible physiological explanation. maybe there are indeed different kinds of neurons with different modes of operations. ...it's all very complicated... :-)

    • @Posesso
      @Posesso 2 роки тому +8

      @@MusicEngineeer Just to add something that might be related. I watched "How we see photons", and in there it was stated that multiple signals need to arrive from the eye to the brain so the brain says 'yeah. a photon was there' and you would see it. This signals come from individual cells, so for me the averaging is a mathematical representation of this mental process.
      I think that in general Brain needs more time than body sensors to produce meaningful outputs, and I think this justifies the opposite question: Would the brain be able to deal with such amount of signals without averaging? To this adds that moving hands, or legs is way slower, and Brain may need to wait.
      So, I would change, it seems to work in most cases for, the results seem to show that this is the most usual brain behavior.
      To me this slightly touches the Free Will debate, there is not an 'instantaneos you' that is worth examining, it is spread over a short period of time.
      @Aitor Thanks a lot for your expert information and super detailed post

    • @zvxcvxcz
      @zvxcvxcz 2 роки тому +4

      @@Posesso I wouldn't characterize it as averaging. It is more like integration, but note that there are negative inputs. Basically, a neuron will fire if it crosses a certain voltage threshold. If it receives a number of positives inputs in succession, then it will cross this voltage, but if the inputs are more spread out, then it may not because there is a leakage current (the voltage tends towards a rest steady state at all times). Not all inputs will be adding either, there are also inhibitory neurons that will lower the voltage of the receiving neuron. To complicate that mechanic further, a neuron that is heavily inhibited can fire when the inhibition stops because the pull back to steady state can actually overshoot and cross the firing threshold. Using averaging is a pretty crude tool here and to assess how well it works... well, there is potentially selection bias at play here (people often won't be publishing the times that averaging failed to get them useful results). Averaging is certainly useful at least sometimes, but there certainly is an instantaneous 'you' with differential merit. How one wants to apply that to the philosophy of free will, ship of Theseus, etc... is a matter I won't go into today. Integration in this sense is not averaging, I would encourage you to look at the classic Hodgkin and Huxley equations for modeling neurons. The Wikipedia article is titled "Hodgkin-Huxley model". This is still the general standard model of neuronal behavior with stronger models of the same derivation but with more kinds of ion channels or involving more complicated neuronal geometry and weaker models utilizing some sort of approximation.

  • @swankitydankity297
    @swankitydankity297 2 роки тому +1

    This absolutely blew my mind. What a fantastic video!

  • @someone5781
    @someone5781 Рік тому +1

    This is such a good video! I loved every minute of it and was absolutely captivated. Great job Artem!

  • @lileo4882
    @lileo4882 2 роки тому +2

    This is literally mind blowing. Awesome video and I’m excited to see more!

  • @bertrenard1
    @bertrenard1 Рік тому

    This is the best explanation of computational neuroscience I have ever heard !

  • @edwardmacnab354
    @edwardmacnab354 2 роки тому +2

    You're a smart guy. This presentation was extremely well done. Very thorough and very clear. The best I've seen yet. Please keep up the good work !

  • @captainjj7184
    @captainjj7184 9 місяців тому

    Amazing... opens up to new possibilities and way of approaches, thank you!

  • @ryanwalters7344
    @ryanwalters7344 Рік тому +1

    Smart man. I appreciate your clear explanations of a concept I find sometimes difficult to explain to laymen. I appreciate the food for thought.

  • @DrDeuteron
    @DrDeuteron 2 роки тому +25

    This is why spinors will never be intuitive...the topology of my neurons has me stuck at SO(3).

    • @davidarvingumazon5024
      @davidarvingumazon5024 2 роки тому

      SO(3)? What's this?

    • @adrianbergesenfedaque1183
      @adrianbergesenfedaque1183 2 роки тому

      @@davidarvingumazon5024 the Special Orthogonal group in R³. Basically a group (algebraic structure) formed by orthogonal 3×3 matrices that just so happens to encode every rotation in Euclidean 3D space.
      In other words, it's notation for the rotations in three dimensions.

  • @harrisashraff
    @harrisashraff 2 роки тому +1

    Though I am not a neuroscience student it is a joy to see how mathematics is being applied creatively to capture the essence of a phenomenon. You got a subscriber for life.

  • @hamiltonianthoughts
    @hamiltonianthoughts 2 роки тому +6

    I really enjoyed your presentation. Well done and looking forward to seeing more of neural manifolds. 💪🏻

  • @davidkhassias4876
    @davidkhassias4876 2 роки тому

    This is absolutely amazing Artem! Please keep doing such videos

  • @Bruno-el1jl
    @Bruno-el1jl 2 роки тому +2

    Unbelievably incredible video. Amaaaazing work!

  • @wideeyewanderer1785
    @wideeyewanderer1785 2 роки тому

    I’m only one minute in, and I already love the intro, no filler bs, instead a man itinerary of today’s presentation

  • @nucderpuck
    @nucderpuck 2 роки тому +6

    Sensory afferents in certain species (e.g. weakly-electric fish) can fire at rates well above 500Hz, but these are exceptional. Very nice video!

  • @NoahSpurrier
    @NoahSpurrier 2 роки тому +1

    Excellent. Nice detour that wraps it up where we began.

  • @BrianAmedee
    @BrianAmedee 2 роки тому

    this is BY FAR my favourite SoME submission!

  • @jacobchateau6191
    @jacobchateau6191 2 роки тому +1

    This is a fantastic exposition! I am excited to have found this channel.

  • @VerifyTheTruth
    @VerifyTheTruth 2 роки тому

    Your Video Is Efficiently Organized, Effectively Explained, And Informative At An Intuitive Level. Sincerely, Well Done.

  • @joseluismancilla-aguilar1183
    @joseluismancilla-aguilar1183 2 роки тому

    Congratulations on this interesting and entertaining video. Very abstract concepts such as manifolds, homeomorphisms, etc. are clearly explained without getting into technical complications.

  • @sedenions
    @sedenions 2 роки тому +4

    Excellent. You make me want to continue neuroscience studies by pursuing my Masters / PhD. I also like how you were able to omit using the word 'attractor' throughout the entire video.

  • @brian-classic
    @brian-classic Рік тому +1

    Amazing job on this video. Bravo. You've clearly taken inspiration from one of YouTub's top science communicators.

  • @RanLevi
    @RanLevi Рік тому

    Fantastic video! Thanks Artem, great explanations.

  • @bogodarov
    @bogodarov 2 роки тому

    Artem let me speak from my heart! I understand well your pronansation. This is first time so good for my listening skill

  • @lythd
    @lythd 2 роки тому +1

    I can't even express how good of a video this was, really good work :)

  • @JulioSantirachi
    @JulioSantirachi 6 місяців тому

    Now I have passion for topology in neuroscience. Thank you for these excellent videos!

  • @AnilKeshwani
    @AnilKeshwani 2 роки тому

    This is just straight-up fantastic. I learnt a lot, thank you.

  • @Nasim_bl
    @Nasim_bl 7 місяців тому

    I was studying the topic through a paper, and honestly I had no idea what is topology's business with neuroscience, your presentation shed a light on that. Thank you a lot.

  • @mateomendoza2232
    @mateomendoza2232 11 місяців тому

    I really love your videos and your job. I'm starting in the field of neuroscience with a background in biomedicine but little in math. Your videos really help me to understand some complex things. Thanks a lot! Also I really like you put reference and suggesting readings in your videos.

  • @KentYTsang
    @KentYTsang 2 роки тому +1

    This video was a huge benefit to me. I always wanted to get into topology, but I am only really interested in how it relates to the brain. Now I understand fundamental topology better, as well as how it applies to neural population dynamics. Thanks!

  • @alexharvey9721
    @alexharvey9721 11 місяців тому +1

    One of the best channels on UA-cam! I've got so much out of this channel, thanks so much 🙏
    I didn't realise head direction cells were strongly associated with the thalamus (always think EC by default).
    Though if we consider all global reference frames as egocentric (just hypothetically, though there is an interesting paper by Flavia Filimon on the concept), then it does kind of make sense that all cortical regions that collaborate spatially would retain strong origin-like components in their "signal".
    Time to impress girls with my new knowledge of topology!

  • @aitor9185
    @aitor9185 2 роки тому +1

    This is top tier! Glad to see the field getting some attention too :)

  • @sirnate9065
    @sirnate9065 2 роки тому

    I paused the video at 20:30 to think about & answer the question, and I got it right, which felt really good and shows that you did a good job explaining the concept. This was so interesting! I might have to check out some of the references in the description!

  • @peskarr
    @peskarr 2 роки тому +13

    Спасибо. Мало того, что тема крутая и необычная, так ещё и русский акцент помогает понимать речь и делает просмотр не таким напряжным, как если смотришь других англоговорящих авторов. Подписался на канал : )

  • @ChrisBennettGameDesign
    @ChrisBennettGameDesign 2 роки тому +1

    You need to do more of these videos. They are really good work. 👏

  • @kalla103
    @kalla103 Рік тому

    Great video! Information dense, yet accessible, well made!

  • @RomanNumural9
    @RomanNumural9 2 роки тому

    This is absolutely amazing. Please don't stop

  • @lowerbound4803
    @lowerbound4803 2 роки тому

    It's so well-done. Appreciate the motivating examples!!

  • @BPrismo
    @BPrismo 2 роки тому

    Awesome video man. when I looked at how many subs you had I had to double take because I thought it said 2 million! Good luck on the contest!

  • @ashzhu3085
    @ashzhu3085 2 роки тому +1

    This is absolutely fantastic! Good job!

  • @rikvermeer1325
    @rikvermeer1325 11 місяців тому

    you are one of the most clear&smart people i know, thank you

  • @klammer75
    @klammer75 Рік тому

    Love this guy! I am all over this approach! Topology, from Poincare, was originally thought of as an approach to the constraints of the solution set to various differential equations….think this approach can bear much fruit in this arena🤩

  • @sebastianmorales9787
    @sebastianmorales9787 2 роки тому

    Amazing video, congratulations! Incredible mix of topics and very exciting conclusion for what's next on neuroscience... Thanks!

    • @hughcaldwell1034
      @hughcaldwell1034 2 роки тому

      Your name is way too similar to Sebastian Moran, the right-hand-man of Professor Moriarty, and I'm a fan of it.

  • @rafaelarevalo8047
    @rafaelarevalo8047 2 роки тому

    how absolutely fascinating. thrilling video. thank you so much.

  • @shanaynay333
    @shanaynay333 2 роки тому

    So psyched that people are addressing this interesting topic. 🤔 Thank you!

  • @cexploreful
    @cexploreful 2 роки тому +1

    far beyond EXCELLENT!!💯

  • @matteobecchi1210
    @matteobecchi1210 2 роки тому

    This is amazingly clear and interesting! Thank you!

  • @santiagomarruffo9611
    @santiagomarruffo9611 2 роки тому +1

    Mind, blown. Brilliant!

  • @anastasiakolos9837
    @anastasiakolos9837 2 роки тому +2

    Great job and very interesting topic!) Good luck with the competition✨

  • @thinkmachine_7370
    @thinkmachine_7370 2 роки тому

    Very beautiful! Thank you so much for this video! This was so intuitive, I had an idea of how algebraic topology comes into play in the functioning of the brain, but to see it happen before my eyes, and visualize it left me wonderstruck and at a loss of words! Once again, thank you so very much for this!

  • @fenglongsong4760
    @fenglongsong4760 4 місяці тому

    Super nice! Thanks for the great explanation!

  • @noiJadisCailleach
    @noiJadisCailleach 2 роки тому

    That... that was incredibly incredible.

  • @JosephTakach
    @JosephTakach 2 роки тому

    Production value is insane!! Good job!

  • @KilgoreTroutAsf
    @KilgoreTroutAsf 2 роки тому

    This was actually a very informative video.
    Thanks and keep up the good work!

  • @ZubairKhan-sp8vb
    @ZubairKhan-sp8vb 11 місяців тому

    Amazing!!
    Artem your work if fabulous!

  • @chenmarkson7413
    @chenmarkson7413 5 місяців тому

    When I first watched this video two years ago, I found the concept of manifold weird and new. Now two years into uni, the chapter we are studying in my math course right now is precisely on writing proofs for smooth manifolds: tangent spaces, diffeomorphism, nonlinear systems, implicit form... How time has passed!

  • @kevinlopez287
    @kevinlopez287 2 роки тому +1

    This is so amazing. I’m a software engineer and using those analysis we could create an interface between brain and software without surgery.

  • @waseemashraf8407
    @waseemashraf8407 2 роки тому

    Loved your energy and the explanation was very smooth

  • @denismedvedev8214
    @denismedvedev8214 2 роки тому

    I think I only understood half of it, but the half I understood is amazing. Thank you for giving us this new perspective on the working of the brain.

  • @ArgumentumAdHominem
    @ArgumentumAdHominem 2 роки тому +21

    Hello Friend, great video, very faithful overview of the subject (I also work in neuroscience). I feel like the video scratched the surface of some very deep reaching concepts, so I was wondering if your could elaborate on a few questions of mine:
    1) What does the number of holes actually mean when applied to neuronal manifolds? Are they somehow a fundamental property underlying a given dynamical system that are invariant of its representation/implementation?
    2) If I understand the setup correctly, the neuronal recordings are a set of samples from a multivatiate probability distribution. Thus, experimentally obtained data is finite and not continuous. How does one proceed to estimate the shape of the underlying manifold given finite data? Is there any way to tell apart a hole from absence of samples, potentially due to undersampling?
    3) How reliable are the statements about intrinsic dimension being smaller than embedding dimension in real data. I would assume that there is some noise in the system, for example that introduced by the instrument or the pre-processing procedure (such as the spike rate approximate estimate). My naive guess would be that there would be non-zero variance along all embedding dimensions. Is this the case? Are there embedding dimensions along which the variance is truly zero? If not, then I presume that a dimensionality reduction procedure (like PCA) is used to drop some dimensions, whose relative variance is much smaller than for other dimensions. Such procedures always destroy information, although the fraction can be well-controlled. So, is it justified to artificially make the manifold smoother by throwing out dimensions of lower variance? In my experience, the largest variance axes are frequently unrelated to the research question (I still wonder what they do), while interesting task-related activity can be found in some of the lower-variance dimensions.
    I would be happy for any feedback (talk in person, answer, link to a paper or two, etc)
    Have a great day

    • @ArtemKirsanov
      @ArtemKirsanov  2 роки тому +14

      Hi! These are all terrific questions!
      1) I think this is one of the "holy grails" of theoretical neuroscience to find whether there is an invariant property, which is shared by all / many neural manifolds. Something about the dynamics that would indicate that it reflects the underlying "neuronal computation". But as far as I know, we currently don't have anything like that. As this stage, we are looking at different properties of different manifolds of neural activity in an attempt to find if there is a pattern and interpret it. For example stability (doi: 10.1016/j.neuron.2017.05.025) and orthogonality (doi: 10.1038/ncomms13239)
      And if we discover that neural activity manifolds of a certain brain areas, for example have different number of holes / number of holes depends on the behaviour, it is a starting point to hypothesise what it might mean for the underlying computation (such as the topology of the encoded variable), design additional experiments etc.
      But interpretation of what exactly these "invariant properties" of neural manifolds actually mean is still at very nascent
      2. Exactly! To uncover the "shape" of data (which are essentially discrete points) we use the algorithm called persistent homology (Great explanation: ua-cam.com/video/h0bnG1Wavag/v-deo.html). It allows us to reconstruct manifolds from a set of points and estimate their topological properties
      3. You are absolutely right, in real experiments data is always N dimensional due to the noise, caused by our measurements or the neural activity itself. I haven't looked into that too deeply, but I think just like with any PCA, we can ignore the dimensions with low variance. And there is some arbitrariness in this threshold. I'm afraid I don't have a great answer right off the bat, perhaps it's explained in more detail in the original paper itself (doi: 10.1038/s41593-019-0460-x)
      Hope some of it made sense ;)

    • @ArgumentumAdHominem
      @ArgumentumAdHominem 2 роки тому +5

      @@ArtemKirsanov thanks a lot for this broad reply! I will certainly look at the references you suggested

    • @lauravincent6935
      @lauravincent6935 2 роки тому +7

      whoa u two enlightened poor 17yo me

    • @aitor9185
      @aitor9185 2 роки тому +3

      Having read the paper I can add to question #3 that yes indeed they do reduce the dimensions with a method called "Isomap" which uses the minimum geodesic distances along the high-D manifold to perform the dimensionality reduction. In principle this method conserves the topology of the manifold, which is also why you can see the rings in their 3D plots.
      PCA is not that good because it assumes each individual dimension (neuron) is normally distributed, which would mean the manifold is a blob or sphere but not a donut.

  • @TimTeatro
    @TimTeatro Рік тому

    Amazing video: extrememly well done, well presented and on a fascinating topic.

  • @SehajSingh94
    @SehajSingh94 2 роки тому +1

    Intuitive indeed, great video!

  • @Mani_Ratnam
    @Mani_Ratnam Рік тому

    Now,this is getting more & more interesting..

  • @kaushaljani6769
    @kaushaljani6769 2 роки тому

    Simply Awesome !!! Hats off man
    ,Get to know about various fields in brief and in fun way...

  • @noiJadisCailleach
    @noiJadisCailleach 2 роки тому

    I'm a layman and played it at 2x speed. And I still understood it.
    Absolutely brilliant. BRAVO!

  • @AnthonyZboralski
    @AnthonyZboralski 2 роки тому

    Artem is an amazing explainer! Bravo!

  • @040_faraz9
    @040_faraz9 2 роки тому +1

    excellent work. Really enjoyed and understood it.

  • @AdobadoFantastico
    @AdobadoFantastico 2 роки тому

    Two of my favorite topics.

  • @homo.shepien
    @homo.shepien 2 роки тому

    Loved it. Very well explained.

  • @marshallross3373
    @marshallross3373 2 роки тому

    Great video. That must have taken a great deal of time to create and sync up all of the animations to the narrative. Well done!

  • @agranero6
    @agranero6 2 роки тому

    Last time I read an article about this idea on a huge dimensional space (one for each neuron) was on an article on a book: Zeeman E. C. The topology of the brain and visual perception. In N . K. Fort (Ed.), Topology of 3-manifolds. Englewood Cliffs, N.J.: Prentice-Hall, 1962 Part 6 The Metric on the cortex. That I read in the 80s. Then I never saw something related again.
    Interesting to see new things about that 35 years later.

  • @prikarsartam
    @prikarsartam 2 роки тому +23

    its a great representation of a good example, although I wonder how complicated structures would form for more complex neural activities like human cognitive behaviours; let's say abstract-decision making, or imagining an expression. I think shapes would turn more interesting in these domains, and would possibly require greater computational powers.
    Anyway, great job!

    • @michaelwinter742
      @michaelwinter742 2 роки тому +1

      Naw, dawg. Complex computations are handled by the way neurons connect together and a ton of repetition. The best studied example of this vision. If you look into it, you can easily see how collections of neurons detect, compare, contrast, categorize, perform analogies, and contextualize. Sure, there are other logical functions of neurons, but it’s a good start. Ya know?
      You wanna might check out the website by McGill University, The Brain From Top to Bottom.

    • @egor.okhterov
      @egor.okhterov 2 роки тому +1

      1. You need to elaborate what you mean by “abstract decision making”. It’s to vague of a question.
      2. The “power” used by brain when you solve complicated problem and when you recognise a celebrity in Instagram is the same. Actually the brain uses more power when you sleep then when you are awake.

    • @noelleletoile8980
      @noelleletoile8980 2 роки тому +1

      @@egor.okhterov do you have a reference for the amount of energy the brain consumes while asleep vs awake?

    • @tyo88teasley
      @tyo88teasley 2 роки тому

      I study questions somewhat related to this (at an intersection of depression and persistent homology) -- I think you're very likely correct about needing higher-dimensional structures, but the computational cost of articulating topological information (i.e., homology, as in the video) at more than 3 or 4 dimensions can be pretty prohibitive, especially on a population-sized dataset.

  • @rupeshmahore
    @rupeshmahore Рік тому

    I'm a physicist and recently discovered these manifolds. Super interesting. Thanks for this, I hope Grant Sanderson see this.

  • @danielayoutube6122
    @danielayoutube6122 2 роки тому

    This is sooooo cool! It's super well done and explained. Greetings from Portugal

  • @olbluelips
    @olbluelips 2 роки тому

    Wowww! This is amazing information and well-explained

  • @baron_xd4633
    @baron_xd4633 2 роки тому

    time flied this video, didnt even notice its length. brilliant! also a very nice ending statement :)

  • @PeteDimitrov
    @PeteDimitrov 2 роки тому

    Quite fascinating and really nicely presented!

  • @TheTHe0DB
    @TheTHe0DB 2 роки тому +1

    Great Video Artem thank you !