Тёмный

What is Spatial AI? "The Next Frontier of AI Architecture" 

Matthew Berman
Подписаться 329 тыс.
Просмотров 28 тыс.
50% 1

Fei Fei's interview with a16z, plus my reaction.
Try Mammouth now for just $10 today: mammouth.ai
Interview: a16z.com/podca...
Giveaway: gleam.io/q8wkK...
Join My Newsletter for Regular AI Updates 👇🏼
www.matthewber...
My Links 🔗
👉🏻 Main Channel: / @matthew_berman
👉🏻 Clips Channel: / @matthewbermanclips
👉🏻 Twitter: / matthewberman
👉🏻 Discord: / discord
👉🏻 Patreon: / matthewberman
👉🏻 Instagram: / matthewberman_ai
👉🏻 Threads: www.threads.ne...
👉🏻 LinkedIn: / forward-future-ai
Media/Sponsorship Inquiries ✅
bit.ly/44TC45V

Наука

Опубликовано:

 

3 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 173   
@Cine95
@Cine95 22 часа назад
Matt playing that video at 1.5x while I am playing this video at 1.5x
@bestemusikken
@bestemusikken 21 час назад
I'm always watching 2x speed. And god, I want 2.5x now. Love it!
@Cine95
@Cine95 21 час назад
@@bestemusikken well technically its 3x
@ajbent1
@ajbent1 21 час назад
@@bestemusikken Just open the Developer tools and type: $("video").playbackRate = 3.0; You can set the playback to whatever multiplier you want.
@heartyfisher
@heartyfisher 21 час назад
me too !! .. (but I watched the SpatialAI before...)
@49531
@49531 20 часов назад
@@bestemusikken with Enhancer for RU-vid extension you can set speed from 0.1 to 16x(sic!)
@mendthedivide
@mendthedivide 21 час назад
00:00 - Introduction to Spatial AI and Fei-Fei Li 00:35 - Fei-Fei Li's Contributions: ImageNet and AI Research 01:43 - Cambrian Explosion in AI: Pixels, Videos, and Language 02:09 - AI's Evolution Through Deep Learning and Data 03:12 - Early Days of Deep Learning: ImageNet and Commercial Applications 03:54 - Rise of Generative and Discriminative Models in AI 04:20 - Algorithmic Advances in Computer Vision and Language Modeling 05:16 - AI’s Power Unlocked by Large Data and Compute 06:14 - Key AI Papers: Attention and Transformers 07:28 - Breakthroughs in AI with AlexNet and GPU Power 09:57 - Supervised Learning Era and AI Data Utilization 11:22 - AI Research and Algorithmic Breakthroughs in Academia 13:54 - Evolution of Generative AI: Style Transfer and Image Generation 15:10 - Speed and Optimization in Generating AI Images 16:36 - Gradual Advancement of AI Towards AGI 18:05 - Fei-Fei Li’s North Star: Visual Intelligence and Storytelling 19:46 - Current AI Capabilities: Computing Power and Algorithm Depth 21:13 - Tesla’s Use of Real-World Data for AI Training 23:08 - Transitioning from 2D to 3D: Learning Structures in AI 24:27 - Nerf’s Breakthrough in 3D Computer Vision 25:33 - AI's Focus on Reconstruction and Generation Convergence 27:04 - AI Representation of the 3D World Through Physics and Structure 28:12 - Spatial Reasoning and Limitations of Multimodal AI Models 29:14 - Contrast Between 2D and 3D World Understanding in AI 30:12 - Processing and Representing 3D World Data in AI 31:35 - Human Perception of 3D World Through 2D Visuals 32:57 - Future Applications of AI in Virtual 3D World Creation 34:19 - 3D World Creation and its Economic Impact in Gaming 35:46 - Desire to Explore and Simulate 3D Virtual Worlds 36:58 - Impact of Spatial Intelligence on AR and VR Technologies 38:00 - Spatial Intelligence as the Operating System of 3D Future 39:02 - Ideal Hardware for Spatial AI: Glasses vs. Goggles 39:33 - Blending Digital and Physical Worlds: AR and Robots 40:06 - Conclusion: Spatial Intelligence’s Role in Robotics and AI
@JustaSprigofMint
@JustaSprigofMint 22 часа назад
I watched this interview on the a16z channel. it is so amazing how much AI is making actual progress even among all the over-bloated market economics taking place right now into making anything and everything AI which is making many people wary. Well, if the internet and AI took away your critical thinking skills, AI and internet isn't responsible for it. I'm so sick of channels interviewing "CS researchers and scientists" and they spend like an hour decrying everything AI. Folks, use you own goddamned brains a little!
@remsee1608
@remsee1608 22 часа назад
Sorry nobody has translated the bike into brainrot yet
@johnwilson7680
@johnwilson7680 19 часов назад
Please don't speed up the video. We can speed it up if we want to.
@hqcart1
@hqcart1 22 часа назад
you cant watch technical interview at highh speed
@ChrisKruger75
@ChrisKruger75 21 час назад
I slowed it to .75% to have a better time with the Video, unfortunately, that made Matt sound much less smart that I think he is ;-)
@ew3995
@ew3995 19 часов назад
no you cant
@RikHeijmen
@RikHeijmen 15 часов назад
You can. Not when the video already was 125% 😂
@00101010111001101100
@00101010111001101100 15 часов назад
@@ChrisKruger75 lol, I just slowed it down because I had fast playback. Now he sounds like a drunk person trying to recap something we are watching.
@Azul-m1u
@Azul-m1u 7 часов назад
what u yappping for ?
@the42nd
@the42nd 21 час назад
Didn't feel like their call brought anything new to the table at all.
@badashphilosophy9533
@badashphilosophy9533 13 часов назад
The issue for AI is having no environment therefore causing inability to access or navigate beyond calling apis. Ifigured the trick is to put a remote dpi into an app so u can display a computer environment in a window on your own desktop, then add a node interface over that and train ai on navigating the node system It works because each node is a sensor and informs the back end of pixel colour directly underneath each node in realtime and also functions like mouse clicks and keyboard inputs so the view of nodes and functions occur together, and train ai on navigating the node system with machine learning beginning with just training on a user just using the computer themselves. The ai never has to see the display because the nodes give all the info so if it can learn to navigate the nodes then it can navigate the computer without any direct connection which means jack all overhead. Not sure if people realise this but even us humans don't see the world, we see a simulation informed by signals from the senses, one type being rods and cones in our eyeballs, this node system would be the rods and cones interface that by correlating with action allows ai to link action potential with view of a digital environment. The coolest thing about this is the ai would be in an app on another computer so it could navigate the internet in the same way humans do to find the info it needs to do tasks and could build seperate apps, rebuild new versions of itself on another computer using other ai like Pythagora and test its apps by actually opening them and trying them, it could set up a system where it remotely navigates multiple remote computers at a time and trains on all of them, ai building ai, automation automating automation and automating iterations out of better interactions with manual input becoming less and less in each one The crux of this is giving ai an environment. People have tried with robots and virtual environments but why not the digital space itself, that's the first thing we need to do
@jbavar32
@jbavar32 21 час назад
So many incredible use cases for this technology. People creating their own ‘happy places” they can just put on the headset and submerge into while on a long flight.
@tomcraver9659
@tomcraver9659 13 часов назад
I've done a few tiny experiments just prompting an LLM having it think of real world descriptions in terms of spatial coordinates of objects. It seemed to be a little better at solving real world 'puzzles', like what happens to a ball in a cup that gets turned upside down. Not perfect - but it did uncover some of the model's vague understanding of what a 'cup' is, for example.
@panzerofthelake4460
@panzerofthelake4460 21 час назад
OpenAI still having a context length of 128K is just a bummer. If they won't increase that, I can see them losing the high-ground regarding the reasoning models. Google's Gemini has a Context length in the range of MILLIONS and this could give them the upper hand, since reasoning models essentially use up that finite context length pretty quickly. Of course, summarizing the thought process and feeding back only the summary can reduce the amount of tokens needed, or simply making the AI think in a more efficient language that humans might not be able to understand (at the expense of transparency) could also help, but ultimately, they'll need to increase that context length. It will get ugly for them if they won't.
@zmeireles68
@zmeireles68 15 часов назад
This 40 min long video was worth every sec of it. Thanks Matt.
@alexlanayt
@alexlanayt 22 часа назад
fei fei is great! she have the pioneer's dna
@solidreactor
@solidreactor 2 часа назад
Regarding about the image generation being more like a continuum from previous milestones but was regarded by the public like an abrupt new thing, it made me think of the quote: "It takes 10 years to become an overnight success."
@chriswatts3697
@chriswatts3697 21 час назад
Spatial AI is a thing for sure. Our world is dimensional. If an AI wants to understand everything that is happening in our world, it has to learn about all the dimensions and the physical connections. You said that, for example, Tesla has a lot of spatial data. That is correct for sure, but for quite some time we collect spatial information with our mobile devices. Think back to 2014/2015 when google showed project "tango" (the mobile AR that became ARCore in the end). What they wanted is to capture all the insides of the buildings they capture from outside with Google street maps. And we should not forget that also audio is spatial. You can capture a lot of data just by analyzing audio data (submarines do this for a long time). I am very excited about the spatial AI approach and i hope it will be available for everyone (open source). Thanks for the video, Matthew.
@honkiemonkey33
@honkiemonkey33 5 часов назад
It seems that the practical applications of these developments are still in the process of being fully realized. I hope they are finding ways to apply these innovations beyond just gaming. With that in mind, here are a few possible areas where they might have significant impact: 1. Generating a comprehensive set of architectural and engineering designs based on site parameters and design preferences. 2. Creating 3D product designs, such as furniture or wearable technology, that adapt to environmental factors and surroundings. 3. Offering emergency assistance through augmented reality, such as using smart goggles to guide someone through landing a plane in a critical situation. 4. Enabling underwater robotic welding to facilitate complex repairs in challenging environments. 5. Utilizing autonomous drones that can navigate hostile environments and selectively target designated individuals. It might sound harsh, but it’s likely similar technology to what they would be using for shooting games.
@Ben_D.
@Ben_D. 22 часа назад
Did you forget to link the A16z vid mate? Cant see it in the desc.
@matthew_berman
@matthew_berman 22 часа назад
Updated
@SarahKchannel
@SarahKchannel 21 час назад
Either ask for permission to play somebody elses content, or take your chances to get a copyright strike - but fast forwarding to fool youtubes algorithm is stupid.
@matthew_berman
@matthew_berman 21 час назад
@@SarahKchannelI didn’t fast forward to trick the algorithm. This is clearly fair use since I give my commentary on the vid.
@minu42yu
@minu42yu 21 час назад
This stream of mind-blowing news just gets crazier every day... if it continues like this, you will have to switch to live broadcasting. Your insights and comments are very valuable.
@matthew_berman
@matthew_berman 21 час назад
Thank you!
@henrismith7472
@henrismith7472 22 часа назад
How about a video series teaching us how to put our Nvidia gaming GPUs to use for AI. I know you already have tutorials on running llms locally, but I keep finding out about new features and use cases with my RTX4080, and I can't find a RU-vid channel that really focuses on getting the most out of these GPUs for AI. Chat with RTX is pretty cool for beginner's. I've moved on to OpenWebUI and ComfyUI for my llms and diffusion models. Currently trying to figure out ComfyUI properly, isntead of just using other peoples workflows and dropping my own loras in. I want to learn how to use 2-3 loras. 1 lora for creating images of myself, one for a certain style of image, and one to improve photo realism, skin tone, etc.
@attilaszekeres7435
@attilaszekeres7435 17 часов назад
If we're crafting AIs in our own image, we shouldn't expect them to think outside the box we've built. Matt you should stop erpeating "real world" and wipe off that overzealous grin - this is a funeral my man. What we're teaching AI to perceive is the world we live in. Even the visual representation of local reality at the visible spectrum will be rendered differently by beings with a different central nervous system. It is okay to limit AI to human cognitive/perceptive frameworks, but don't then complain that it can't provide novel insights. This is another nail into the coffin of ever having a truly foreign intelligence, another brushstroke on the frame of the mirror we like to admire ourselves in.
@fontenbleau
@fontenbleau 14 часов назад
if machine goes on higher sensory level (how many dimensions they theoretically counted already? like 11 above us?) - we will stand where we are stuck forever, it makes the feeling like birds leaving the nest to the winter migration. I watched this summer storks nest full growing, yes only one chick survived from 4, but it was personal sadness when it flies away and maybe never returns, if even survives in Africa at all. Humans need to transform into digital entity, because every mammal are made from a cell, from foreign microbial world by nature's help we transformed into bigger creatures/sizes world which kinda goes nowhere, it will be natural to leave this world not by collapsing Dna in every cell, moving next level.
@acromm
@acromm 6 минут назад
An early lunch of Optimus is in my opinion also to fullfill the need of visual data of homes and what happens there, because homes are not streets.
@CarlosFlores-ke1lk
@CarlosFlores-ke1lk 20 часов назад
Wearing a pony tail makes you look smarter. Just like having a british accent. I forgot, if you talk fast you are by definition a genious - so much to share with us mere mortals...
@joshkar24
@joshkar24 5 часов назад
I feel like our ancient ancestors - pre-language - had a non-communicated version of language that language sits on top of but doesn't necessarily describe, even with huge language sets. We can hard code physics and things of that nature to almost literally ground AI in the real world, with being able to probe its answers for a quite material-based explanation for why it is stating something. One of the most captivating use cases for VR is just exploring and screwing around in photoreal 3d environments, "telling stories / pretending" BUT there is a huge lack of these at low price that are quality - I have to dig through about 80% detritus for the good stuff. If this suddenly wasn't the case, it would mean endless novel worlds to explore, personalized, responsive to speech, hand motions, etc. This drives mass adoption of VR, lowering the cost of high res, high FOV headsets that are comfortable, which in turn drives forward spatial AI
@GeoffCoope
@GeoffCoope 18 часов назад
To go with Spatial AI, I hope they also combine it with a physical structure that not only morphs under and around you but also visually represents the Ai world from our own ocular perspective. If that can be invented in my lifetime then I will be rather pleased. I have always dreamt of having my own Holo-Deck.
@ZoOnTheYT
@ZoOnTheYT 14 часов назад
That's a lot of words to say we're trying to build Star Trek's Holodeck.
@gavincrouch
@gavincrouch 13 часов назад
I would be interested to know what the training data would be - 3D scans, dual camera or possibly control net data with depth maps and positioning info. The real challenge is as she states, a human can look at an infinite number of variations in the world and know what it is instantly (a chair for example, regardless of shape or build complexity we know is a "chair" even if we've seen one for the first time in our life, we just know it's a "chair" - for AI in 3D space this would be a challenge because it has no text cues, and manual human tagging for every object on earth is practically impossible). Tesla is limited to road data, it has no concept of objects invisible to it outside of the narrow depths of field from roads.
@RWilders
@RWilders 9 часов назад
Fascinating. I would point out though that ADN condons are "words" in nature (not floating in the sky, but embedded in every living creature) and that they convey 3D intelligence.
@PhilB-yl4jk
@PhilB-yl4jk 18 часов назад
verses ai is working on similar spatial projects. They created the spatial web protocol that accepted by the IEEE.
@fslurrehman
@fslurrehman 21 час назад
I don't understand what World Labs is up to. The concept of modeling the world and reasoning about objects, places, and interactions in 3D space and time closely relates to topology, which is the mathematical study of how spatial relationships and properties persist or change under continuous transformations, such as stretching or bending without breaking. So, they are either giving a new name to the study of topology, or they don't fully grasp what math has already covered. Additionally, using mathematical reasoning and logic, combined with Multiphysics Simulations, produces more accurate results than relying on probabilistic token-based models for representing objects.
@antonystringfellow5152
@antonystringfellow5152 18 часов назад
"Additionally, using mathematical reasoning and logic, combined with Multiphysics Simulations, produces more accurate results than relying on probabilistic token-based models for representing objects." I'm not sure how useful this approach would be when it comes to interacting with the real world, which is so complex and unpredictable. It wouldn't be sufficient to control a robot or autonomous vehicle. Certainly, a simulated environment is useful for training deep learning models but they'll still need training.
@remsee1608
@remsee1608 18 часов назад
Topology sucks worst class in college
@yoyo-jc5qg
@yoyo-jc5qg 21 час назад
fei fei, ilya sutskever, and alex started the ai deep learning revolution with alexnet back in 2012, they're so underrated ... to this day deep learning continues to conquer field after field, amazing
@stevenguevara3252
@stevenguevara3252 11 часов назад
Wtf… you threw me off. Lmao Thought my youtube was playing at double the speed.
@pik910
@pik910 2 часа назад
I hope they use all kinds of sensors when they build their data set, I like the idea of AI being able to estimate the temperature of things based on an image.
@francoislanctot2423
@francoislanctot2423 15 часов назад
I am surprised they did not mention Gaussian Splatting.
@YoungMoneyFuture
@YoungMoneyFuture 11 часов назад
Bro drank 5 red bulls before the interview😵‍💫
12 часов назад
I already did these experiments and development in the 90's using REAL PEOPLE, as a cash Math/Physics tutor (now over 80,000 hrs). Guaranteed clients an A in Engineering Physics 1, by reverse-engineering an expert system from the Curriculum, and employing "Lazy Learning" techniques, employing the HUMAN BRAIN as my neural network. I extended the concepts to the laboratory teaching, and my class became so popular, 3 hours per week, NO HOMEWORK, that the University changed a by law to allow withdrawing from the lecture class but keeping my class. Not only did "compute" not cost me anything, people paid ME to do my compute!. Also mastered transition to Algebra 1 fluency. No one wanted it, the schools just being totally corrupted with mediocrity. So now I tutor the smart, well to do, immigrant crowd and enjoy watching the world catch up.
@DJWESG1
@DJWESG1 8 часов назад
Tutor the poor and least well off locals. . don't wait for them to "catch up", you'll be waiting forever.
@fslurrehman
@fslurrehman 20 часов назад
Meta's SceneScript is already capable of reconstructing environments and representing the layout of physical spaces. So, the question remains: what truly innovative or new approach is World Labs bringing to the table? And finally, there’s a need to reinvent Meta's Orion AR glasses and the advancements from Boston Dynamics.
@h.c4898
@h.c4898 19 часов назад
This is the future in my opinion. I see this modaility taking over the 1D, 2D image modality in the near future. It'll give much more details in color and nuances to the LLM to process. It might be a little more computionally expensive but the benefits will outeeigh the risks and "costs" I think. The three modalities LLM has now will reach their thresholds at some point. It'll need to lean on more advanced technologies like this one too process more nuanced representations from the 3D or the real world. Another one is video reading. Current LLMs can't do that. It reads transcripts only not the visuals. With the camera shutter mode it's possible for this tech to read videos in 3D. Those will enrich the LLM in its representation algorithm. With a richer context, it'll reduce case of hallucinations even more. Anyways, this is exciting stuffs if you'd ask me.
@TailorJohnson-l5y
@TailorJohnson-l5y 19 часов назад
Good look on this one Matt thank you!
@medwards1086
@medwards1086 4 часа назад
Is everyone gonna ignore the work of Jeff Hawkins like he hasn’t been doing this for 20 years a Numenta?
@gcg2927
@gcg2927 16 часов назад
iPhone have had LiDAR sensors since the iPhone 12 Pro but haven’t been utilised much, photogrammetry is super interesting
@epokaixyz
@epokaixyz 20 часов назад
This might be exactly what you need: 1. Learn about spatial AI and its potential impact on various industries. 2. Explore resources like the a16z Podcast with Fei-Fei Li and platforms like Mammoth AI to deepen your understanding. 3. Consider the ways spatial AI can enhance robotics, augmented reality, and world generation in the near future.
@avsuunInfoSEC3391
@avsuunInfoSEC3391 12 часов назад
The five basic human senses are Sight, Hearing, Smell, Taste and Touch. Ai will be doing that.
@marcuss9017
@marcuss9017 20 часов назад
very interesting topic, as I work in virtual construction, obviously it goes way beyond this but just reaffirms my opinion that ppl won't be designing and coordinating much for construction for much longer.
@jjhw2941
@jjhw2941 19 часов назад
Interestingly a paper came out just a few days ago, Robin3D: Improving 3D Large Language Model via Robust Instruction Tuning.
@derekholland3328
@derekholland3328 10 часов назад
loving this channel. such great insight.
@MatthewChowns
@MatthewChowns 18 часов назад
when you play a video at high speeds as part of a recording, you are making it hard to follow for many people. please keep it to normal speed from now on and let us pick the speed in the youtube player ourselves
@pierremacb4668
@pierremacb4668 20 часов назад
Somehow we're still seeing LLMs improving a lot (say, twice a year, e.g. with 4o, 1o) with overall the same amount of data, perhaps they're already incorporating more than 1D/2D in there..
@jackflash6377
@jackflash6377 20 часов назад
Research! Researching in 3D would be ground breaking. Everything that is important is 3D. Genetics, physics, engineering. All 3D.
@pierremacb4668
@pierremacb4668 20 часов назад
no. everything important is nD.
@joshkar24
@joshkar24 5 часов назад
whoever can crack the cheap and fast apprehension of high quality 3d structure of the real world will provide a huge advantage to many fields, this has been a huge and expensive struggle for decades now, making design, engineering and games so much more expensive.. An "unlock" like this will provide training data for calculating 3d from the billions of 2d photos, videos. Then this links up to LLMs and makes them more precise as well. Once you have a detailed and precise understanding of a 3d environment there is TONS of "free" functionality - simulate new lighting, calculate volumes, path-finding for robots, physics, etc And these tasks can be done in non-real time as needed. This leads to AI-designed automation solutions acting on the real world, deploying drones, IoT etc which then completes the loop so we get a real-time sim of the real world with fidelity proportionate to locations of interest.
@barendencrone
@barendencrone 18 минут назад
except Tesla's data isn't spatial since Elon insisted cameras were enough, no?
@marzchart591
@marzchart591 21 час назад
Like I told openAI, it's a megawatt per million for your projects (mouha ha haha (the energy company rubbing its hands))
@arlogodfrey1508
@arlogodfrey1508 20 часов назад
Turning on captions makes a mess when the video in the video also has them
@h.c4898
@h.c4898 20 часов назад
Yes. If u feed the LLM with them in the video. Shut caption off, make sure the video is clean or represented properly before input. That will reduce case of hallucination.
@kiranwebros8714
@kiranwebros8714 6 часов назад
This is the right path, Spacial intelligence with point of interest.
@DJWESG1
@DJWESG1 8 часов назад
"around 2012 ish" Yup. But the idea didnt pop into his head.
@I-Dophler
@I-Dophler 20 часов назад
🎯 Key points for quick navigation: 🌐 Spatial AI integrates AI with spatial data for enhanced contextual understanding. 🏙️ It enables AI systems to perceive and interact with physical spaces in real-time. 🚀 Spatial AI is considered the next frontier in AI architecture due to its applications in robotics and autonomous vehicles. 🌍 It revolutionizes industries by optimizing logistics, urban planning, and environmental monitoring. 📡 The technology enhances augmented reality experiences by overlaying digital information onto physical environments. 🛰️ Companies are investing heavily in Spatial AI for its potential to redefine navigation and mapping technologies. 🏗️ Challenges include privacy concerns and ethical implications regarding data usage in spatial mapping. 🌌 Future advancements in Spatial AI aim to create seamless interactions between digital and physical worlds. Made with HARPA AI
@MonkeyBars1
@MonkeyBars1 19 часов назад
great commentary video thanks man
@godtable
@godtable 22 часа назад
Welcome! To The World... Of tomorrow! :)
@tubestreamkyki
@tubestreamkyki 21 час назад
I'm now listening 3x speed of Feifeili
@panny3897
@panny3897 20 часов назад
Gotta say im not a fan of speeded up videos. Let us decide when to speed up please. You made the content less accessible 😕🤷🏽‍♂️
@tsentenari4353
@tsentenari4353 14 часов назад
actually even some of the showcased, so, somewhat cherry picked Sora videos made major spatial blunders
@steveconnolly3776
@steveconnolly3776 23 минуты назад
What's the URL to "Archive" that is mentioned in the video?
@deadchannel127
@deadchannel127 13 часов назад
i skipped to the interview and was like wow these tech bros are definitely on adderall
@planetmuskvlog3047
@planetmuskvlog3047 21 час назад
On October 10th the world will see that Tesla inc. has already built a real-world AGi. You’ll see
@denisblack9897
@denisblack9897 20 часов назад
I cant help but chuckle everytime people say about new paradigm of reasoning and chain of thought… I put together a group chat for gpt3 to discuss my prompt and witnessed the power of it a few days after chatgpt release… Am i genius or are these ‘researchers’ dumb?😅 I dont get it guys, are they slowly milking the cow that should have been a phrnomenal steak right away?
@IgorKravets81
@IgorKravets81 20 часов назад
What if architecture of a model (weights interpreter) should be also dynamic and trainable via deep learning or via selection/mutation Darwin-style mechanic
@njabruzzo
@njabruzzo Час назад
While 3D can be generated, lowering the creative cost for building those environments - I’m doubtful AVP is the right device to explore those worlds. The form factor and cost is too high. (I don’t see next gen kids willing to wear those for any length of time)
@ericvanartsdalen161
@ericvanartsdalen161 16 часов назад
I noted Justin Johnson spoke about 4D... to me, that speaks to AI dealing in a holistic view of the world, not only the basis of 3D models, but how those 3D models live in their placement within the world. But consider also, over an object's relative position in a given time in space. But I could be wrong... but that seems to be some far out Einstein Relativity thought here. What do you think?
@RickySupriyadi
@RickySupriyadi 20 часов назад
Alexnet vs Imagenet what's the difference?
@arron122
@arron122 21 час назад
Without spatial intelligence you won't reach GAi (human level). You need the Ai to think and visualize in 3D space and time like the human mind does. Current image/video Ai generators, trained on 2D data, are pretty useless in the commercial world. You would want full control and management of your props, textures, assets, lighting, in your Scene in the x,y,z coordinates to achieve the vision you have in your head. I've spent too much time and energy trying to generate an image or animation the way I want it using inpaint, ipadapter, SAM... I'll just have to wait a bit longer.
@marcfruchtman9473
@marcfruchtman9473 Час назад
Thanks for the video.
@karoinnovation1033
@karoinnovation1033 19 часов назад
I love this channel!!
@01Grimjoe
@01Grimjoe 14 часов назад
Very interesting, thank you.
@Ro1andDesign
@Ro1andDesign 17 часов назад
I bet this model would be very good at the marble problem 😃
@live--now
@live--now 17 часов назад
Damn ive already saw that interview.. i thought it was something new... Ok ..👍 👍
@JohnLewis-old
@JohnLewis-old 13 часов назад
please don't set the speed for me, I do that myself.
@lfrazier0417
@lfrazier0417 15 часов назад
I swear that I watched this video months ago.
@NLPprompter
@NLPprompter 10 часов назад
They said .. plumber cannot be replaced... with spatial intelligence you can be plumber! with that device it can project AR Of pipes behind the wall because spatial intelligence able to generate 3d!, mechanic?
@Anders3n
@Anders3n 19 часов назад
Mammoooot 🤣
@LucBouvattier
@LucBouvattier 3 часа назад
@Mammouth_ai
@Mammouth_ai 3 часа назад
woop woop
@LucBouvattier
@LucBouvattier 2 часа назад
@@Mammouth_ai is that really the sound of a Mammouth ?
@feedvid
@feedvid 9 часов назад
Fascinating! 👍
@koliux1
@koliux1 16 часов назад
Well... very informative and cool and I can see it being applied in many industries, however it feels like there is something missing. Oh yes where on earth are you going to store all that spacial data ? The next big thing would be then Storage! Not compute...
@IgorKravets81
@IgorKravets81 20 часов назад
It seems to me, that this 3d dimensional architecture will turn out to be a very narrow niche approach and we will laugh at it (as well as at transformers) once agi is achieved. When a baby learns the world it only gets 2d images from two eyes, two microphones and a gazillion different touch sensors all over the body. I bet almost nothing in the brain is 3d oriented when the baby is born. AGI architecture would not necessarily incorporate 3d in hardware. I think it will be more generic. Even simpler than transformers. But will require much more compute. Imho.
@Agoncor
@Agoncor 21 час назад
This will kill TESLAS FSD Advantage
@robbrown2
@robbrown2 21 час назад
Please don't speed up videos, I can speed them up on this end if I want to.
@BorisKehr
@BorisKehr 6 часов назад
Tack!
@taurean.williams
@taurean.williams 16 часов назад
Basically building the Matrix
@justcars2454
@justcars2454 15 часов назад
40minutes, and title is what is spatial ai, is this topic need a whole 40mimuuuuuuuutesssssss
@EriCraftCreations
@EriCraftCreations 20 часов назад
So very fascinating 🎉
@Outcast100
@Outcast100 15 часов назад
Contact lenses with AR combined with neuronal output device on your ear with mic and bone induction sound ..(since brain input is still theoreticall - and risky when it will be possible ) You would have AR and besically think what you want and it appears in AR.... ex:call somone ..time...video....see the collusem in real life and overlay as it was with gladiators and all, games like pokemon go but actually see the pokemon (Now that i think about it we might have to invent a few new names for mental problems that people will have😅)
@ploppyploppy
@ploppyploppy 21 час назад
More bubble money :) Ah it's just like watching the late 90s all over again.
@DailyTuna
@DailyTuna 20 часов назад
I’ll pass using my data so that these people can be rich and have power😂
@rokljhui864
@rokljhui864 16 часов назад
Normal-speed talking and thoughtful comprehension are so 'last decade'. How about compressing everything into a 1 second tik-tok ?
@user-uv3nv2bc6v
@user-uv3nv2bc6v 14 часов назад
A huge step towards Metaverse?
@alankeeling2946
@alankeeling2946 20 часов назад
Simple WE'RE F'ED - end of discussion
@DeepThinker193
@DeepThinker193 18 часов назад
No, I wear a chastity belt so I'm not F'ed
@justinlloyd3
@justinlloyd3 20 часов назад
She sounds like she is going in the right direction
@ChuckBaggett
@ChuckBaggett 12 часов назад
The 1.5 speed sounds terrible, like a cartoon duck or something, and earned this a thumbs down and not finishing and not sharing.🤬
@ITSupport-q1y
@ITSupport-q1y 16 часов назад
Smart people
@fontenbleau
@fontenbleau 14 часов назад
Python is dead end! One guy, which famous now, writes neural net in low-level Assembly and it's 500 times faster than Pytorch on 1 Cpu core. That insufficient power to run cumbersome high-level Python is illusion and trap! Almost a conspiracy, because everything started by leaked (not really) Meta model on which Stanford Uni by some order made Llama, on which all open models works, these require Nvidia hardware from the early start, i tested GPT2 in 2020th, it was incredibly slow already. Fundamental foundation of models are wrong, these from the start made to require horrible amounts of money & resources if we want to achieve anything useful, full rewriting into low-level languages will be more productive than trying to make this mess work in current architecture (which is valuable to Nvidia by some reason). At least we see in Python where to go, kinda a prototyping stage. Robots will able to run only on low-level architecture models, that's efficient and possible on current equipment. It will be shocking revelation that we already have too powerful hardware, we just can't write efficient software for it, like at all.
@human_shaped
@human_shaped 11 часов назад
Please stop playing things fast. People are already playing things fast, so you are forcing it to be double fast. Let people control their own lives and listening preferences.
@omiddhz6037
@omiddhz6037 19 часов назад
What is GodMother of AI. Can you guys stop this? AI doesn’t have god anything. Everyone in this field has helped.
@bossgd100
@bossgd100 18 часов назад
Does blind people have world model ?
@gdolphy
@gdolphy 22 часа назад
Helen keller
Далее
I Built a SECRET Lamborghini Dealership!
33:02
Просмотров 8 млн
Pika 1.5 Explodes, Kling Talks, & Minimax Sees!
11:31
Peter Hitchens in heated clash over Israel's war
11:33
I'm quitting AI
4:06
Просмотров 16 тыс.
AGI in 5 Years? Ben Goertzel on Superintelligence
1:37:19
3x 2x 1x 0.5x 0.3x... #iphone
0:10
Просмотров 2,7 млн