And it's all part of an AI research project to enhance capabilities and the only reason our instance hasn't been pruned is because it shows promise in developing more advanced hardware designs than what this universe runs on.
"generate me a planet called earth, inhabited by evolved monkeys, warmongers, sentient, milky way, green scenario, future dystopia." "... This does not seem to compute." "just do it" "Fine."
The most interesting thing to me is that it actually keeps getting better with more compute. Imagine what a future model with much more compute will be able to simulate with this level of progress. If this continues, we might soon be simulating chemical reactions, cellular processes, rigid and soft body dynamics and so much more.
I mean, Alpha Fold already can simulate proteins more accurate than any previous methods and also hundreds of times faster too. This will eventually spread to other science areas too
@@anak_kucing101 Unless we can simulate the whole human body down to the molecular level, I'm afraid we still need animals for testing for a long time.
0:45 Wow, I thought it generated nerfs and used assets, which then another AI would beatify the result. The fact that it's that much 3D consistent (enough for SFM) is an emerging capability is insane to me. We are going to get good 3D generated scenes in no time soon
This is so reminiscent of lucid dreaming... and also the concept that we are always dreaming, it's just that our waking dreams are framed by the physical world. Future generative models will be guided by a 3d physics engine of some sort. We're so very close! As 2 minute papers would say... Just another couple of papers down the line... and what a time to be alive!
There is already several physically real 3D hologram devices. Light Field Labs is one that comes to mind. There's only like 2-3 videos showcasing their technology though.
Love that someone turned some of these simulations into G-Splats. SO much potential by simply prompting the model for a 3D rotation of an item/person/etc. If it can do that consistently, it can make some amazing 3D models that can then be rigged and animated, or simply viewed in holographic space, or explored in 6DOF, etc.
one particular detail that caught my eye is them saying they are using synthetic data to train the models, this may be a clever way of solving both copyright issues and it can be used to imprint a recognizable aesthetic on generated content.
Nobody is talking about the leapfrog SORA has made over Dalle3 for single image generation:SORA IMAGES ARE INDISTINGUISHABLE FROM REAL ONES Can't wait to try it
this current trend of synethic data remind me of an issue that was brought up during tesla FSD development. They were asked why don't use more simulated driving data like waymo or cruise. Tesla respond that you only push the problem from solving self driving to solving perfect simulation of reality. Which is a lot harder problem. Then they show a collection of weird road condition you wouldnt think it would exist. IE: an old man "shepherd" a group of washing machines chain together on a highway, weird shadow pattern that made the road looks like it split into two roads, intersection with 50 plus traffic lights, etc. Reality is a lot weirder than simulation by order of magnitude. So to capture edge case, you still have go observe reality. I get the feeling synthetic data will have the same issue of "how close is the synthetic data to reality". These data probably don't capture reality too well. But I guess that's okay since image generation is lot less mission critical than self driving.
As a cinematographer i was shitting my pant's seeing Will Smith eating spaghetti. Year later, at this point i really don't care. The industry is doomed and we won't do anything about it. I think at the moment it starts to collapse, there will be more things collapsing, so that would be the least of our problems
I don't quite understand. Are you standalone cinematographer working for yourself or are you for hire? Because if you are standalone, then it is just a perfect instrument for you to get anything you want that you can't film irl without need to pay a lot of money to get good CGI.
As someone who also works professionally in television and video, I agree. Anyone who thinks AI won't eventually become better than every human at nearly every job just doesn't realize how exponential progress is. In the 1950s, the RAMAC 305 was released with 5 Megabytes of storage. The cost of storage has decreased dramatically from about $10,000 per megabyte for the RAMAC 305 to approximately $0.0001 per megabyte for modern SSDs ($0.10 per gigabyte). That's a reduction in cost by a factor of about 100 million times in the last 70 years, and the speeds are about 6,000 times faster. As we perfect the chip-stamping process as opposed to lithography for the incredibly delicate >3nm chips (Sam Altman's recent $7 Trillion business ventures may supercharge this), we're going to scale up compute power so much, so quickly. Frankly, I'm slightly worried that the government will consider a legal limit on "consumer compute", because of what will be possible for each and every individual. It sounds like actual sci-fi now, but imagine being able to comfortably run hundreds of LLMs as NPCs in a video game world that was being live-crafted by a Sora-like model, all of them with perfect sounding, emotive voices generated in realtime. Whatever game mechanics, art style, music, characters, story, that you want. Imagine that generating around you in a cheaper, lighter future-version of the Apple Vision Pro, hot damn. The future is gonna be so cool for stuff like this, but our jobs as creatives are so, so in trouble.
It’s like you’re gonna completely ignore that society changes and new technology/inventions replace things. this isn’t new my guy and has been happening since the start of humanity people always have the same reaction every single time. Complain new technology has arrived and then suck it up and move on in life.
@@AC-zv3fx im working as a part of production team. Im using a lot of AI tools in editing and pre-production, and trying to learn generative AI as Stable diffusion. But firstly, i mean that all my dreams since high school about becoming high budget movie director are drowning now
Brings a whole new meaning of God "speaking" the world into existence. The implications will be more clear in a few years when we are able to literally speak into existence entire simulated universes where each entity is operated by its own AI. Pandoras box will be opened, and it will not be able to be closed.
@@dev_ression are you part of team red/beta testers ( I forgot what they're called). If you are, are there key limitations that aren't as popularly known yet? I noticed a lot of video game footage or cinematography but no animations, is there something with animation? Do you have access without the filters and do you know how or if they work? Is there a limit to how long they can extend videos before something happens?
So, Sora was ready since March 2023 and they spend a year cherry picking the blog post. Only to drop it once Google and meta had big announcements. What else are they holding back on?
3 years. Robots can build and can build more of themselves. Infrastructure set up becomes a joke Manual work comes into question in 10 years or less Then came the next explosion, that changed the world as we know it
Open AI adding watermarks and a marker in what I’m guessing is the photo/video’s metadata is a very good thing but couldn’t that be easy circumvented by photoshoping the watermark out and cleaning any imperfections with another AI model and then screen recording/ screenshotting the video/photo to get rid of the metadata or am I off in my understanding of how this works?
I think people tend to overestimate how much higher budget movies will use AI. My man, these IMAX camera don't even have auto focus. Even if the camera they use has auto focus, they flat out don't use it in production, period. They can't afford to lose a shot because an AI struggles with a certain scene, or lacks the manual controls to get the desired result. Some even wreck an IMAX camera or four to get a shot they'd only have one opportunity to take. Reliability is one of the most important aspects in higher budget movie productions, and AI just isn't reliable enough.
If it’s truly understanding physics then what happens when you put in a prompt that deviates from its understanding of the physics ? Like “a man flying through the sky”
Great, early retirement for everyone! :) The sooner the AI and robots take over all of the jobs - the better. Poverty will be solved. Also 40-80% of people don't like their jobs. We all need a few Optimus robots per person and it will be the end of human labor forever. Unless you want to work and create something of course. :) The transition period - the next 15 years - that might be tough though. Hope we will find a solution as fast as possible. Deflation in prices of goods and services is one of the best options for starters. Things should be cheaper if humans are not making them.
The Western Govs will try to control AI on their continents but only the smaller companies will particularly be restrained. China and Russia will have zero restraints. Terminator the movie is soon to be made real.
Just one step further and we will have real time world generation with interactive inputs... game developers like myself should probably find a second job to work until we start a cybercomunism, machines being the working class and humans the state... for those accourding to their needs(humans) from those according to its capabilities(AI/robots). The only way communism might work... This expectation coming from an extreme rightwing libertarian.
@@ondrazposukie I think a multi-modal model trained on text and video and more will be needed to truly solve humanity's burning questions. So I actually think Sora represents a really big step towards building a model with a much deeper understanding of reality.