Тёмный

Apple M3, M3 Pro & M3 Max - Chip Analysis 

High Yield
Подписаться 42 тыс.
Просмотров 299 тыс.
50% 1

In-depth analysis of Apple's new 3nm chips: M3, M3 Pro and M3 Max. Silicon deep-dive, die-shot analysis and a closer look at CPU, GPU, NPU and the TSMC N3B process node.
Follow me on Twitter/X: / highyieldyt
Support me on Patreon: www.patreon.com/user?u=46978634
0:00 Intro
0:47 M3 Silicon Analysis
4:17 M3 Pro Silicon Analysis
6:26 M3 Max Silicon Analysis
8:25 Why is the M3 Pro a downgrade?
10:47 NPU deep-dive
12:16 CPU deep-dive
13:17 GPU deep-dive
14:44 GPU architecture / Apple family 9 GPU
16:22 TSMC N3B Process Node
18:44 Wrap-up

Наука

Опубликовано:

 

15 май 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 900   
@silverc4s146
@silverc4s146 5 месяцев назад
I watched the entire 20 minutes. As a long retired chip designer, this 90B active device world is incomprehensible, but I enjoy following anyway.
@eth_saver
@eth_saver 5 месяцев назад
this channel is so good, idk why it doesnt have more subs, all videos are excellent
@JosPoortvliet
@JosPoortvliet 5 месяцев назад
Yeah it was an interesting video. I am looking forward to a deep dive on the cpu architecture! The p and e cores for 25% more transistors, what are they used for?
@MFoley-tv3zh
@MFoley-tv3zh 5 месяцев назад
Fascinating!!!
@hackerlicher889
@hackerlicher889 5 месяцев назад
Can you guide me as I also aspire to be a chip designer
@jimgolab536
@jimgolab536 5 месяцев назад
Same.
@DeadCatX2
@DeadCatX2 5 месяцев назад
As an embedded and FPGA engineer, CPU design like this has always felt like the major leagues. Watching this video must feel to me like watching a sports game with good color commentary feels to a typical American. Thank you for producing this
@yoomy_gums
@yoomy_gums 5 месяцев назад
As an autodidactic hardware designer. I feel related, why the transistor count increases on the cores on next models? what new is implemented? how is everything working in harmony? how it schedule tasks to all parts on the computer without struggling. Me and everyone else are NOT capable to completely understand how this works. An engineer involved on Apple knows way better than us, but still it’s almost impossible that he know all of this, so insane. That’s what I think that these machines are miracles, almost magical! but actually it’s the hard work of many incredibly smart people. So almost all is out of my ligue, despite the 4 years of researching that I dedicated to the topic. 🤯
@mikafoxx2717
@mikafoxx2717 4 месяца назад
​@@yoomy_gumsYou basically have to start with the history of architecture, and the ways they improved on the previous generation, like say the 486 vs pentium vs pentium pro. Yes, these are arm, but these days instruction sets make far less difference than the architecture fundamentals.
@NothingXemnas
@NothingXemnas 4 месяца назад
Same. I love all the intricacies of GPUs and CPUs, even ones with included graphics in the same die, as Intel and AMD still do in some chips. But massive dies with everything embedded into it (REALLY making "System on a Chip" mean what it says) is just incredible. Of course there is more wasted material (making such massive single dies also means more dies having defects), but it is still such a respectful "balls to the wall" approach! This is why the Snapdragon X Elite is so exciting to me!
@starkead3087
@starkead3087 Месяц назад
@@yoomy_gums I guess the transistor count increased into the m3 with the ray tracing
@yyyy-uv3po
@yyyy-uv3po 12 дней назад
@@yoomy_gums "why the transistor count increases on the cores on next models" For starters, they went from ARM-v8.5 to ARM-v8.6. It also depends on which extensions they actually implement.
@HighYield
@HighYield 5 месяцев назад
Sorry for the long wait, the video got longer and longer the more I worked on it... Let me know if you enjoy these (very) deep-dives, or if it's too long/detailed for you. PS: the dynamic caching doesn't have anything to do with the system memory, but it's about the on-chip GPU memory. The whole GPU seems to be complete game changer, something a lot of ppl seem to have missed. This might very well be the most advanced GPU architecture right now and it will take a while until we see it's full potential.
@MrXeuke
@MrXeuke 5 месяцев назад
thanks for your hard work!! 🙏
@giordano7703
@giordano7703 5 месяцев назад
Love these deep dive videos!
@EmreHepsag
@EmreHepsag 5 месяцев назад
I would love deep dive long detailed video. Great job!
@LucianoBelotto
@LucianoBelotto 5 месяцев назад
I enjoyed it, thank you! Many RU-vid videos reviewing the M3, but no other video I found like this one
@debojitmandal8670
@debojitmandal8670 5 месяцев назад
But how come it's still not as good as an rtx 4090 or even 4080 minus the power draw plus i thought ray tracing by nvidia was a major architectural innovation. Plus if nvidia already did this minus implementing with the registers it just means nvidia had this before apple
@c11p
@c11p Месяц назад
Easily the most informative M3 breakdown. Kudos.
@zkeltonETH
@zkeltonETH 5 месяцев назад
Still watching and man, these deep dives are so fascinating to learn more about silicon design and engineering in our current era. Absolutely amazing work!
@denumerable
@denumerable 5 месяцев назад
Here with you. Very fascinating.
@DevDunkStudio
@DevDunkStudio 5 месяцев назад
I'm watching all the way through :p
@jsl8461
@jsl8461 5 месяцев назад
Based on benchmark tests, the M3 Pro chip doesn't suck the way its specs on paper would suggest (ie, it has a lower transistor count, lower P-core count, and 1 less GPU count than M2 Pro). It performs similarly to the M2 Pro chip on most tests (on a few tests it performs modestly better), but it seems to use a lot less power as the M3 Pro laptop has a significantly better battery life even with its 100 nit brighter display. Market segmentation is a consideration but it looks like Apple actually delivered a product that is as performant (and in some ways better) than the previous generation while offering significantly better efficiency at the same time. Quite different from the M3 chip, where most of the improvements over M2 seem to be on performance rather than efficiency.
@Fractal_32
@Fractal_32 5 месяцев назад
Who wouldn't watch your entire breakdown of apples silicon? Personally I enjoy how this channel focuses on the less talked about features of hardware design, it really makes you understand how much a company can care or not about a product they are launching into the market. Keep up the great work, I cannot wait to watch more of these breakdowns in the future!
@Frytech
@Frytech 5 месяцев назад
I'm certainly watching your every video till the end! Just recently discovered your channel and it's a godsend in terms of amazing in-depth explanations of how exactly all those performances and features are achieved and realized on the silicon level! I've always wanted for someone to explain things like that, like on a truly low level - in terms of hardware - literally talking about transistor counts and how it's all allocated on a chip, designed, interconnected, etc. Thank you so so much for what you're doing on this channel! Keep these amazing videos coming!
@ancientsword
@ancientsword 5 месяцев назад
It doesn't matter how transistors change, I just know that the knife skills are superb and they are getting expensive again.
@ikarosav
@ikarosav 5 месяцев назад
Great video I was really looking forward to this one! On the "Dynamic Caching" in the new shader core (aka. register file + image block + group shared = L1). You've watched Apple's video already so I'll try to add some additional practical context to why it's important: It doesn't require new shaders to be written, old shaders are forward-compatible with taking advantage of this feature, however most shaders were indeed written with the limitations that came before it, thus the big advantage would only be felt on shaders that had low occupancy previously and can now maybe have higher occupancy. A lot of shaders are written with say reading a bunch of buffers, and reading a bunch of textures at -some point- typically early, and at this point they'll greatly benefit from high occupancy to hide latency and avoid stalling. But typically, later in the shader, you do a bunch of math that require -a lot of registers- for a short time, and this spike in register count in the old method required that the whole shader demand many registers the whole time, even though for fetching buffers and textures it only needs enough to store the read results in just then. So the benefit here is that you get to have low register pressure when you need high occupancy early in a shader to hide memory latency, and later during "just math" where you don't need occupancy to saturate the math you can now go nuts with registers. Having the freedom to use many registers can make for better algorithms that can take advantage of large amounts of registers without worrying about hurting memory latency in another part. It also provides freedom, you don't have to spend a lot of optimization time getting a magical register count, the shader core does it for you (almost, you still need to make sure you don't need many registers at the time of doing these memory reads), and most importantly, you can now make dynamically branching uber shaders that don't trash your register file usage! Previously we've always had to make many shader variants for specialized cases and compile them either at build or run-time, because a huge shader with tons of branches would have register pressure as bad as the worst case "everything is on" scenario, well now the register pressure is dynamic based on what's enabled! I probably got some parts wrong but I think it's really interesting how much having an L1 cache changes for shaders.
@BurritoKingdom
@BurritoKingdom 5 месяцев назад
I believe the new shaders he's talking about is mesh shaders. Which do have to be completely rewritten. It is why in the PC space there was an up roar when Alan Wake 2 was released, it's the first major game to use mesh shaders and made older GPUs obsolete since it's not forward compatible.
@ikarosav
@ikarosav 5 месяцев назад
@@BurritoKingdom ah yeah my bad. just wanna add that mesh shaders and amplification, as well as ray tracing have been part of metal for some years, although internally running in a software implementation, so developers have been able to write tech that took advantage of these for some time pre-emptively. I know octane used the ray tracing api for a while before the agx9 came out, but yeah i don't know of anything that has taken advantage of the mesh shading api so that indeed would be novel to see used now.
@hishnash
@hishnash 5 месяцев назад
While it doe snot require shaders to be re-writen (this is very nice) you can get a good bit more pref by making changes. It is common to break up long running shaders into smaller shaders were each of these smaller shaders has a more constant register/threadgroup usage. This adds some overhead as you dispatch extra shaders but on older gpus it requests in better avg occupancy as the parts of your application with lower pressure can run higher ocupancy than if you just dispatch a single longer running shader that has some very high peak local mem or register usage. This new dynamic register/local/cache system means you can now just stitch all these shaders together (reducing the dispatch overhead) so you can now have much longer running single dispatch shaders without the occupancy hit that this has on most other gpus.
@gimmedaloot754
@gimmedaloot754 5 месяцев назад
Love these long deep dive videos. When executed well they provide extraordinary value. Time is valuable and this video did not disappoint. Keep up the great work!
@mrfin
@mrfin 5 месяцев назад
You have a gift for articulating these subjects. I have zero chip background but was easily able to follow through to the end.
@jimgolab536
@jimgolab536 5 месяцев назад
I watched the whole thing and subscribed. This was a very nice level of analysis for me, and I think you did a great job of overviewing the changes. It seems to me that this gen is taking to heart one of the original RISC tenets, where spending transistors on caches (vs cpu, etc) is a huge win. The tricky part, also from RISC heritage, is that you have to have compilers that can take advantage of the opportunities for caching (and the exposure of opportunities for parallelism). I enjoyed your video a lot. Thanks.
@schwartn
@schwartn 5 месяцев назад
I’m a high school computer teacher and I played it for my students. My students love to keep up with the latest chip news. Thanks for sharing!
@MoonshineOctopus
@MoonshineOctopus 5 месяцев назад
I’ve been waiting for this deep dive! Thank you!
@kristiandilov5249
@kristiandilov5249 5 месяцев назад
this is exactly the type of breakdown/content I was looking for. Really loved watching it, I need to deep dive into each topic and learn more 😀
@Cofenotthatone
@Cofenotthatone 5 месяцев назад
I always watch your videos from the beginning to the end, since your content is excellent. Thank you again this time.
@bjesuiter
@bjesuiter 5 месяцев назад
I love your breakdowns, it’s so fascinating! Watched to 11 min now and aiming for the full 20 min!
@balegdeh.
@balegdeh. 5 месяцев назад
im in LOVE with these deep dive videos and i dont even feel like i just watched a 20min video, please keep this video style as long as you can😭
@papsaebus8606
@papsaebus8606 5 месяцев назад
It’s remarkable how much effort you’ve put into producing and researching this, keep it up! 👏
@HardwareScience
@HardwareScience 5 месяцев назад
I watched the entire video and I don't think 20 min is particularly long for this kind of content, you did a great job 👌👌
@timl2k11
@timl2k11 5 месяцев назад
Watched it all, both an overview and a deep dive and I think you nailed the pacing.
@Fidalailama
@Fidalailama 5 месяцев назад
Ive never seen any of your videos before but thank you for the depth and relatively simple explanations. Great video!
@TensorXR
@TensorXR 5 месяцев назад
I watched the entire 20:12 minute video, It was very informative. I personally appreciate these deep dive technical analysis type videos, I learn a-lot more about semiconductor engineering and about the hardware we all take for granted. I am deeply fascinated about where the industry is headed with these process nodes and there optimizations.
@stefanbuscaylet
@stefanbuscaylet 5 месяцев назад
I’m in a somewhat similar industry trying to rebalance our product line portfolio and create distinct segmentation and know how many meetings and difficult it is. Im sure there was a ton of stress by folks at Apple (and thus a ton of meetings) when they relanded the M3 Pro calling it a “downgrade”. I can see the product planners and engineers arguing in my head. Watched the whole thing and subscribed. Thanks for doing this.
@TheGarrison89
@TheGarrison89 18 дней назад
Love watching your deep dives all the way through, I learn a lot. Thanks!!
@alexanderstohr4198
@alexanderstohr4198 5 месяцев назад
Thanks for that one - its length fits quite well to the topic.
@andrewbulloch
@andrewbulloch 5 месяцев назад
Superb deep dive, incredible detail you're covering here. Silicon has come a long way from my early days in the 90's in semi-conductors.
@beaudanner
@beaudanner 5 месяцев назад
Apple silicon has been such a fantastic leap that I have to remind myself that my M1 Pro is _so_ good that I shouldn't have any need to upgrade. Despite the amazing performance gains of later chips I'm sticking to my 5-8 year upgrade cycle
@robblincoln2152
@robblincoln2152 4 месяца назад
Same. That gives me time in the down years to upgrade my iPad!
@salty4
@salty4 5 месяцев назад
always love your long deep dive videos
@Gwrod.s
@Gwrod.s Месяц назад
I understood like half of the content but you made it super easy to digest! Great job!
@rainerzufall9881
@rainerzufall9881 5 месяцев назад
I always watch your videos from start to finish, as they are high quality content!
@woolfel
@woolfel 5 месяцев назад
I really wish Apple would give more details about the dynamic caching stuff. I read the patent filing and it looks interesting. I was hoping the new GPU design is optimized for training ML. Watched the whole video. Hopefully as more people analyze the chip, you can update and identify where the dynamic caching logic sits on the Max chip.
@pham3383
@pham3383 5 месяцев назад
if you do ML,just stick to CUDA...
@Demopans5990
@Demopans5990 5 месяцев назад
@pham3383 Or wait for the day Apple stops being Metal exculsive and adopt something like OpenGL at a hardware level
@HyperMario64
@HyperMario64 5 месяцев назад
@@Demopans5990 I think you meant Vulkan ;)
@SuperWookie01
@SuperWookie01 5 месяцев назад
Loved the video especially the visualisation of where everything is on the chips. Watched every second of it.
@EricCanton
@EricCanton 5 месяцев назад
Watched till the very end! I love technical deep dives like this, looking at chip pictures, talking about process nodes.. awesome vid 😎
@HighYield
@HighYield 5 месяцев назад
Glad you enjoyed it :)
@m.s.psrikar8681
@m.s.psrikar8681 5 месяцев назад
Just completed watching this video. As a current chip designer, absolutely love your content and this video in particular was very well done. Would like to see more deep dives like this video.
@HighYield
@HighYield 5 месяцев назад
More to come!
5 месяцев назад
Don't worry, 20 minutes for such a subject is definitely not too long. Would be great to go even more to the depth. Anyway, great and informative video! 👍
@ogheneochukooputu9725
@ogheneochukooputu9725 5 месяцев назад
Nice! I watched the entire video and even returned to watch some parts. Thank you for the comparison.
@surftec
@surftec 5 месяцев назад
Fantastic video. Thanks for the detailed review. Length of video is fine, detail was exceptional and delivered with enough speed it wasn’t at all boring. Some drag things out so much I fall asleep listening them slowly waffle. Yours was perfect for me.
@cyan_aura
@cyan_aura 5 месяцев назад
Thanks for such a detailed analysis. I find these deep dives really interesting and I'm pretty sure many others would agree too. Would surely love to see more of these in future. Cheers!
@HighYield
@HighYield 5 месяцев назад
More coming for sure. Thanks for your support, much appreciated!
@KellyWu04
@KellyWu04 5 месяцев назад
I love your channel. It’s more or less unbiased. Praising where praises are deserved and criticizing when it’s due. I also love the detail-ness of your content.
@LenardGunda
@LenardGunda 5 месяцев назад
Watched from beginning to end, thank you for the deep dive. Very informative!
@swlak516
@swlak516 5 месяцев назад
Loved the video. Watched the whole thing. Always look forward to your deep dives.
@MichaelSinz
@MichaelSinz 5 месяцев назад
The number of transistors on a single chip is breaking my brain. But then, the last chip I was involved with was a long time ago (I ended up going into more software work as we were building software to test and validate our chip designs and that ended up being where I found the love of software engineering)
@antenedilbert7191
@antenedilbert7191 5 месяцев назад
This is the most intensive yet easily informative piece of Video. I'd say its not long. It's full of info that it never felt long. Lets see what Qualcomm does with their designs after their new acquisition.
@joeldelacl4206
@joeldelacl4206 5 месяцев назад
Great video man... Love learning more about chip design and watching new products from a design perspective. Keep the great work!
@moinulhossain7468
@moinulhossain7468 5 месяцев назад
I don’t remember when was the last time I watched a 20 minutes video at a stretch. Great job making it information rich and right on the money. Enjoyed the video and how you compiled it. Please carry on.
@paul94040
@paul94040 5 месяцев назад
Excellent video. I watch a lot of processor analysis videos and product teardowns, and yours is one of the best I've seen. Interestingly, there's a bunch of initial M3 product comparison videos that are reporting M3 as a failure because their benchmark software doesn't take advantage of the improved architecture that Apple has delivered. I would love to see your analysis on the Apple audio chip improvements and the Closed Loop Controller used in the camera system...
@David_Best
@David_Best 5 месяцев назад
Still watching
@khemw.9510
@khemw.9510 День назад
Appreciated the deep drive, you compared and explained them clearly. Keep up the great content.
@kwameYuTu
@kwameYuTu 5 месяцев назад
This is the in depth analysis I have been waiting for!!!! You must be doing something right as I had never heard of your channel before this video. Keep up the great work!
@Dominic416_
@Dominic416_ 5 месяцев назад
Keep ‘em coming!
@daemontus
@daemontus 5 месяцев назад
Watched 'till the end :) I'd very much like to see a price estimate for these chips on N3B. Everybody keeps complaining about the SSD and RAM prices that Apple is charging, but my guess would be that the high-spec models are actually *heavily* subsidizing the price of the low-end configurations. A complete laptop with 96B transistors for $3500 vs. a 4090 GPU with 76B for $2000-3000 is very interesting.
@robblincoln2152
@robblincoln2152 4 месяца назад
This is it exactly! There not selling parts but performance envelopes.
@dmanistheshit
@dmanistheshit 5 месяцев назад
Hey love the deep dive videos the longer the better :)
@CraigBlack123
@CraigBlack123 5 месяцев назад
Watched it all the way through. Your deep dives are excellent. Thoroughly enjoyable. Much appreciated.
@juliandelphiki7381
@juliandelphiki7381 5 месяцев назад
Great video. First one I saw on your channel and you have earned a new subscriber. As an avid gamer I’ve been interested in seeing Apple push the GPU side since the m1 came out and they started using the same underlying hardware across Mac and iPad. Curious to see both apples support for this segment and adoption by studios.
@IakobusAtreides
@IakobusAtreides 5 месяцев назад
Exceptional content
@brxrmr
@brxrmr 5 месяцев назад
Still watching and enjoyed the deep dive. It flies over my head but still interesting to me and learning a lot!
@adamw.7242
@adamw.7242 5 месяцев назад
Commenting for the long watch survey. Your content is great; always finish the deep dives! ❤
@rangerBlu
@rangerBlu 5 месяцев назад
Came to hear your thoughts on M3 Pro - great breakdown of key components! Although I've read many complaints about the M3 Pro being a "downgrade", I don't think it's a big deal. I think your analysis is sound although on another site they speculated the new Pro design may have been due to issues with achieving an adequate supply of M2 Pro chips through binning of the M2 Max alone which resulted in cost and supply issues. Nicely done - thanks!
@ALifeWithoutBreath
@ALifeWithoutBreath 5 месяцев назад
Still watching at 10:50 and watched to the end. As a diver the wording makes me chuckle but this dive was neither too deep nor too long. 😉🤿 Last time I got into the nitty gritty of Apple Silicon was when the M1-Series was released. Now that the 3nm chips are out I'm catching up on the developments. It's neat to see an analysis of the actual dies. Good work! ☺
@KnuffinMight
@KnuffinMight 5 месяцев назад
Ive watched this video like 3 times this stuff is so interesting. I appreciate the level of effort you put into this video.
@m89hu
@m89hu 4 месяца назад
same! I hope his next video is an hour long 😅
@harounelmir2632
@harounelmir2632 2 месяца назад
Watched all the way through, time went by like a breeze with your explanations, cheers
@JerrodLandon
@JerrodLandon 5 месяцев назад
Wow thank you I was wanting someone to do a deep dive into the M series, chip sets. I don’t know really anything about how all this works but absolutely love hearing how things are made on a very detailed level. Thank you so much for the undoubtedly, very hard work and research you put into this!
@goldrunner18725
@goldrunner18725 5 месяцев назад
I watched the whole thing. You're, at the moment, the closest thing we have to AnandTech that I can think of. I'm a big fan of Chips & Cheese as well, but those folks sometimes take a long while before coming around the lastest chips like AT used to do. Thanks for confirming the A9 Family GPU is entirely new, some did not believe at. As well as having the common sense to see the M3 Pro is not a downgrade - is its own custom design chip aiming for something differnet than the M3 and the M3 Max.
@Vilverin
@Vilverin 5 месяцев назад
This was super insightful! Though perhaps it wasn't just sales/profit targets driving the differentiation between the M3 Pro and Max. The Max is really over-the-top for a laptop CPU/GPU and runs quite hot and loud (almost Intel-era-loud) especially in the 14 inch MBP, much more so than the M2 Max did. Battery life is also quite modest relative to other Apple M machines. So the M3 Pro serves as a new middle ground for people who want more performance and features than what the base M3 offers, but don't want the battery life and noise compromises that come with the Max, as insanely powerful as that chip may be.
@daniloamchagas
@daniloamchagas 5 месяцев назад
Another excellent video! I always go all the way to the end in your videos. Keep up with the great work!
@blazemasterb4018
@blazemasterb4018 5 месяцев назад
I love your Deep dive Videos i've watched all of them till the end you know your stuff very entertaining keep up the good work!
@MrDejvidkit
@MrDejvidkit 5 месяцев назад
This is great video!
@albuslee4831
@albuslee4831 5 месяцев назад
This was the best analysis I've seen so far on Apple M3 chips, especially on the M3 Pro chip and it's design purpose. It makes so much more sense and was constructive, objective analysis than 99% of other reviewers simply bashing on Apple without clear explanation.
@user-kt2kz5qg4z
@user-kt2kz5qg4z 5 месяцев назад
Wonderful video. The whole thing was interesting and well constructed. I especially liked the summary at the end.
@markusklingsiek2366
@markusklingsiek2366 5 месяцев назад
Thanks for that very comprehensive video. I watched it until the end and it was a very interesting deep dive. Thanks for that!
@robertpearson8546
@robertpearson8546 5 месяцев назад
The M3 NPU may be multiplexing the processing units, time-sharing 8 NPUs to get a (slower) 16 NPUs group. Like switched capacitor op-amps allow multiple poles/op-amp.
@jasonjames2778
@jasonjames2778 3 месяца назад
Would that not take a hit to the TOPS? Just curious about what you think regarding that. Greater fan in and fan out of signals on the silicon does effect the throughput performance right?
@robertpearson8546
@robertpearson8546 3 месяца назад
@@jasonjames2778 Of course it would. But TOPS and MIPS (Meaningless Information Propagated by Salesmen) do not measure actual performance. Since "operations" and "instructions" have no semantic meaning, the numbers have no semantic meaning. They are just a measure of clock speed, not performance. I was lucky enough to take a course in quantitative performance evaluation. The examples in the text were not imaginary problems but were case studies on real problems in actual computer centers. One example was a system with a fast and a slow disk drives. The vendor suggested doubling the CPU speed. That resulted in a 3% increase in throughput. Moving some files from the fast disk to the slow disk increased the throughput by 100%. Compare the 68000 and the Novix microprocessor. The 68000 is clocked at 10 MHz so it has a high MIPS. But the Novix's throughput is 3 times that of the 68000.
@sloanNYC
@sloanNYC 5 месяцев назад
Fantastic work! The M3 definitely looks like maybe the TSMC troubles with the new node may have forced Apple make some tradeoffs they didn't want. I wonder if the advances in the GPU cores is also changing the needs of the NPU compute types... can more AI workloads move to the GPU with the new architecture?
@tonyburzio4107
@tonyburzio4107 5 месяцев назад
TSMC is launching a new 3nm process, from lessons learned, in February.
@andyxiao8913
@andyxiao8913 5 месяцев назад
watched the entire video, really good analysis. subbed.
@Epicgamer_Mac
@Epicgamer_Mac 4 месяца назад
I’m still watching, and I’m planning to watch the entire thing! I appreciate the detailed discussion and I find it very fascinating.
@nematodes5113
@nematodes5113 5 месяцев назад
These deep dives are great - I did watch the whole thing (though at 2x speed). A lot of the interesting architectural details just aren't captured in specs like core and transistor counts.
@christopherleubner6633
@christopherleubner6633 5 месяцев назад
Yup I noticed that immediately regarding the NPU subassembly. It looks like they built bigger arrays in each core vs just more cores. Given their function, it would be a smart move as it would allow for more efficient process optimization for larger and more complex routines. ❤
@jonanddy
@jonanddy 5 месяцев назад
I just tried the 2x speed idk how you understand anything
@Theinatoriinator
@Theinatoriinator 5 месяцев назад
@@jonanddy You get used to it, eventually you get an extension and watch stuff at 3x or 4x speed and still take it in. It's crazy how much time you can save by doing that.
@The_Dark_Hood
@The_Dark_Hood 5 месяцев назад
If I had to take a guess at the why the NPU hasn't been taking up much more die area I would say that it's likely because they haven't found an architecture they love that doesn't take a lot of power. Instead they're keeping with their current solution and upgrading with space and power being the driving forces, especially since that's Apples Silicon Architecture design default. It's also hard to tell if it's actually that much more powerful or not between generations since TOPS doesn't tell you much about an NPU. For example, is this combined int8 and float32 TOPS? Did they add support for smaller int4 TOPS? How does is the NPU caching affecting this number?
@Jamieibrahimm
@Jamieibrahimm 5 месяцев назад
This is from Apple’s ML website “The first generation of the Apple Neural Engine (ANE) was released as part of the A11 chip found in iPhone X, our flagship model from 2017. It had a peak throughput of 0.6 teraflops (TFlops) in half-precision floating-point data format (float16 or FP16), and it efficiently powered on-device ML features such as Face ID and Memoji. Fast-forward to 2021, and the fifth-generation of the 16-core ANE is capable of 26 times the processing power, or 15.8 TFlops, of the original.”
@R2DHue
@R2DHue 5 месяцев назад
With your power consumption theory duly noted, even if it were true that Apple wasn’t happy with its NPU design, there’s no reason why they couldn’t have increased the NPU core count (apart from power consumption - if true). I surmise Apple wasn’t ready for LLM & AI - OR - GPUs are more suited to these tasks than NPUs after all. That’s certainly what Nvidia found out…
@boshi9
@boshi9 5 месяцев назад
Having tested this, same models on Apple's NE are significantly faster compared to the GPU while consuming negligible amount of power (it's almost as if nothing is running at all). The caveat is that not every model can be run on the Neural Engine since it's very specialized by nature (in which case Core ML automatically falls back on the GPU), but there's no power efficiency issue even if they decided to scale it up. I suspect that Apple simply believes that its performance is sufficient for the current tasks they have in mind.
@SimRacingSK
@SimRacingSK 5 месяцев назад
long videos i love it, watched it all keep m coming
@FLUFFSQUEAKER
@FLUFFSQUEAKER 5 месяцев назад
Don't worry about 10+ minute videos, i could watch an hour of this!
@tobi6758
@tobi6758 5 месяцев назад
I'm still puzzled why apple did not introduce a Server / AI focussed chip yet, seeing how Nvidia, AMD and Intel make massive amounts of cash with that. Something like a M3 Server edition with either a ton of CPU or GPU or NPU cores for their Mac Pro. The energy efficiency would certainly make them a great competitor to the already mentioned.
@little_fluffy_clouds
@little_fluffy_clouds 5 месяцев назад
It’s because Apple isn’t in the server market and isn’t interested in entering that market
@Demopans5990
@Demopans5990 5 месяцев назад
Not really. AMD's newest Threadrippers somehow have better performance/watt than m3 in some cases, despite gulping more than 300 watts of power at max speed. Guess that is what 3 figure thread counts will do. Also, Apple will be effectively required to support open source software (and open hardware standards), which doesn't sound like an Apple thing to do
@GarryMah85
@GarryMah85 5 месяцев назад
Love these deepdive vidso. I think these chip are a modern wonder of mankind, its mind blowing how we are able to design and produce something with 92 billion switches
@MarkEvans5
@MarkEvans5 5 месяцев назад
Well done. 👍🏼 I also watched the whole video, quality information. Thank you!
@VideogamesAsArt
@VideogamesAsArt 4 месяца назад
This is exactly the high quality and great content I learnt to expect from this channel. It seemed like a 5 minute video to me. Right at my knowledge/interest level, incredibly educational and engaging/entertaining. Looking forward to all your videos
@torb-no
@torb-no 5 месяцев назад
I’ve kind of been thinking that in some ways the base M2 was underpowered (specifically lacking the extra display controller), yet the M2 Pro was maybe overkill. I’d prefer the M3 also have an extra display controller, but making a smaller M3 Pro is also one way to do it. The M3 Pro seems like a pretty good solution it you want a MacBook that runs cool. And yeah, watched the whole video. Very fascinating. Thanks for making it!
@alb.1911
@alb.1911 5 месяцев назад
Thank you. 🙏
@RomPereira
@RomPereira 5 месяцев назад
I did like your approach while explaining the dye of the SoC. Keep up the good work! Thank you.
@kidman2505
@kidman2505 5 месяцев назад
Your videos are amazongly well made, and the cintent is mint. These technical videos are all Id ever watch if possible lol.
@01chohan
@01chohan 5 месяцев назад
Kind of impressive how M3 Pro is slightly faster than M2 Pro (CPU wise) despite using fewer transistors and keeping the same number of total cores.
@alanmay7929
@alanmay7929 5 месяцев назад
They are literally using different settings and newer chip node
@sweealamak628
@sweealamak628 5 месяцев назад
Fantastic analysis. The M3 N3B node entered mass production late last year which implies that it was designed about 1.5 years ago or longer. This was all before ChatGPT and AI changed the world. So I suspect the Neural Engine to grow in size for the M4.
@m89hu
@m89hu 4 месяца назад
yesssssss!!!!
@AncientAviator
@AncientAviator 5 месяцев назад
I watched the entire video and was quite fascinated. Thanks very much for taking the time to produce this.
@Parseee
@Parseee 4 месяца назад
going to uni next year and really interested in these analysis videos. you're making a truly stunning job!
@Jamieibrahimm
@Jamieibrahimm 5 месяцев назад
There has been rumors of the iPad Pro getting a huge price increase to $1500-$1800 So I think maybe M3 Pro might be a “downgrade” so they can put it in the iPad to push more gaming on the iPad and to further segment iPad Pro from the iPad Air. And a few months ago, someone said Apple was working on a 14.1 inch iPad Pro with M3 Pro and when I saw the announcement of less CPU and GPU cores it made that leak even more believable
@jihadrouani5525
@jihadrouani5525 5 месяцев назад
I'm so hungry for these types of videos, I wish it had another 20 minutes in it comparing the GPU of Apple silicon to other GPU's like RTX 4090, since you calculated the estimated size percentage of GPU vs the entire die, you could have easily compared it to other GPU's, for example: + You said M3 Max's GPU takes roughly 35% of the 92B chip, that's roughly 32B transistors, for comparison RTX 4070 is a 35.8B chip. + You said M3's GPU takes about 23% of 25B, that's 5.75B transistors, for comparison a GTX 1650 is a 4.7B chip, and 1660 is a 6.6B chip. It kind of puts things into perspective and how much Apple need to get competitive with say a 4090 (76B)...
@HighYield
@HighYield 5 месяцев назад
I was actually going to compare the GPU size, but even if I know the transistors count of the entire chip and the GPU area, it's hard to be sure, since not all parts of the chip have the same transistor density. That's what stopped me, because I don't want to make claims I can't fully back up.
@jihadrouani5525
@jihadrouani5525 5 месяцев назад
Yeah it's definitely in the realm of severe speculation...
@GlobalWave1
@GlobalWave1 5 месяцев назад
Well i'd like to see what the M3Ultra will scale up to... That'll be interesting.
@__aceofspades
@__aceofspades 5 месяцев назад
It's very clear that Apple is throwing money around to try and be top dog. Their transistor budget is way higher than Intels and even higher than many dGPUs. Its not a good situation to be in for Apple, as they are too reliant on TSMC not to stumble (like they just did with N3).
@maltoNitho
@maltoNitho 5 месяцев назад
11:00 I just paused to comment… I’m watching every second because I haven’t found this level of detail about these chips said in such a succinct way. Thank you for keeping it entertaining and informative.
@davidbolus
@davidbolus 5 месяцев назад
watched it all - thanks for the hard work.
@bushgreen260
@bushgreen260 5 месяцев назад
*I wonder if we will finally get the true apple silicon Mac Pro this generation. With up and maybe beyond 128 performance CPU cores.*
@mikebruzzone9570
@mikebruzzone9570 5 месяцев назад
always, every segment, all the way through. mb
@halwye
@halwye 5 месяцев назад
Fascinating, and the depth of review is much appreciated. The lack of bias in particular makes this.
@marcin1337_
@marcin1337_ 5 месяцев назад
I watched it all to the end, your analysis are excellent and the deep dives is what I love
Далее
RTX 4090 Chip deep-dive
19:44
Просмотров 37 тыс.
How Apple Just Changed the Entire Industry (M1 Chip)
26:28
КОРОЧЕ ГОВОРЯ, 100 ДНЕЙ В СССР
09:31
Why next-gen chips separate Data & Power
18:56
Просмотров 115 тыс.
The billion dollar race for the perfect display
18:32
Apple's Silicon Magic Is Over!
17:33
Просмотров 837 тыс.
X86 Needs To Die
1:09:15
Просмотров 341 тыс.
Apple M2 Pro & M2 Max Chip Analysis
10:26
Просмотров 22 тыс.
How this tiny GPU invented the Future
18:00
Просмотров 217 тыс.
I tried to make a camera sensor
30:00
Просмотров 410 тыс.
How does Computer Memory Work? 💻🛠
35:33
Просмотров 3,5 млн
M3 Max blew my mind.
14:39
Просмотров 196 тыс.
Дени против умной колонки😁
0:40