NVIDIA Ampere (RTX 30) Architecture Deep-Dive: RT Cores, GDDR6X vs. GDDR6, & More

Ойындар

We're deep-diving into the NVIDIA RTX 3080 and 3090 architecture (Ampere), prior to our RTX 3080 review going live in the next video. This looks at memory, RT cores, and arch changes.
Sponsor: Get the Thermaltake Core P3 on Amazon - geni.us/Yf93
The best way to support our work is through our store: store.gamersnexus.net/
Like our content? Please consider becoming our Patron to support us: / gamersnexus
In this video, we're talking about NVIDIA's Ampere architectural changes versus Turing (and, to some extent, Pascal). We'll be covering the block diagrams for SMs and the GPU on the whole, but also talking about the major changes to RT cores and to the Tensor logic. Memory and memory bandwidth are also major performance drivers, and get additional discussion. This video will help you understand how the RTX 3080, 3090, 3070, and other NVIDIA Ampere (30-series) GPUs work at a lower level, and will help you understand why the performance changes are what they are. You'll have to check back for the review on the 16th for benchmarks.
TIMESTAMPS
00:00 - NVIDIA Ampere Architecture Overview
01:50 - NVIDIA Process Changes & 8nm
05:05 - Ampere SM vs. Turing SM Differences
12:56 - Ampere L1 Cache Changes
14:55 - Ampere 3rd Gen Tensor Cores
19:14 - Changes with 2nd Gen RT Cores (3080 vs. 2080 RT)
24:14 - Quick Summary
24:44 - GDDR6X vs. GDDR6 & Memory Improvements
28:07 - Conclusion: A Lot More to Cover, But...
** Please like, comment, and subscribe for more! **
Links to Amazon and Newegg are typically monetized on our channel (affiliate links) and may return a commission of sales to us from the retailer. This is unrelated to the product manufacturer. Any advertisements or sponsorships are disclosed within the video ("this video is brought to you by") and above the fold in the description. We do not ever produce paid content or "sponsored content" (meaning that the content is our idea and is not funded externally aside from whatever ad placement is in the beginning) and we do not ever charge manufacturers for coverage.
Follow us in these locations for more gaming and hardware updates:
t: / gamersnexus
f: / gamersnexus
w: www.gamersnexus.net/
Host, Additional Reporting: Steve Burke
Editorial: Patrick Stone
Video: Keegan Gallick, Andrew Coleman

Пікірлер: 1 000

  • @GamersNexus
    @GamersNexus3 жыл бұрын

    Here's our new GPU testing methodology for the rest of 2020: kzread.info/dash/bejne/X4Rrj6-FXcW2c6g.html The best way to support our work is through our store: store.gamersnexus.net/ Like our content? Please consider becoming our Patron to support us: www.patreon.com/gamersnexus

  • @paulhoeksema7200

    @paulhoeksema7200

    3 жыл бұрын

    Steve? How Is Your Day?

  • @eric4946

    @eric4946

    3 жыл бұрын

    Congrats on a million subs!

  • @stephenblanck6986

    @stephenblanck6986

    3 жыл бұрын

    Great information as always. I am especially curious on the cost of minimizing pam4 interference in the lanes as well as how long it will take micron to have 2 GB gddr6x chips.

  • @lb5928

    @lb5928

    3 жыл бұрын

    Bottom line: Amphere is power hungry and has nearly zero effeciency gain over turning with little to no ray tracing performance improvement due to a lack of increased cache space to match the increased RT cores.

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    If i move my hand in front my face fast enough i see motion blur.

  • @pieflies
    @pieflies3 жыл бұрын

    Brilliant! I have absolutely no idea what’s going on.

  • @MrThirumalar

    @MrThirumalar

    3 жыл бұрын

    👍 Most of us don't ! 😄

  • @legoNerd01245

    @legoNerd01245

    3 жыл бұрын

    same, i like listening to these videos while I do homework

  • @mustangthings

    @mustangthings

    3 жыл бұрын

    So.... I think the gist of it is that I do indeed need a 3090 for my KZread and Netflix machine.

  • @hhectorlector

    @hhectorlector

    3 жыл бұрын

    I’m only here for Snowflake. No Snowflake = downvote. Snowflake = upvote. 😐

  • @hmmfb6805

    @hmmfb6805

    3 жыл бұрын

    I feel this on another level

  • @Tiki832
    @Tiki8323 жыл бұрын

    By the way, something this video has made me realise is that i don't think the significant difference between how Nvidia and AMD GPUs handle instruction processing has been broken down cleanly in a way that would help a lot of people understand how drastically different the two are. Obviously the next week or two is likely to be busy with 3000 series coverage but could be a interesting topic for later October whilst waiting for formal announcements regarding Navi 2x.

  • @trevorlangdon

    @trevorlangdon

    3 жыл бұрын

    Or after navi's engineering day.

  • @Alien8ed96

    @Alien8ed96

    3 жыл бұрын

    Yes please ! I'd like to see that piece !

  • @LumpySpaceLeah

    @LumpySpaceLeah

    3 жыл бұрын

    like how much of a pain it was to get high occupancy on GCN

  • @Poisoned2010

    @Poisoned2010

    3 жыл бұрын

    Would def watch such a comparison video; thumbs up.

  • @jordanrodrigues8265

    @jordanrodrigues8265

    3 жыл бұрын

    AMD is proud to tell the public how their products interact with the rest of a computer system. There are actually open source drivers for Radeon and they're not half bad. Basically they're a hardware company and quite happy to collaborate on software. (The flip side of this is the disappointing closed source drivers they're infamous for.) nV has a long tradition of secrecy and treating their hardware plus software as a unified black box. I'm really kinda surprised by this video, but it simply doesn't get into the kind of detail that I'm used to having available even as a casual hobbyist developer. I have PDF manuals and the actual source code that my computer runs, no NDA required.

  • @thepolticalone961
    @thepolticalone9613 жыл бұрын

    GN: here's a video at 1:13am. Also GN: here's one at 7am. Me: do I go sleep late or wake up early?

  • @bengrogan9710

    @bengrogan9710

    3 жыл бұрын

    The only answer is Yes

  • @thepolticalone961

    @thepolticalone961

    3 жыл бұрын

    @@bengrogan9710 it is

  • @chemicalyinbalanced

    @chemicalyinbalanced

    3 жыл бұрын

    You saying review will be 7am UK time?

  • @Qneng

    @Qneng

    3 жыл бұрын

    Yes.

  • @BrumBrumBryn

    @BrumBrumBryn

    3 жыл бұрын

    Don't go to sleep.

  • @xDevscom_EE
    @xDevscom_EE3 жыл бұрын

    Tear it apart already, would you? :D

  • @smegmosisjones5645

    @smegmosisjones5645

    3 жыл бұрын

    If it’s even possible 😂

  • @skoopsro7656

    @skoopsro7656

    3 жыл бұрын

    Thanks for all you're hard work @Tin

  • @Lishtenbird

    @Lishtenbird

    3 жыл бұрын

    @@smegmosisjones5645 That reddit post with a blue jackhammer, titled "GN FE teardown kit".

  • @FerscMoses

    @FerscMoses

    3 жыл бұрын

    He already did - he just can't talk about it xD

  • @MH-hs7ie

    @MH-hs7ie

    3 жыл бұрын

    Lol.. That's no way even possible for OEMs...

  • @tibortheman
    @tibortheman3 жыл бұрын

    CANNOT WAIT for the 3000 series GPU test results. Best tech benchmarker out here. period.

  • @Lionheart1188

    @Lionheart1188

    3 жыл бұрын

    Hardware Unbox

  • @anafabula

    @anafabula

    3 жыл бұрын

    If you just want pure benchmarks and nothing else Hardware Unboxed is still the benchmark king. If you want a more in depth review Gamers Nexus is the best.

  • @AC-cg4be
    @AC-cg4be3 жыл бұрын

    I'm glad I'm not the only one that gets perturbed by the Bps vs bps screwup by people that should know better. As a network guy, it is extremely annoying when people say something like "We just got a 1 gigabyte circuit!" No, you have a 1 gigabit circuit. There's almost an order of magnitude difference.

  • @andersjjensen

    @andersjjensen

    3 жыл бұрын

    Three orders of binary magnitude to be precise :P

  • @AC-cg4be

    @AC-cg4be

    3 жыл бұрын

    @@andersjjensen 8 bits per byte. So a 1 GigaBYTE network circuit runs at 8 gigaBITS. An order of magnitude faster would be 1Gbps vs 10Gbps.

  • @andersjjensen

    @andersjjensen

    3 жыл бұрын

    @@AC-cg4be That's an order of magnitude in base 10, yes :P

  • @formdoggie5

    @formdoggie5

    3 жыл бұрын

    @@andersjjensen beat me to it

  • @pizzaman11

    @pizzaman11

    3 жыл бұрын

    Granted I blame the idiot who came up with the bit to byte naming scheme, he was just asking for people to confuse the 2.

  • @Fritzsch82
    @Fritzsch823 жыл бұрын

    "...and because no-one in the entire world likes motion-blur" :D damn right

  • @OdinAlgeron

    @OdinAlgeron

    3 жыл бұрын

    Sometimes I wish I could roll up a old phone book and smack developers over the head NO, NO! No motion blur!~ BAD DEVELOPER!~

  • @hhectorlector

    @hhectorlector

    3 жыл бұрын

    Actually not all motion blur is bad anymore. Shadow of the Tomb Raider has excellent motion blur. Most of the time it sucks but sometimes developers do it well!!

  • @ze_rubenator

    @ze_rubenator

    3 жыл бұрын

    @@OdinAlgeron I used to love it when playing NFSU 1/2 and NFS MW, it gave such an immense feeling of speed and danger that simply wasn't there without motion blur.

  • @pokeguy742

    @pokeguy742

    3 жыл бұрын

    @@hhectorlector true. I enabled motion blur in sotr and I only vomited for 20 minutes. Best implementation of a garbage concept yet.

  • @michelvanbriemen3459

    @michelvanbriemen3459

    3 жыл бұрын

    Amen to that. You don't need motion blur when your brain can -fail to properly interpret the signals- do the job on its own just fine.

  • @ethzero
    @ethzero3 жыл бұрын

    The section on sparse matrices fried my dense matrix. Cheers.

  • @CaveyMoth

    @CaveyMoth

    3 жыл бұрын

    It's taking over my brain. It's like a dominatrix.

  • @CynHicks

    @CynHicks

    3 жыл бұрын

    🙉

  • @maxbet3968

    @maxbet3968

    3 жыл бұрын

    Matrices: Revolution was my favorite movie

  • @trevorlangdon

    @trevorlangdon

    3 жыл бұрын

    Actually it fried your sparse matrix...

  • @soldni

    @soldni

    3 жыл бұрын

    tbh it was really not that well explained; it’s clear that this is not Steve’s area of expertise, so he relied heavily on the documentation provided by NVIDIA instead.

  • @stelp7617
    @stelp76173 жыл бұрын

    LOVE the t-shirt!! some love for us down under. Really looking forward to the tear down and benchmarking!

  • @Lishtenbird
    @Lishtenbird3 жыл бұрын

    Having watched this deep dive, I now feel the opposite of sparse.

  • @Klaeyy

    @Klaeyy

    3 жыл бұрын

    I have also been called dense since watching this video, yes.

  • @IngwiePhoenix

    @IngwiePhoenix

    3 жыл бұрын

    *d e n s e*

  • @Kennykazey
    @Kennykazey3 жыл бұрын

    Kudos for all the work going into the presentation, with timestamps, timed animations etc.!

  • @adjoho1
    @adjoho13 жыл бұрын

    "Because nobody in the entire world likes motion blur, but we should still spend hardware on adding it". Vintage Steve right there.

  • @OldBuford
    @OldBuford3 жыл бұрын

    Damn, this video really made me realize how little I understand how GPUs actually work...

  • @phutureproof

    @phutureproof

    3 жыл бұрын

    I think thats a fair comment for 90% of us here

  • @XDbored1

    @XDbored1

    3 жыл бұрын

    i don't usually think about it either but here's my guess probably half wrong texture mapping is just reading and aligning grids of data to the correct part of the GPU memory, and then FPUs are fed more numbers from the CPU about vertices and then process those to twist all the grids to the correct angles to turn your pile of textures into a 3D vector image, then ROPs read that vector image and place a grid of pixels at your viewport equal to the render resolution over that 3D picture and samples each square to get a ideal colour for each pixel and then that rasterized 2D picture is processed by shaders that can read and edit individual pixels to any colour based on software scripts and filters more bigger textures means more TMUs and memory is needed, more vertices means more CPU and FPUs are needed, more pixels means more ROPs, memory, and shaders, are needed. when we talk about GPU rasterization performance in games we typically mean the ROPs,memory.shaders. as texture mapping is a very simple 2D operation that is only ever bottlenecked by memory not really the number of TMUs GPU architecture can still beniift from having TMUs closer to memory or ROPS which is why large GPUs still have more TMUs then smaller ones modern GPUs are extremely fast at low precission 16but floating point operations only 2nd to dedicated crypto mining Asics or many core multisocket server CPUs so they will typically become bottlenecked by either CPU or ROPS/memory/shaders before running out of FPUs if a GPU did run out of FPUs it look like not gaining any extra fps from etiher upgrading the CPU or reducing the render resolution which would be awful and weird and probably blamed on a poorly optimized game engine if it was ever encountered 32buit or 64bit floats are another story computer resolutions are not high enough that there would be any worry of 3D lines not being straight enough having there angles stored as only 16bit floats and it is doubtful that they ever will be, high precision FPU performance for GPUs seems to get worse with every new generation and is often intentionally gimped to push more expensive workstation cards on anyone with a super niche specific use case like exporting almost impossibly precise CAD files to the worlds most precise 3D milling/printing machines that mere mortals could not afford anyway.

  • @callofduty611
    @callofduty6113 жыл бұрын

    As always, great research with a presentation of the facts and any distinction between what isn't. We need more like you, Steve.

  • @KLCII88
    @KLCII883 жыл бұрын

    Can’t wait for the in depth analysis y’all are going to bring us!

  • @ebsmith1102
    @ebsmith11023 жыл бұрын

    Will Top Benchmarker and CEO, Snowflake be in the video tomorrow?

  • @GamersNexus

    @GamersNexus

    3 жыл бұрын

    Yes!

  • @digital_zero1698

    @digital_zero1698

    3 жыл бұрын

    @@GamersNexus Gotta have the A-Squad out for the big videos.

  • @hugevibez

    @hugevibez

    3 жыл бұрын

    @@GamersNexus Tbh more excited about this than any new GPU coming out

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    Your boi on his channel dropped the intel compiler video i bitched about earlier a few weeks ago. It works in games too. Not just math libraries.

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    The microstutter on ryzen is why the low fps score amd vs intel is what it is. Starcraft 2 and farcry 5 are the deep dives you probably won't explore. I might beat you to it.

  • @rndargis0
    @rndargis03 жыл бұрын

    As I said last video : I’m a professional cuda programmer. My job was to literally test the gpu performance in a certain system with dual cpu socket. I’ve made custom software that test memory bandwidth between the ram and gpu , wrap size optimisation ,kernel optimisation ect... if you ever need help for workstation testing and what the data means I can help you with that.

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    Remember the coreteks reveal on a.i. rearranging os kernals to allow more threads on x86 to formulate solutions for otherwise single threaded applications without context switching? Its the same thing with multi gpu. Its heralded as hard to implement but is it really in truth... Industry needs demand and technology is only slowing down as it speeds up.

  • @rndargis0

    @rndargis0

    3 жыл бұрын

    Lucifers Paradise It all depends on your implementation of dual GPU. If you have 1 CPU and 2 GPU you might need to split pcie lines and the total bandwidth will stay the same but the FLOPS might be higher if your algorithm is more computing bound than memory bound. If you use 2 cpu with 2 ram bank now it’s starting to be more tricky since you can assign memory in different places and access from the other bank.

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    @@rndargis0 We live in the age of automation. What the fucks taking so long.

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    Like low level api adoption. We should have the thread scaling and retain hardware abstraction for gpu until A.I. can make non fixed hardware serviceable in a financially responsible way. Industry i know for a fact CAN do better but in a lot of ways wont. Like why hasn't amd called intel out on their compiler yet? Its EASY to prove.

  • @remasteredretropcgames3312

    @remasteredretropcgames3312

    3 жыл бұрын

    A lot of performance is held hostage for the fear that we have enough and for years. Its a conspiracy i have wrestled against in my personal sphere for years. Out of order execution is great but Microsofts thread scheduler abuses that far far more heavily than a party of competent software engineers should. Im better at telling it how to think and im a fucking hobbyist. Thats pathetic.

  • @larrydugger2971
    @larrydugger29713 жыл бұрын

    Excellent 'introduction'...time to head to your store...better yet Patreon, as always thanks for what you do!

  • @Keith2XS
    @Keith2XS3 жыл бұрын

    Wow, thanks for explaining things better than most professors! You have a great team there, keep it up! (Your motion blur comment was priceless)

  • @coolbrotherf127
    @coolbrotherf1273 жыл бұрын

    In a 30 min video that sounds like a college lecture, I didn't expect to laugh as hard as I did at 22:02.

  • @LinkinRepublic
    @LinkinRepublic3 жыл бұрын

    Gamers watching this, thinking “damn I’m smart..” while zoning out on some of this deep stuff

  • @AB_538

    @AB_538

    3 жыл бұрын

    Attacked.

  • @DravicPL

    @DravicPL

    3 жыл бұрын

    @SIG /r/gatekeeping I'm SmArT, UnLikE YoU.

  • @UaE0BoSS1

    @UaE0BoSS1

    3 жыл бұрын

    Ali Baba look at me, i use reddit, r slash HAHA

  • @DravicPL

    @DravicPL

    3 жыл бұрын

    @@UaE0BoSS1 tell me how I am wrong though. This clown saying a technical breakdown is shallow, like no shhh Sherlock, it's not that serious. We can still enjoy it and "nvidia's fault" implies that Nvidia is wrong for releasing white paper to the masses, are you insane?

  • @nanoprehistoric

    @nanoprehistoric

    3 жыл бұрын

    @SIG you're smart then, I have to rewind some parts over and over again to understand it hahaha.

  • @nanoprehistoric
    @nanoprehistoric3 жыл бұрын

    Whoaw this is heavy topic, I have to rewind some parts over and over to understand it hahaha. I love it, thanks!

  • @benchmarksandmore948
    @benchmarksandmore9483 жыл бұрын

    It’s just wonderful to have Steve dive into the deeper aspects of the graphic cards. It’s really very enjoyable. Coming from someone whose studying computer science

  • @ApocDevTeam
    @ApocDevTeam3 жыл бұрын

    Have you locked Buildzoid in a cage yet? I need his PCB reviews for the custom cards when the embargo drops tomorrow.

  • @trevorlangdon

    @trevorlangdon

    3 жыл бұрын

    Custom lifts the day after. I expect the FE will be the most interesting one this time.

  • @_--_--_

    @_--_--_

    3 жыл бұрын

    @@trevorlangdon Yeah for that price certainly, it will be in the higher end regarding AIB cards. Nvidia certainly did a dirty one on all the AIB partners.

  • @DSP_Visuals

    @DSP_Visuals

    3 жыл бұрын

    Seems like Buildzoid is the type who locks himself in his room.

  • @ScytheNoire
    @ScytheNoire3 жыл бұрын

    19GB on the RTX 3080. Confirmed. Heard it on the internet.

  • @jayhill2193

    @jayhill2193

    3 жыл бұрын

    It's GigaBIT, so the 3090 will only have about 2.4 GB of memory, duh

  • @darkshadow7709

    @darkshadow7709

    3 жыл бұрын

    @@jayhill2193 memory bandwidth*

  • @russellneal1263
    @russellneal12633 жыл бұрын

    I appreciate the timestamps a lot, I had to go back and refresh myself on some of the acronymns being used and it was so easy because of that. Thank you guys for the technical break down as always.

  • @pr1musinterpares100
    @pr1musinterpares1003 жыл бұрын

    Congratulations on the Million subscribers. You guys produce some of the best tech content.

  • @hectorohh
    @hectorohh3 жыл бұрын

    pfft, this is why I dropped out, I can watch a KZread video and then have an intellectual conversation with my dog reciting 2% of the information you provided to prove my self worth!

  • @StubbornProgrammer
    @StubbornProgrammer3 жыл бұрын

    Audience: Comments before watching the whole video. Steve: ಠ_ಠ

  • @hhectorlector

    @hhectorlector

    3 жыл бұрын

    The people who comment before finishing the video are the only people that get their comments “loved” by GN

  • @mirageman85
    @mirageman853 жыл бұрын

    The best information on Ampere out there on KZread is here on this Gamers Nexus video. Kudos !!!!!!

  • @RackBaLLZ
    @RackBaLLZ3 жыл бұрын

    Soooo Happy to see you've broke the 1M sub barrier!! CONGRATULATIONS 🎊

  • @sm1tty.
    @sm1tty.3 жыл бұрын

    this such a tease! I thought it was the review when I lazer'd the notification!

  • @robertandrews6915

    @robertandrews6915

    3 жыл бұрын

    Not really his fault. I'm sure he'd love to do a review but Nvidia wouldn't let him publish it. The reviews will probably come out on the release date

  • @sm1tty.

    @sm1tty.

    3 жыл бұрын

    Robert Andrews already confirmed in another GN video they come out tomorrow but I thought maybe he was dropping it early

  • @Carnyzzle
    @Carnyzzle3 жыл бұрын

    this wait for the benchmarks has felt so dang long

  • @PavanSuresh
    @PavanSuresh3 жыл бұрын

    I cannot WAIT for the teardown and benchmark videos tomorrow. Seriously, cannot wait. Can you post them already?!

  • @MatrixRocker
    @MatrixRocker3 жыл бұрын

    Love how Steve and his team gets deep in this stuff. I'm hyped for their tests on the 3000 series.

  • @__aceofspades
    @__aceofspades3 жыл бұрын

    Wait for benchmarks.

  • @longjohn526
    @longjohn5263 жыл бұрын

    Steve reminds of a professor I had in college who would always drop these droll little insider jokes into his lectures. I asked him one time why he did it and he told because he could judge how the students were picking up on the lectures by how many smiled and if a student laughed outloud (guilty) then he knew they really got it As a side note I think we should change Steve's nickname from Tech Jesus to The Professor ......

  • @BloodBoughtMinistries

    @BloodBoughtMinistries

    3 жыл бұрын

    Yeah seeing Jesus of Nazareth was a Jew who did not have long hair as the Catholics portray Him and those living in places like the USA. Pretty ignorant.

  • @Anipixelz

    @Anipixelz

    3 жыл бұрын

    @@BloodBoughtMinistriesI mean it's just a joke, who cares about actual jesus you have got something better right here!

  • @ZioYuri78

    @ZioYuri78

    3 жыл бұрын

    Steve is to hardware as Professor Leonard is to math.

  • @longjohn526

    @longjohn526

    3 жыл бұрын

    @@BloodBoughtMinistries Ignorance is claiming to know what Jesus looked like since there are no paintings of him or any known description that have been validated .... And besides it's the message not the man's looks that's important

  • @StevenCasteelYT
    @StevenCasteelYT3 жыл бұрын

    I just learned about the word "architecture" regarding chips and your video looks to be the lesson I need.

  • @rookie22222
    @rookie222223 жыл бұрын

    loved the deep dive ,can't wait for the reviews and tear downs!

  • @000NULL
    @000NULL3 жыл бұрын

    Are we getting...amp’d here?! Also thanks for the fast shipping on the mouse mat and tool kit! Giant mat.....

  • @MakenaKai
    @MakenaKai3 жыл бұрын

    I am most interested on how they were able to bypass and thus offload so much from the CPU

  • @LuisHansenNH

    @LuisHansenNH

    3 жыл бұрын

    By taking data directly from RAM, not having to ask the CPU every time.

  • @DravicPL

    @DravicPL

    3 жыл бұрын

    I don't think they've done anything of the sort but maybe I am missing something. I watched this entire video and the entire Ampere presentation by Jen-Hsun Huang, and apart of the song of the future called DirectX12_2 and its direct storage technique, nobody said anything about offloading CPU load.

  • @DravicPL

    @DravicPL

    3 жыл бұрын

    @Nelson Lambert great, what?

  • @DravicPL

    @DravicPL

    3 жыл бұрын

    @Nelson Lambert did you fookin not read my comment? Read it. I mentioned direct storage but that literally is game engine dependant and won't come into play until like 2022.

  • @IngwiePhoenix

    @IngwiePhoenix

    3 жыл бұрын

    A CPU is great at a general broad of tasks - GPUs however excel in simple math that is either repeated over and over or is just...plainly too simple for a CPU to handle in time. Now, a CPU and a CPU aren't exactly the same; the x86 architecture is different to ARM and RISC-V. A game usually uses a graphics API like OpenGL, Vulkan or DirectX. Each of those _actually_ implement the GPU communication by going through the NVIDIA driver which is the endpoint that implements those graphics APIs. The game is actually unaware who provides the OpenGL API or the Vulkan API. The driver itself knows how to translate calls to those to actual GPU instructions, communicates with it and exchanges data - and that _is_ happening on the CPU, initially. Optimized applications may however write parts of themselves to the GPU memory and tell it to execute it, much like a CPU would, but with a different instruction set tailored to that GPU. There is, for instance, a C/C++ toolchain designated for programs to embed CUDA. So, you could build your main game on the CPU and offload tasks to the GPU directly by working off it's bare metal and profiting from it's vastly different architecture. This is something Blender does, for instance. Now, that's just a very, very basicallyfied description of what's going on - but maybe you get what I am trying to say. ^_^

  • @first-thoughtgiver-of-will2456
    @first-thoughtgiver-of-will24563 жыл бұрын

    awesome content! I love hearing Steve talk about linear algebra/machine learning specs.

  • @hi_tech_reptiles
    @hi_tech_reptiles3 жыл бұрын

    This is phenomenal content, and why I am a patron. Maybe a dumb question, but now with the FP/INT dual area of the GPC, can that area do both at the same time or is that half of the GPC only do one or the other at the same time?

  • @Wanooknox
    @Wanooknox3 жыл бұрын

    TIL that Steve is a "ph" Steve. Phteve from Gamers Nexus.

  • @albertotr1

    @albertotr1

    3 жыл бұрын

    The ‘s Nexus’ is silent - cause he’s a gamer

  • @AAULTEN
    @AAULTEN3 жыл бұрын

    Gray banding on the backgrounds for some of the motion blur slides, needs more 10bit

  • @resColts

    @resColts

    3 жыл бұрын

    Nah just needs more blur

  • @CaveyMoth

    @CaveyMoth

    3 жыл бұрын

    I get banding on every gradient background in KZread videos =/.

  • @resColts

    @resColts

    3 жыл бұрын

    Video isn't HDR if you have HDR on that's why

  • @igoresque

    @igoresque

    3 жыл бұрын

    dithering fixes banding even at 6bit depth, even though it’s noisier, but youtube won’t let you have any gradients

  • @TheRealLink
    @TheRealLink3 жыл бұрын

    Great deep dive! Understood some. Excited to see reviews in the coming days.

  • @rosshalz
    @rosshalz3 жыл бұрын

    A 30 minute architecture deep dive gn video on a gpu series that I was eagerly waiting for? This is gonna be gooood...

  • @thorinoakenshield663
    @thorinoakenshield6633 жыл бұрын

    Present for class!

  • @erikhendrickson59
    @erikhendrickson593 жыл бұрын

    It's almost time, boys!

  • @netrex9916

    @netrex9916

    3 жыл бұрын

    Yes! It might actually show up before Christmas!

  • @CarthagoMike
    @CarthagoMike3 жыл бұрын

    Always nice seeing a new Gamers Nexus video at 3am. Time to get a cup of coffee, sit back and relax.

  • @jonhughes4079
    @jonhughes40793 жыл бұрын

    Fuck yeah, I've been waiting for this, see you tomorrow for the review too!

  • @Andarus
    @Andarus3 жыл бұрын

    I understood some parts of it!

  • @nofekun1889

    @nofekun1889

    3 жыл бұрын

    So did us!

  • @joeljacobson9423
    @joeljacobson94233 жыл бұрын

    Aw man. GN got me. I thought this was the 3000 review

  • @squirt626

    @squirt626

    3 жыл бұрын

    Tomorrow

  • @terbo2000

    @terbo2000

    3 жыл бұрын

    "Jebaited" - AMD, probably

  • @dude3278

    @dude3278

    3 жыл бұрын

    @@squirt626 tomorrow at midnight or tomorrow in the morning?

  • @xM4XT0Nx

    @xM4XT0Nx

    3 жыл бұрын

    @@dude3278 www.timeanddate.com/countdown/to?iso=20200916T14&p0=136&msg=+RTX+3080+review+embargo+drop&ud=1&font=cursive&csz=1

  • @dude3278

    @dude3278

    3 жыл бұрын

    @@xM4XT0Nx Thanks

  • @speedibusrex
    @speedibusrex3 жыл бұрын

    Congrats on the 1000,000 :)

  • @rmg110
    @rmg1103 жыл бұрын

    Congrats on the 1 000 000 !

  • @charlesballiet7074
    @charlesballiet70743 жыл бұрын

    1:15 "...on how stuff works.." didn't you see the presentation by jensen "It just works....." ok my meme quota has been filled for the day

  • @Adromelk
    @Adromelk3 жыл бұрын

    Me understanding about 10% of this : yes, very interesting

  • @brycesimon4755
    @brycesimon47553 жыл бұрын

    This was a very helpful piece, the changes to the RT cores make me excited to see benchmark numbers for Octane Renderer! if only there was someone around here dedicated to benchmarking...

  • @brycesimon4755

    @brycesimon4755

    3 жыл бұрын

    I'll buy a second mouse mat if that helps :p

  • @sagumekishin5748
    @sagumekishin57483 жыл бұрын

    Always good to see the technical side of new GPUs. It would be better if comparison to the data center version can be made. For example GA100's SM has 192KB of cache vs GA102 only has 128KB per SM. GA100 also has 8 LD/ST unit per SM vs GA102's 4 LD/ST per SM.

  • @josebr0
    @josebr03 жыл бұрын

    They literally had to change the hardware so that people could see less with motion blur enabled. How is that this thing still exists on PC games is beyond me.

  • @ujiltromm7358

    @ujiltromm7358

    3 жыл бұрын

    *C I N E M A T I C E X P E R I E N C E*

  • @Verpal
    @Verpal3 жыл бұрын

    Still wondering whether Micron can pull a Miracle and we will see a single sided RTX 3080ti with 20GB GDDR6X this year.

  • @Wellemy79
    @Wellemy793 жыл бұрын

    Grats on 1 mil subs!

  • @fsi2210
    @fsi22103 жыл бұрын

    Thank you, your educational videos, are the one's I most look forward to and appreciate.

  • @DSP_Visuals
    @DSP_Visuals3 жыл бұрын

    Nvidia has said that the 3080 is the new flagship, and not the 3090, but keeps comparing the 3080 to the 2080 and 2080 Super instead of the 2080Ti. Then there's also the fact that the 3090 exist is not sold as a Titan and is being sold by AIBs. They're just trying to say they didn't raise the price on their flagship card this generation.

  • @nry3223
    @nry32233 жыл бұрын

    Videos like these really make me want to get my degree in computer engineering. Because I love stuff like this. I want to know how they get the TMC's, the SM's, Cuda cores, etc. I don't really like the software side, but the hardware is so juicy. Getting a business degree though currently...maybe I'll also go for a CompEng eventually.

  • @national_security

    @national_security

    3 жыл бұрын

    I love listening to stuff like this, but when you get to low levels of hardware it becomes so dirty. I am studying computer science/software engineering and really dislike hardware based courses.

  • @nry3223

    @nry3223

    3 жыл бұрын

    @@national_security I guess everyone is different. I found out when I initially was going to uni for a CompSci degree that I really, really don't like coding. I can read it but I can't do it myself. Makes it hard to succeed. However I really love the super nitty gritty details of hardware, and I want to at least get an education in it so I can understand it more, and be able to look at something like a gpu and know to a degree how it's constructed. Or look at a motherboard and be able to do stuff like Buildzoid does where he basically covers everything on the board edge to edge and throughout. I don't plan to do much in life, if I actually make it in CompEng then I do. But I'd rather have my own small business tech shop where I build computers, or keyboards, etc. Kind of like a hobby, but I get paid for it. But I'd love what I'd be doing.

  • @national_security

    @national_security

    3 жыл бұрын

    @@nry3223 If you like it, go for it. Wish you all the best. I just wanted to note that high level hardware is beautiful, but lower levels are a lot different. It is good to see people interested in a wide variety of tech fields. We can all complement each other. I would love to have small business, but where I live market is really small. It would never succeed.

  • @nry3223

    @nry3223

    3 жыл бұрын

    @John Statesman I already wasted $25,000 for one year. Back to community college to not waste money. I know how it feels. Unfortunately it's less about the education these days and more about the connections you make. Where the unfortunate part comes in is most of those connections have to be made through a portal, and the easiest and best one to do it through is education. Spend money to make money essentially. Although I'm using the term money loosely here.

  • @d8nish951

    @d8nish951

    3 жыл бұрын

    Same if I could I wish I could go back and do a computer science degree instead of biomedical science 😭

  • @theOSMaxwell
    @theOSMaxwell3 жыл бұрын

    These kind of videos encourage me even more to keep studying computer engineering. Good one GN.

  • @enoktheewok4821
    @enoktheewok48213 жыл бұрын

    Needed this vid on the gddr6 difference, thanks!

  • @smoketinytom
    @smoketinytom3 жыл бұрын

    I’ve never smashed the notification faster than this! RIP Early Morning Start!

  • @dvtye3378
    @dvtye33783 жыл бұрын

    God I can't wait for the 16th to come already so I can comment on GN's RTX 30s series vids before watching the whole thing.

  • @shinji3rd403
    @shinji3rd4033 жыл бұрын

    gratz for the 1M, dear Stephen. ;)

  • @Lets_get_wealthy
    @Lets_get_wealthy3 жыл бұрын

    Just wanted to thanks for the content. Keep up the great work. Cheers.

  • @DarkSoulaire
    @DarkSoulaire3 жыл бұрын

    I havent finished the video, only just started, but i felt like i wouldve heard u touch on this before in a video somewhere considering how particular you are... Their "1.9 perf/w" is using the wrong units (at least the way the diagram shows it). since it's /w, the watt is what should be unchanged for that comparison as it is the denominator (the line they draw from either graph should be verticle, not horizontal). I understand that they wanted to show at 60 fps, but that unit should then be watts/frame ("it uses 2.1w/frame vs 4w/frame at 60fps" for example). in actual percent, the efficiency ("1.9 more") is it uses only 52.6% of the power at 60 fps. If u look at the 240w mark where it is for turing, its 60fps, while ampere at 240w is ~90 so its actual "perf/w is about 1.5x at 240w. Its like u said on other videos about the arbitrary "its 3x quieter" type marketing for products... It technically doesnt make sense, generally what they mean is it produces 33% of the sound that the other does. Sorry had to say it somewhere because it bugged me - when i saw it in the live show i was like wait, thats not quite how that works.

  • @iskierka8399

    @iskierka8399

    3 жыл бұрын

    Putting a parameter as the denominator doesn't mean it has to be controlled, just that it's a figure being divided by. The exact same number can be expressed by inverting it with no change in meaning; compare the difference of measuring a car's fuel efficiency in miles per gallon or litres per 100 km - both describe the exact same thing regardless of how it was measured. What you're picking at would only be equivalent to how driving the car faster results in a lower efficiency - that doesn't end up meaning though that the comparison at a constant speed was invalid, or that you can't express the constant speed efficiency in one of the measurements. Both are valid as long as you explain how the number was obtained, which nvidia did - performance per watt at a constant performance. Though either way it makes AMD's advertising an interesting comparison when they claim a 50% efficiency gain, which is what nvidia has when measured the worse way, and AMD would have no reason to make their number look worse by comparison. The sound being incorrect is totally valid, though, there's no universal noise floor that can be used to measure it being "three times quieter".

  • @DarkSoulaire

    @DarkSoulaire

    3 жыл бұрын

    No, i stand by what i said. The main thing is that if they are saying you are getting more performance per watt, then the performance number (fps) should be increasing, not staying the same at 60fps. so the watt does need to stay constant. Also, "more efficient" is what this graph is claiming, but as a higher efficiency means a lower number, 1.9x is multiplying in a negative direction (visually you're going from turing to ampere, the line should actually be an arrow like

  • @nathangamble125

    @nathangamble125

    3 жыл бұрын

    ​@@DarkSoulaire "No, i stand by what i said." Then you're wrong. Nvidia made it clear what they were showing in the graph, and there's nothing about efficiency comparisons that says you have to control for one variable or another. In my opinion the way Nvidia compared it is better (it's more useful for people who are targetting a specific frame rate), but that's just an opinion as much as yours - both ways are valid, and Nvidia showed the graph they were using for the comparison so anyone who looks at it can see what they mean, so there should be no ambiguity in what they're claiming.

  • @DarkSoulaire

    @DarkSoulaire

    3 жыл бұрын

    No you're wrong. You literally started as saying I'm wrong and then contradicted yourself by saying that both ways are valid. The fact that both ways are valid, as well as the fact that it had to be clarified that it was efficiency not performance, is basically the literal definition of ambiguity. However if they choose to present it as i said there is literally no other way to interpret it

  • @UnknownEAFCPlayer
    @UnknownEAFCPlayer3 жыл бұрын

    I have a 2080 Ti and I still haven't sold my 1080 Ti. I'm sad and lonely. I need your support during these tough times.

  • @mjc0961

    @mjc0961

    3 жыл бұрын

    I have a 1080Ti and still haven't sold my 970. ...I also have an R5 3600 and still haven't sold my i7-3770. RIP RIP

  • @Safetytrousers

    @Safetytrousers

    3 жыл бұрын

    I got a good price for my Strix 1080ti. It took a lot of the sting out of the 2080ti price for me.

  • @BrendonCamm
    @BrendonCamm3 жыл бұрын

    Wow, thank you so much for breaking this down! Super interesting stuff. Have a great night :)

  • @zmb5501
    @zmb55013 жыл бұрын

    I wish I was this knowledgeable on computer stuff like you steve. some of this goes over my head.

  • @diegotambor9283
    @diegotambor92833 жыл бұрын

    Steve has warned us for tomorrow...

  • @patricktho6546

    @patricktho6546

    3 жыл бұрын

    you mean today? Or Australia tomorrow? Witch Timezone (in GMT +-)

  • @BloodBoughtMinistries

    @BloodBoughtMinistries

    3 жыл бұрын

    Its today, not tomorrow. Tomorrow will be the 17th

  • @jamesmcclain3588
    @jamesmcclain35883 жыл бұрын

    Their need to be an emote for a hand flying past the top of someone’s head

  • @GFClocked
    @GFClocked3 жыл бұрын

    Great video. Can't wait for more in-depth.

  • @BH-xe7nb
    @BH-xe7nb3 жыл бұрын

    I’m watching this and definitely understanding what you’re talking about. Yep, makes total sense to me.

  • @WildkatPhoto
    @WildkatPhoto3 жыл бұрын

    GN content at a normal hour? What is this witchcraft?

  • @joshlampe3458
    @joshlampe34583 жыл бұрын

    I'm VERY interested in your information on this, as I'm looking for a GPU upgrade now there is an actual, sizable difference in this generation. Probably going to buy a 3080, but will end up waiting until AMD announces their offering in October (I'd rather buy a AMD if nothing more than to support Freesync). Thanks for the excellent, unbiased info. I'd expect nothing less from the best computer information source on this planet. Well done GN.

  • @hhectorlector

    @hhectorlector

    3 жыл бұрын

    But all Nvidia cards since the 1050ti support Freesync?

  • @joshlampe3458

    @joshlampe3458

    3 жыл бұрын

    @@hhectorlector You need to buy a Gsync monitor to get the card/monitor frame synchronization from Nvidia. They basically charge the monitor manufacturer a hefty fee to make their monitors 'Gsync'. If you look at monitor prices, the end consumer pays that cost, which is, generally speaking, in excess of $100. There is no fee to the manufacturer (and ultimately you) for AMD's Freesync.

  • @ChristopherGoggans

    @ChristopherGoggans

    3 жыл бұрын

    @@joshlampe3458true, there is a "fee" but my understanding was that there was also extra hardware and substantial certification to prove the quality of the display before being G-Sync certified. In short, it's not all smoke and mirrors and there are actual hardware differences between freesync and gsync. As to whether those differences should cost as much as they do, that's a separate discussion lol.

  • @Karrosh1234

    @Karrosh1234

    3 жыл бұрын

    @@joshlampe3458 This is no longer the case. Nvidia cards support Freesync at the driver level and have for a while now.

  • @andersjjensen

    @andersjjensen

    3 жыл бұрын

    G-Sync, Freesync and VESA Adaptive Sync all accomplishes the same thing. Nvidia (as usual) jumped the gun and went ahead with a proprietary standard (at a markup), but as soon as both AMD and VESA caught up Nvidia (as usual) went "look at us, we're the good guys who support everything". So the others are right. Nvidia now works with everything (they enabled it in the drivers and a fair few "fans" were pretty darn miffed because they could have saved $100 on getting the non-G-Sync version of their monitor).... This has happened with everything from texture formats to (the now irrelevant) PhysX and eventually the CUDA language will go the same way, as the industry is starting to demand the vendor neutral OpenCL... And before you ask.. yes, I've been around for a while :P

  • @SonGoku-97
    @SonGoku-973 жыл бұрын

    been watching this channel for years now and somehow still learn some new shit on almost every video that drops

  • @LunarLaker
    @LunarLaker3 жыл бұрын

    Darn shame Anandtech hasn't uploaded their deep-dive yet, very grateful you guys did though! Guess I have to read 2 Ampere deep dives :P

  • @SinisterPuppy
    @SinisterPuppy3 жыл бұрын

    Wonder if game engines will ever leverage the AI parts of GPUs? I'd love to have (mostly) competent characters in games, even if for only one or two main characters. We'll have these 8k, ultra raytraced, near photorealistic characters that are dumb as bricks otherwise.

  • @LaughingMan44

    @LaughingMan44

    3 жыл бұрын

    That's...that's not how it works...the AI has nothing to do with game AI...I....jesus Christ someone help me

  • @andersjjensen

    @andersjjensen

    3 жыл бұрын

    It is still pretty far away. The thing with "deep learning" (correct term is neural network simulation) is that it is incredibly susceptible to the saying "garbage data in, garbage data out". What that means is that you need to feed it incredibly controlled data with only one type of difference in it.. or you'll get wildly unpredictable results. Example: And institute wanted to experiment with recognition of cars. So they send out a team of students to take photographs of places with cars. The next day they took another round and waited until there were no cars, but took the pictures from slightly different angles, and also took a some more random pictures without cars.... They trained the network, and it worked 99.99% on the reference data, but failed miserably on the (bought) control set... They bugged around with it for ages until someone noticed: on day one there was blue sky... on day to there was an ever so slight over draft... So they had trained the network to recognize cars on over draft days! Now, when we get to something as complex as "behavior" of characters in a game you can see why producing clean training data might become a serious head ache, and why you'd need to "stack" several networks on top of each other to get something reasonably coherent, right?

  • @tacticalcenter8658

    @tacticalcenter8658

    3 жыл бұрын

    Forza uses ai for the cars in game to pick better lines.

  • @andersjjensen

    @andersjjensen

    3 жыл бұрын

    @@tacticalcenter8658 That is also a very "mono task"-like operation to do. It is fairly easy to generate quality training data for.

  • @LaughingMan44

    @LaughingMan44

    3 жыл бұрын

    @@tacticalcenter8658 what has that to do with nvidia?

  • @paulosouza9329
    @paulosouza93293 жыл бұрын

    Never been so early. 6 seconds wow

  • @samuelbaird4983

    @samuelbaird4983

    3 жыл бұрын

    That's what she said

  • @keuj517
    @keuj5173 жыл бұрын

    CONGRATS 1M Subscriber!

  • @xOubax
    @xOubax3 жыл бұрын

    Well, I feel like I just got out of class after this video! Very informative Steve! 👍🤘

  • @uss_04
    @uss_043 жыл бұрын

    Stop asking for Steve to drop the benchmarks already. That’s Linus job.

  • @World_Theory
    @World_Theory3 жыл бұрын

    A non-binary memory data channel? Interesting! How long before they're straight up using analog signals? (I'm joking. Analog would have *all* the errors inside a noisy environment like a computer. Discrete voltage steps will be preferable to that, assuming you have the necessary sensitivity and error checking built in, like it seems Nvidia has here. Although… If anyone were to get optical signals working on that scale, there may yet be hope. Don't quote me on that, though; my physics knowledge isn't quite that good, and that's getting rather close to the quantum scale of physics for my comfort. But, optical computers!)

  • @cheozuka
    @cheozuka3 жыл бұрын

    Keep up the great work! ps what is your t-shirt from? (I am curious as I am from australia)

  • @markisfish4489
    @markisfish44893 жыл бұрын

    You are awesome! Thank you for the reviews, informational news, facts, etc.

  • @MattJDylan
    @MattJDylan3 жыл бұрын

    I can't be bothered to watch it right now because it's almost 4.30 AM here, but I appreciate your hard work guys

  • @chincotaco
    @chincotaco3 жыл бұрын

    Is there any info out there about the Max Transition Avoidance Coding? You mention "similar techniques" in network encoding… is this basically just a Gray code? Couldn't find much on the Google. Thanks! Great work as always, learned a ton.

  • @mikelliteras397
    @mikelliteras3973 жыл бұрын

    Congrats on 1 mil subs

  • @andydbedford
    @andydbedford3 жыл бұрын

    GN, I have been waiting for this video, great job 👍

  • @uss_04
    @uss_043 жыл бұрын

    While you have your review samples, do you try to keep the card in a bench 24/7 for sustained /idle temps and early mortality testing?

  • @kashem425
    @kashem4253 жыл бұрын

    Got my GN tool kit today. Fast shipping and very nice ty!

  • @makingPAIN
    @makingPAIN3 жыл бұрын

    So if I wake up at 12am you will have the review live? You guys are gonna have a busy week!!!!

  • @scottgoodson4838
    @scottgoodson48383 жыл бұрын

    Thank you for publishing at this level! As someone who does read the whitepapers, I am quite sure there is no other KZread creator with your finesse and content quality in such coverage. Keep it up, even if you don’t get blockbuster views on these types of videos - they’re truly part of the Gamer’s Nexus culture!

Келесі