• Hey, guest user. Hope you're enjoying NeoGAF! Have you considered registering for an account? Come join us and add your take to the daily discourse.
  • Hey Guest. Check out the NeoGAF 2.2 Update Thread for details on our new Giphy integration and other new features.

NVIDIA GeForce RTX 3090 to have 24GB GDDR6X VRAM, RTX 3080 10GB VRAM

supernova8

Member
Jun 2, 2020
1,720
2,393
420
Current predictions look right +- 5%


The fucking nerve if they charge almost double 3080 for the 3090 when it's only slightly better. Plus, $400 for the 3060 is a fuckin dick move. I hope AMD wipes the floor with their 3060 and shows NVIDIA out to be the price-gouging bastards they've always been.

edit: probably a little too harsh. I guess if you're getting close to 2080 Ti performance for 1/3 of the launch price it's not that bad but I still hope AMD comes out with a real corker.
 
Last edited:
  • LOL
Reactions: TyphoonStruggle

Rikkori

Member
May 9, 2020
2,232
4,134
410
The fucking nerve if they charge almost double 3080 for the 3090 when it's only slightly better. Plus, $400 for the 3060 is a fuckin dick move. I hope AMD wipes the floor with their 3060 and shows NVIDIA out to be the price-gouging bastards they've always been.
That's why they're giving it 24 GBs. Have to say it does make me sad how much they gimped the vram. My only hope is that nvcache rumours are true.
 

Spukc

Member
Jan 24, 2015
16,672
17,455
920
The fucking nerve if they charge almost double 3080 for the 3090 when it's only slightly better. Plus, $400 for the 3060 is a fuckin dick move. I hope AMD wipes the floor with their 3060 and shows NVIDIA out to be the price-gouging bastards they've always been.
good luck waiting for that magical AMD card that will never release
 
  • Like
Reactions: Noboru Wataya

supernova8

Member
Jun 2, 2020
1,720
2,393
420
good luck waiting for that magical AMD card that will never release

To be clear I'm not expecting them to beat NVIDIA in performance, I'm just expecting a relatively competitive card at a low price. Since AMD has pretty much dominated Intel in CPUs, there's more headroom to turn their attention to GPUs.

They spent a few years getting Ryzen into full swing and I think last year's 5700XT was the first sign of AMD starting to get back into the GPU fight.

Just remember when Ryzen first came out and the reviews were like "yeah pretty good but err.. stick with Intel for now" and then the next year it was like "shiiiiiiiit". I doubt we're that far away from the "shiiiiit" moment with Radeon. We've already seen what they can pack into a very tight power envelope for RDNA2 with the next-gen consoles so imagine that as a full-fat GPU and you have a potential killer on your hands.
 
Last edited:

Spukc

Member
Jan 24, 2015
16,672
17,455
920
To be clear I'm not expecting them to beat NVIDIA in performance, I'm just expecting a relatively competitive card at a low price. Since AMD has pretty much dominated Intel in CPUs, there's more headroom to turn their attention to GPUs.

They spent a few years getting Ryzen into full swing and I think last year's 5700XT was the first sign of AMD starting to get back into the GPU fight.

Just remember when Ryzen first came out and the reviews were like "yeah pretty good but err.. stick with Intel for now" and then the next year it was like "shiiiiiiiit". I doubt we're that far away from the "shiiiiit" moment with AMD. We've already seen what they can pack into a very tight power envelope for RDNA2 with the next-gen consoles so imagine that as a full-fat GPU and you have a potential killer on your hands.
just around the corner yeah ANY DAY NOW :pie_roffles:
 

supernova8

Member
Jun 2, 2020
1,720
2,393
420
just around the corner yeah ANY DAY NOW :pie_roffles:

Haha yeah I know it's somewhat of a running joke. Totally get that. But that was back when AMD was totally hopeless with no money, no hope. It's a different company now.
 

Spukc

Member
Jan 24, 2015
16,672
17,455
920
Haha yeah I know it's somewhat of a running joke. Totally get that. But that was back when AMD was totally hopeless with no money, no hope. It's a different company now.
Aye i would be the first in line to stand having a laugh at team green getting fucked by AMD like they did intel.
But i honestly can't care until the moment they release something. The last stuff AMD did was a joke. And people should not care about them until they show something.
 
Last edited:

Ellery

Member
Feb 17, 2019
1,611
2,977
525
35
Manches, Amsterdam, Berlin
I think it is always best to be realistic about AMD. I don't know how much R&D money they have now, but it is probably 1/5th or less than what Nvidia has.

And I guess we can take it for a fact now that Ampere is going to be a gigantic GPU with extremely high power draw and if Nvidia sees the need to bring something this brutal then it will be very hard for AMD to catch up with that. I would still expect the Big Navi to slot in somewhere between the 2080 Ti and the RTX 3090, but I also expect the gap between the 2080 Ti and 3090 to be atleast 60% in terms of raw performance we see in gaming and synthetic benchmarks.

Going from a 5700 XT to something that would compete with a 2080 Ti + 60% is probably way out of reach. I hope they can make a well rounded high end GPU that competes with the RTX 3080.
 

supernova8

Member
Jun 2, 2020
1,720
2,393
420
Aye i would be the first in line to stand having a laugh at team green getting fucked by AMD like they did intel.
But i honestly can't care until the moment they release something. The last stuff AMD did was a joke. And people should not care about them until they show something.

5700XT seemed to review pretty well but it came out too late. I believe it was July 2019? RTX 2060 was already out by Jan 2019 so anybody looking for a mid range card probably had that. Plus, NVIDIA killed off the 5700XT with the Super cards before AMD even got their products out.

I guess it might be fair to say AMD has actually made some good progress in the GPU space, the only problem is that they're up against actual competition.
 
  • Like
Reactions: Coulomb_Barrier
Sep 25, 2016
2,049
516
540
What will the 3070 and 3080 use power draw wise vs the 2070 I currently have? I don't want to upgrade PSU. Are those two standard 8pin too?

I personally can't even imagine going with AMD just because of DLSS. It's basically 100% more performance in the games that support it if you're targeting 4k.
 
Last edited:
  • LOL
Reactions: AquaticSquirrel

supernova8

Member
Jun 2, 2020
1,720
2,393
420
I think it is always best to be realistic about AMD. I don't know how much R&D money they have now, but it is probably 1/5th or less than what Nvidia has.

And I guess we can take it for a fact now that Ampere is going to be a gigantic GPU with extremely high power draw and if Nvidia sees the need to bring something this brutal then it will be very hard for AMD to catch up with that. I would still expect the Big Navi to slot in somewhere between the 2080 Ti and the RTX 3090, but I also expect the gap between the 2080 Ti and 3090 to be atleast 60% in terms of raw performance we see in gaming and synthetic benchmarks.

Going from a 5700 XT to something that would compete with a 2080 Ti + 60% is probably way out of reach. I hope they can make a well rounded high end GPU that competes with the RTX 3080.

(1) If you think NVIDIA is bigger than AMD, have you seen how much bigger Intel is than AMD? Personally I'd suggest NVIDIA going for a massive GPU with an extremely high power draw suggests they've been spooked a little by AMD.

(2) Don't forget they went from Excavator CPUs to Ryzen in the space of two years (I'm not saying it took 2 years to make it). Plus since Raja Koduri left AMD, Lisa Su has been running Radeon herself in the interim. I think we should be cautiously optimistic about what AMD produces.
 

Ellery

Member
Feb 17, 2019
1,611
2,977
525
35
Manches, Amsterdam, Berlin
(1) If you think NVIDIA is bigger than AMD, have you seen how much bigger Intel is than AMD? Personally I'd suggest NVIDIA going for a massive GPU with an extremely high power draw suggests they've been spooked a little by AMD.

(2) Don't forget they went from Excavator CPUs to Ryzen in the space of two years (I'm not saying it took 2 years to make it). Plus since Raja Koduri left AMD, Lisa Su has been running Radeon herself in the interim. I think we should be cautiously optimistic about what AMD produces.

I agree with both. Currently rocking a Ryzen CPU 3700X because I was extremely impressed by the impact Lisa Su had and the stock went through the roof ever since she took over. The 5700 XT looks promising, but unlike Intel I don't see Nvidia screwing up. They seem much more focused and passionate about pushing.
 

magaman

Banned
Oct 19, 2017
546
998
360
The fucking nerve if they charge almost double 3080 for the 3090 when it's only slightly better. Plus, $400 for the 3060 is a fuckin dick move. I hope AMD wipes the floor with their 3060 and shows NVIDIA out to be the price-gouging bastards they've always been.

edit: probably a little too harsh. I guess if you're getting close to 2080 Ti performance for 1/3 of the launch price it's not that bad but I still hope AMD comes out with a real corker.

Make more money. Then you can afford nicer things.
 

DeaDPooL_jlp

Member
Sep 20, 2015
2,584
3,726
580
3080ti with the rumored 20gb sounds perfect for me but of course its not launching in September and I really want a ray tracing card that can push cyberpunk in November so 3090 looks to be what I need.
 
Last edited:
  • Empathy
Reactions: BlueXImpulse

llien

Member
Feb 1, 2017
9,844
7,763
935
To be clear I'm not expecting them to beat NVIDIA in performance,
5700 series do exactly that: beat similarly priced NV cards on performance.

until the moment they release something.

5700 was a joke? Release "Something"?
What the heck is wrong with you?

Ampere is going to be a gigantic GPU with extremely high power draw and if Nvidia sees the need to bring something this brutal then it will be very hard for AMD to catch up with that.
What does "catching up with that" even mean? We are talking about less than 1% of the GPU market.
 
Last edited:

scraz

Member
Feb 10, 2014
153
122
520
So whats the chance the 3080Ti has higher clocks then the 3090 at this point?
 

dave_d

Member
Jan 31, 2005
1,855
678
1,580
Current predictions look right +- 5%

I'm wondering if the 3070 at $600 is the 8GB or 16GB model. (Since earlier on there were rumors that there were going to be 2 models of it, one launching in Sept, and the other in Oct)
 

kiphalfton

Member
Dec 27, 2018
1,098
1,455
470
3090 is nowhere near a titan, titan v-ram pool should be 32gb+. 24gb is basically what u would expect the 3080ti to sit at.

then about 3080.

3080 is a budget card sold for top end money. Anything below 16gb of v-ram has no future simple as that.

I would not even consider the card as a 3060 option.

People are currently looking through this generation glasses. Next generation baseline for v-ram will be 16gb of memory. the base line will go up massively.

This reminds me of a 580 1,5gb, totally solid gpu performance and memory performance, v-ram was fine for ultra settings on any game until PS4 games arrived. It couldn't even run unity or watch dogs on any setting because v-ram bottleneck. Now it probably won't be as bad for the 3080, but expect medium settings being a thing for the card through the entire generation if not low with it.

Any gpu at 400+ when next gen consoles are out needs 16gb of v-ram modules.

10gb makes absolute no sense at all in any way other then they eyeballed for that number. The only way that card has any reason to exist is if its dirt cheap like 200 bucks. And we all know that won't be the case.

It's clear to me that the entire 3000 series besides the 3090 is going to be replaced in 6 months with proper versions when amd has there cards on the market. Anybody buying into these cards will be burned hard.

And for me with a 1080ti, i will be sitting this clusterfuck of a card generation out until they offer something half decent.

That doesn't make sense. 780Ti > 980Ti > 1080Ti > 2080Ti was 3GB > 6GB > 11GB > 11GB. If what you said is true, and the RTX 3080Ti is simply the RTX 3090, RAM wouldn't just jump from 11GB to 24GB "just because", especially when purported RAM for the RTX 3070 and RTX 3080 are pretty close to what they were for Turing.

The RTX 3090 being a Titan replacement would make sense as Nvidia isn't going to indirectly admit they overpriced the Turing Titan by releasing an Ampere Titan for less. $2500 was too much, and they know it. Instead they change the name, and act like them pricing the Turing Titan at $2500 didn't happen. It's correcting course, for pricing, as they went from $1200 with the Titan Xp to $2500 for the Titan RTX. Same thing with the RTX 3080Ti is probably going to happen, where it's obviously not going to be as expensive as the RTX 2080Ti, since pricing wise it would alot in right between the RTX 3080 and RTX 3090 (i.e. somewhere between $800 and $1400, assuming leaked prices are correct).
 
Last edited:
  • Thoughtful
Reactions: llien

GHG

Member
Nov 9, 2006
23,983
28,817
1,845
Leaked specs in one chart:



Also:

NVIDIA GeForce RTX 3090 features GA102-300 GPU with 5248 cores and 24GB of GDDR6X memory across a 384-bit bus. This gives a maximum theoretical bandwidth of 936 GB/s. Custom boards are powered by dual 8-pin power connectors which are definitely required since the card has a TGP of 350W.

The GeForce RTX 3080 gets 4352 CUDA cores and 10GB of GDDR6X memory. This card will have a maximum bandwidth of 760 GB/s thanks to the 320-bit memory bus and 19 Gbps memory modules. This SKU has a TGP of 320W and the custom models that we saw also require dual 8-pin connectors.


Good news on the dual 8 pins for both the 3090 and 3080.

Come on EVGA, don't let me down.
 

Rikkori

Member
May 9, 2020
2,232
4,134
410
I'm wondering if the 3070 at $600 is the 8GB or 16GB model. (Since earlier on there were rumors that there were going to be 2 models of it, one launching in Sept, and the other in Oct)
That would be 8 GB. We have no leaks about the double vram variants besides that they might exist. Sadly I expect those to be $100-$200 more.
 
Sep 29, 2011
4,249
2,870
850
Man, I used to get so excited when a new gen of videocards was about to release. Maybe it's all in my head, but it feels like NV just keeps getting more and more [blatantly] money-grubby as the years go by. Like, fuck me in the ass if you absolutely must, but don't laugh maniacally in my ear the whole time you're doing it.
 
  • Like
Reactions: Dr_Salt

FireFly

Member
Aug 5, 2007
1,491
1,046
1,300
That doesn't make sense. 780Ti > 980Ti > 1080Ti > 2080Ti was 3GB > 6GB > 11GB > 11GB. If what you said is true, and the RTX 3080Ti is simply the RTX 3090, RAM wouldn't just jump from 11GB to 24GB "just because", especially when purported RAM for the RTX 3070 and RTX 3080 are pretty close to what they were for Turing.

The RTX 3090 being a Titan replacement would make sense as Nvidia isn't going to indirectly admit they overpriced the Turing Titan by releasing an Ampere Titan for less. $2500 was too much, and they know it. Instead they change the name, and act like them pricing the Turing Titan at $2500 didn't happen. It's correcting course, for pricing, as they went from $1200 with the Titan Xp to $2500 for the Titan RTX. Same thing with the RTX 3080Ti is probably going to happen, where it's obviously not going to be as expensive as the RTX 2080Ti, since pricing wise it would alot in right between the RTX 3080 and RTX 3090 (i.e. somewhere between $800 and $1400, assuming leaked prices are correct).
The reason for the jump is likely because on the 3090 they're using a 384-bit bus with 12 x 32-bit memory controllers that would otherwise have been addressing 1 GB of memory each. So if Micron only do GDDR6X in 1GB and 2 GB flavours, which seems plausible, their choices for a "full" configuration are 12 GB or 24GB. And increasing memory by only 1 GB over the 2080 Ti probably wasn't going to cut it, especially if AMD are going for 16 GB.

The Titan is $2500 because it's not a gaming card, but rather intended for developers and scientists. In the context of a business purchase, I don't think $2500 is a huge amount of money, and if it wasn't selling, Nvidia was free to reduce pricing. I presume that if a successor to the Titan was created, it would be based on Nvidia's A100 architecture.
 
Last edited:
Mar 23, 2018
6,957
8,351
765
Man, I used to get so excited when a new gen of videocards was about to release. Maybe it's all in my head, but it feels like NV just keeps getting more and more [blatantly] money-grubby as the years go by. Like, fuck me in the ass if you absolutely must, but don't laugh maniacally in my ear the whole time you're doing it.

It's because of no competition really. I hope AMD curp stomps nvidia to the shitter this generation. But who am i kidding i have zero faith in there gpu department.
 
Jan 7, 2018
1,936
2,227
590
I thought 5700XT was one terrific card from AMD. 2080 perf for half the price. [2080 launched at $700]

RDNA2 replacement of 5700xt [6700xt?] should put it in the 2080ti territory with Raytracing support for what 1/3rd the cost. Simply put amazing. But AMD definitely need something to counter DLSS.
 

martino

Member
Apr 25, 2013
6,034
3,851
705
France
I thought 5700XT was one terrific card from AMD. 2080 perf for half the price. [2080 launched at $700]

RDNA2 replacement of 5700xt [6700xt?] should put it in the 2080ti territory with Raytracing support for what 1/3rd the cost. Simply put amazing. But AMD definitely need something to counter DLSS.
it would be great and would make sure i won't buy nvidia card before long.

edit : i forgot there is also hdmi 2.1 support for tv.
 
Last edited:
  • Like
Reactions: MightySquirrel

GHG

Member
Nov 9, 2006
23,983
28,817
1,845
I thought 5700XT was one terrific card from AMD. 2080 perf for half the price. [2080 launched at $700]

RDNA2 replacement of 5700xt [6700xt?] should put it in the 2080ti territory with Raytracing support for what 1/3rd the cost. Simply put amazing. But AMD definitely need something to counter DLSS.

The first one of the two to implement something like DLSS at the driver level wins.

Knowing Nvidia they will probably do it but make it exclusive to the 3xxx series onwards.
 

supernova8

Member
Jun 2, 2020
1,720
2,393
420
5700 series do exactly that: beat similarly priced NV cards on performance.

5700XT is better than 2060 Super in most games and is often neck and neck with 2070 Super but I suppose if you consider stuff like ray tracing and DLSS, the NVIDIA cards are better.
 

baphomet

Member
Oct 29, 2011
6,471
1,972
870
Denver, CO
so are we expecting these to go on sale on the 1st, or will they be at some point later in the month?

going by say the 2000 series?
 

llien

Member
Feb 1, 2017
9,844
7,763
935
if you consider stuff like ray tracing
Yeah, let's consider stuff like "ray tracing",which % of games supports it, or, which % of games needs it, to look great:


Looks like 0 (zero, null).

Saying "but dlss" is basically saying "but fancy upscaling".

Well at the point of writing this post I was refering to having the best infrastructure to deliver the best GPU.
We have exactly zero RTX Titan owners in this thread, the "the best infrastructure" song is clearly not yours.
 
  • Like
Reactions: Coulomb_Barrier

GymWolf

Member
Jun 11, 2019
15,012
21,833
615
It seems like many people are gonna wait for the inevitable 3080super...

The question is, can i survive with my 2070s for another 12-18 months before that thing come out? Is it enough to play cyberpunk at 1440p60 with decent details? This shit can't even sustain 1440p60 ultra in remnant of the fucking ashes with hell on screen:ROFLMAO:

It's bad...
 
Last edited:
Feb 15, 2013
8,900
7,583
855
London
I'm in the market for a $200-300 replacement for my 1070 Ti. Looks like I'll be lucky to get a GeForce 3040 with that money seeing as how they keep increasing prices.
 

Ellery

Member
Feb 17, 2019
1,611
2,977
525
35
Manches, Amsterdam, Berlin
Why wouldn't they just go with 12gb for the 3080?

  1. For the next few years 10GB is enough for 1440p.
  2. Good 4K monitors are expensive. Nvidias knows their audience and they can milk 4K gamers for more money with the 3090. Those people have no alternative anyway.
  3. Having just 10GB is cheaper for Nvidia. VRAM is one of the most expensive things for them.
  4. A 10GB card puts people on a more frequent upgrade interval, because they sooner realize that the VRAM is a potential bottleneck.
  5. People look at performance first in reviews and then at the price. The price would go up with more VRAM, but the performance wouldn't
  6. GPU Reviews are a reflection of the current PC gaming landscape with titles from the last few years and the 3080 10GB will be enough for them and reviewers won't run into VRAM shortcomings.
  7. Based on that people are going to see that 10GB is enough for them currently. What the future holds nobody knows.
  8. The 10GB 3080 gives a very clear distinction to the 24GB 3090 and it makes the extremely expensive 1500$ RTX 3090 look like a reasonable buy because it flaunts so much VRAM. Basic consumer psychology.
  9. It leaves more room for future cards that may slot between the 3080 and 3090 to be impressive. Nvidia can go with anything between 10 and 20gb then and people will be impressed.
 
  • Like
Reactions: Rikkori

Eliciel

Member
Oct 30, 2009
1,671
824
1,095
Would anyone recommend me selling 5700XT right now and rather focus on the new cards? What do you say..