NVIDIA G80: Architecture and GPU Analysis
Moderator: Thanas
- Ace Pace
- Hardware Lover
- Posts: 8456
- Joined: 2002-07-07 03:04am
- Location: Wasting time instead of money
- Contact:
NVIDIA G80: Architecture and GPU Analysis
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
Haha! I KNEW that was going to be your reaction! Don't worry, I plan on gathering some data from BF2142, Oblivion and NWN2 tonight with my GX2, and again tomorrow on the GTX (don't expect a lot, though, I want to play with my new toy).Ace Pace wrote:Fuck you, fuck you hard. You WILL bench that thing under realistic conditions.
Also, if there's demand for it, I could do Company of Heroes and Fear timedemos this weekend (but those would be GTX only).
Artillery. Its what's for dinner.
- Ace Pace
- Hardware Lover
- Posts: 8456
- Joined: 2002-07-07 03:04am
- Location: Wasting time instead of money
- Contact:
FEAR timedemos.
Am I reading this right? Is this basicly free AA aslong as the ROPs arn't busy actully writing data to the display?Ah, the ROP hardware. In terms of base pixel quality it's able to perform 8x multisampling using rotated or jittered subsamples laid over a 4-bit subpixel grid, looping through the ROP taking 4 multisamples per cycle. It can multisample from all backbuffer formats too, NVIDIA providing full orthogonality, including sampling from pixels maintained in a non-linear colour space or in floating point surface formats. Thus the misunderstood holy grail of "HDR+AA" is achieved by the hardware with no real developer effort. Further, it can natively blend pixels in integer and floating point formats, including FP32, at rates that somewhat diminish with bandwidth available through the ROP (INT8 and FP16 full speed (measured) and FP32 half speed). Each pair of ROPs share a single blender (so 12 blends per cycle) from testing empirically.
Sample rates in the ROP are hugely increased over previous generation NVIDIA GPUs, a G80 ROP able to empirically sample Z 8 times per clock (4x higher than G7x ever could), with that value scaling for every discrete subsample position, per pixel, bandwidth permitting of course. Concluding 'free' 4xMSAA from the ROP with enough bandwidth is therefore an easy stretch of the imagination to make, and the advantages to certain rendering algorithms become very clear. The 24 ROPs in a full G80 are divided into partitions of 4 each, each partition connecting to a 64-bit memory channel out to the DRAM pool for intermediary or final storage.
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
- InnocentBystander
- The Russian Circus
- Posts: 3466
- Joined: 2004-04-10 06:05am
- Location: Just across the mighty Hudson
Right now? Nothing.
R600, which is should be a February launch, will be unified as well (although supposedly only 64 shaders) and possibly a 512-bit bus with 1GB of RAM. If that last part is true, the R600 will probably be an ultra expensive card, and the extra memory and bandwidth probably won't help it much against the G80. On the flip side, no one knew what the G80 was going to be until a month ago, so ATI may well have some surprises in store.
R600, which is should be a February launch, will be unified as well (although supposedly only 64 shaders) and possibly a 512-bit bus with 1GB of RAM. If that last part is true, the R600 will probably be an ultra expensive card, and the extra memory and bandwidth probably won't help it much against the G80. On the flip side, no one knew what the G80 was going to be until a month ago, so ATI may well have some surprises in store.
Artillery. Its what's for dinner.
I would suspect ATI and NVidia have enough spies in each other's operations or at the very least a good enough team of reverse engineers to compete with anything the other throws out.Arrow wrote: [...] no one knew what the G80 was going to be until a month ago, so ATI may well have some surprises in store.
Children of the Ancients
I'm sorry, but the number you have dialed is imaginary. Please rotate the phone by 90 degrees and try again.
I'm sorry, but the number you have dialed is imaginary. Please rotate the phone by 90 degrees and try again.
- Fingolfin_Noldor
- Emperor's Hand
- Posts: 11834
- Joined: 2006-05-15 10:36am
- Location: At the Helm of the HAB Star Dreadnaught Star Fist
- Captain tycho
- Has Elected to Receive
- Posts: 5039
- Joined: 2002-12-04 06:35pm
- Location: Jewy McJew Land
- NRS Guardian
- Jedi Knight
- Posts: 531
- Joined: 2004-09-11 09:11pm
- Location: Colorado
Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.Captain tycho wrote:I really want one of these, but I simply can't justify spending 500+ dollars on a video card when the one I have now does everything I need it to do at a very acceptable speed, and when DX10 cards aren't that far off.
"It is not necessary to hope in order to persevere."
-William of Nassau, Prince of Orange
Economic Left/Right: 0.88
Social Libertarian/Authoritarian: 2.10
-William of Nassau, Prince of Orange
Economic Left/Right: 0.88
Social Libertarian/Authoritarian: 2.10
Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
- ThatGuyFromThatPlace
- Jedi Knight
- Posts: 691
- Joined: 2006-08-21 12:52am
I on the other hand haven't upgraded since 6 series and have been waiting and saving for the DX10 cards for months.
Can't wait till february.
Can't wait till february.
[img=right]http://www.geocities.com/jamealbeluvien/revolution.jpg[/img]"Nothing here is what it seems. You are not the plucky hero, the Alliance is not an evil empire, and this is not the grand arena."
- The Operative, Serenity
"Everything they've ever "known" has been proven to be wrong. A thousand years ago everybody knew as a fact, that the earth was the center of the universe. Five hundred years ago, they knew it was flat. Fifteen minutes ago, you knew we humans were alone on it. Imagine what you'll know tomorrow."
-Agent Kay, Men In Black
- The Operative, Serenity
"Everything they've ever "known" has been proven to be wrong. A thousand years ago everybody knew as a fact, that the earth was the center of the universe. Five hundred years ago, they knew it was flat. Fifteen minutes ago, you knew we humans were alone on it. Imagine what you'll know tomorrow."
-Agent Kay, Men In Black
- Ace Pace
- Hardware Lover
- Posts: 8456
- Joined: 2002-07-07 03:04am
- Location: Wasting time instead of money
- Contact:
phongn wrote:Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Architecting such a massive GPU has taken NVIDIA a great deal of time and money, four years and $475M to be exact.
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
The 8800 GTX isn't that much more expensive than the 7800 GTX when it launched (I paid $579 for an OC'ed version when it launched). In fact, you could possibly argue that its cheaper, since I needed two 7800 GTXs to drive my 24" LCD (1920x1200), while it looks I'll only need one 8800 GTX for the same display.phongn wrote:Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Artillery. Its what's for dinner.
- Ace Pace
- Hardware Lover
- Posts: 8456
- Joined: 2002-07-07 03:04am
- Location: Wasting time instead of money
- Contact:
No kidding.phongn wrote:Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Eighty. From a 300mm wafer. And thats just gross numbers...Nvidia's isn't handing out exact die size measurements, but they claim to get about 80 chips gross per wafer.
Last edited by Ace Pace on 2006-11-09 09:54am, edited 1 time in total.
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
That's comparing apples and oranges. We're comparing stock board to stock-board, not pre-overclocked variants or SLI.Arrow wrote:The 8800 GTX isn't that much more expensive than the 7800 GTX when it launched (I paid $579 for an OC'ed version when it launched). In fact, you could possibly argue that its cheaper, since I needed two 7800 GTXs to drive my 24" LCD (1920x1200), while it looks I'll only need one 8800 GTX for the same display.
- The Kernel
- Emperor's Hand
- Posts: 7438
- Joined: 2003-09-17 02:31am
- Location: Kweh?!
Mostly this was due to higher costs associated with doing transistor level design apparently.Ace Pace wrote:phongn wrote:Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.Architecting such a massive GPU has taken NVIDIA a great deal of time and money, four years and $475M to be exact.
- Arthur_Tuxedo
- Sith Acolyte
- Posts: 5637
- Joined: 2002-07-23 03:28am
- Location: San Francisco, California
Still, the 7800 GTX was a somewhat anomalous case because there were so many available on launch and priced dropped by like $75 under MSRP scant weeks after release. Top end cards have been pretty consistently priced at $600 or so at release, and this is no exception. The GeForce 3 came out at $600, after all.phongn wrote:That's comparing apples and oranges. We're comparing stock board to stock-board, not pre-overclocked variants or SLI.Arrow wrote:The 8800 GTX isn't that much more expensive than the 7800 GTX when it launched (I paid $579 for an OC'ed version when it launched). In fact, you could possibly argue that its cheaper, since I needed two 7800 GTXs to drive my 24" LCD (1920x1200), while it looks I'll only need one 8800 GTX for the same display.
"I'm so fast that last night I turned off the light switch in my hotel room and was in bed before the room was dark." - Muhammad Ali
"Dating is not supposed to be easy. It's supposed to be a heart-pounding, stomach-wrenching, gut-churning exercise in pitting your fear of rejection and public humiliation against your desire to find a mate. Enjoy." - Darth Wong
"Dating is not supposed to be easy. It's supposed to be a heart-pounding, stomach-wrenching, gut-churning exercise in pitting your fear of rejection and public humiliation against your desire to find a mate. Enjoy." - Darth Wong
Ok, to hell with the price debate. I have my card. I'll post pics and screen shots tomorrow (sorry Ace, but this is just too much!). First impressions:
1) This is EVGA's best packaging yet. It reminds of the same packaging used by a vendor I deal with at work for their $25,000 and (way, way) up FPGA boards. Your card will not be damaged during shipping.
2) Easy install, as expected. Only problem was that I managed to unseat my X-Fi.
3) Ok, I think I'm done tormenting Ace.
4) Maybe not!
5) The first thing you notice with this card over all previous Nvidia cards is the image quality. It is literally a night and day difference. When I fired up 3DMark06 (default settings), I was completely blown away. For a second I wasn't sure it was the same program. The screenshots in the reviews simply do not do this new card justice. Everything looks more real and life like.
6) The Medieval 2 demo absolutely LOVES this card. Using the same settings as the GX2, the game ran two to three times faster (15-20 FPS to 35 to 50 FPS). The G80 has so much power, that the pussy French broke and ran without my noble Englishmen so much as lifting a finger! So I maxed everything out (but the AA - I set it to 4x because I want to play with the 8x CSAA override later) and it was still butter smooth at 1920x1200.
Ok, now its time to try a few other games, but I think I'm going to need to borrow a OMGWTFPWNED image and the Samual L Jackson pic to do a proper critique.
1) This is EVGA's best packaging yet. It reminds of the same packaging used by a vendor I deal with at work for their $25,000 and (way, way) up FPGA boards. Your card will not be damaged during shipping.
2) Easy install, as expected. Only problem was that I managed to unseat my X-Fi.
3) Ok, I think I'm done tormenting Ace.
4) Maybe not!
5) The first thing you notice with this card over all previous Nvidia cards is the image quality. It is literally a night and day difference. When I fired up 3DMark06 (default settings), I was completely blown away. For a second I wasn't sure it was the same program. The screenshots in the reviews simply do not do this new card justice. Everything looks more real and life like.
6) The Medieval 2 demo absolutely LOVES this card. Using the same settings as the GX2, the game ran two to three times faster (15-20 FPS to 35 to 50 FPS). The G80 has so much power, that the pussy French broke and ran without my noble Englishmen so much as lifting a finger! So I maxed everything out (but the AA - I set it to 4x because I want to play with the 8x CSAA override later) and it was still butter smooth at 1920x1200.
Ok, now its time to try a few other games, but I think I'm going to need to borrow a OMGWTFPWNED image and the Samual L Jackson pic to do a proper critique.
Artillery. Its what's for dinner.
Ok, I actually got some benchmarking done.
First up, my system configuration:
Here's the 8800GTX 3DMark06 score:
Right in line with everyone else's stock single card results.
Ace asked to do a Shadermark (v2.1) comparison between the 7950GX2 and the 8800GTX.
Here's the GX2 results
Here's the 8800GTX results
The 8800 pretty much blows the GX2 out of the water on that one.
Ok, now time for Oblivion. I currently have a couple dozen mods install, the ones that affect the following results are the "Low-Poly Grass" mod, the "Unique Landscapes - Ancient Redwoods 1.3" mod and an LOD texture mod. The low-poly looks 99% as good as the stock grass, and it runs better. The Ancient Redwoods is probably the single most demanding area I've ever seen in Oblivion, mod or stock. This test run started at Gottlesfont Priory, and I ran due north until I hit the Black Road, which pretty much goes straight the heart of the Redwoods mod. I stuck the same path as much as humanly possible for these runs.
7950GX2: 2xAF, 1/3 grass, 1/4 item, 1/3 shadows, HDR, all else max
Frames: 1973 - Time: 73047ms - Avg: 27.010 - Min: 15 - Max: 44
8800GTX: Default AF, 1/3 grass, 1/4 item, 1/3 shadows, HDR, all else max
Frames: 3179 - Time: 71654ms - Avg: 44.366 - Min: 15 - Max: 63
8800GTX: 8x CSAA, 16xAF, HDR, EVERYTHING maxed!
Frames: 2178 - Time: 71695ms - Avg: 30.379 - Min: 18 - Max: 45
Uh, yeah, I'll play with everything cranked, thankyouverymuch!
And here's some screenshots of Oblivion with the maxed out settings. Also, you should note that I turned on all reflections in the .ini file, and each screenshot has the FPS stamp in the upper left. 56k download warning applies as well.
Redwoods
Crowd at War
The above screen is should be very representative of the worst case with the max settings, and the game was still playable.
Here's a nice HDR shot
Lighting on armor
The lake from the Ancient Yews mod
Swamps near Fort Teleman
And last but not least, Hardware p0rn (all in 7.2 megapixel glory!):
The Box
The Card
The Computer
And fuck, I'm out of webspace...
First up, my system configuration:
Here's the 8800GTX 3DMark06 score:
Right in line with everyone else's stock single card results.
Ace asked to do a Shadermark (v2.1) comparison between the 7950GX2 and the 8800GTX.
Here's the GX2 results
Here's the 8800GTX results
The 8800 pretty much blows the GX2 out of the water on that one.
Ok, now time for Oblivion. I currently have a couple dozen mods install, the ones that affect the following results are the "Low-Poly Grass" mod, the "Unique Landscapes - Ancient Redwoods 1.3" mod and an LOD texture mod. The low-poly looks 99% as good as the stock grass, and it runs better. The Ancient Redwoods is probably the single most demanding area I've ever seen in Oblivion, mod or stock. This test run started at Gottlesfont Priory, and I ran due north until I hit the Black Road, which pretty much goes straight the heart of the Redwoods mod. I stuck the same path as much as humanly possible for these runs.
7950GX2: 2xAF, 1/3 grass, 1/4 item, 1/3 shadows, HDR, all else max
Frames: 1973 - Time: 73047ms - Avg: 27.010 - Min: 15 - Max: 44
8800GTX: Default AF, 1/3 grass, 1/4 item, 1/3 shadows, HDR, all else max
Frames: 3179 - Time: 71654ms - Avg: 44.366 - Min: 15 - Max: 63
8800GTX: 8x CSAA, 16xAF, HDR, EVERYTHING maxed!
Frames: 2178 - Time: 71695ms - Avg: 30.379 - Min: 18 - Max: 45
Uh, yeah, I'll play with everything cranked, thankyouverymuch!
And here's some screenshots of Oblivion with the maxed out settings. Also, you should note that I turned on all reflections in the .ini file, and each screenshot has the FPS stamp in the upper left. 56k download warning applies as well.
Redwoods
Crowd at War
The above screen is should be very representative of the worst case with the max settings, and the game was still playable.
Here's a nice HDR shot
Lighting on armor
The lake from the Ancient Yews mod
Swamps near Fort Teleman
And last but not least, Hardware p0rn (all in 7.2 megapixel glory!):
The Box
The Card
The Computer
And fuck, I'm out of webspace...
Artillery. Its what's for dinner.
- Uraniun235
- Emperor's Hand
- Posts: 13772
- Joined: 2002-09-12 12:47am
- Location: OREGON
- Contact:
What's that card next to the vid card?
"There is no "taboo" on using nuclear weapons." -Julhelm
What is Project Zohar?
"On a serious note (well not really) I did sometimes jump in and rate nBSG episodes a '5' before the episode even aired or I saw it." - RogueIce explaining that episode ratings on SDN tv show threads are bunk
"On a serious note (well not really) I did sometimes jump in and rate nBSG episodes a '5' before the episode even aired or I saw it." - RogueIce explaining that episode ratings on SDN tv show threads are bunk