NVIDIA G80: Architecture and GPU Analysis

GEC: Discuss gaming, computers and electronics and venture into the bizarre world of STGODs.

Moderator: Thanas

User avatar
Ace Pace
Hardware Lover
Posts: 8456
Joined: 2002-07-07 03:04am
Location: Wasting time instead of money
Contact:

NVIDIA G80: Architecture and GPU Analysis

Post by Ace Pace »

Brotherhood of the Bear | HAB | Mess | SDnet archivist |
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

I'm reading the HardOCP review right now. I'm LOVING this new architecture.

And my EVGA 8800 GTX will be here tomorrow! :twisted:
Artillery. Its what's for dinner.
User avatar
Ace Pace
Hardware Lover
Posts: 8456
Joined: 2002-07-07 03:04am
Location: Wasting time instead of money
Contact:

Post by Ace Pace »

Fuck you, fuck you hard. You WILL bench that thing under realistic conditions.

I'm reading the beyond3d review first. Discussion of actual benchs to the other thread.
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

Ace Pace wrote:Fuck you, fuck you hard. You WILL bench that thing under realistic conditions.
Haha! I KNEW that was going to be your reaction! Don't worry, I plan on gathering some data from BF2142, Oblivion and NWN2 tonight with my GX2, and again tomorrow on the GTX (don't expect a lot, though, I want to play with my new toy).

Also, if there's demand for it, I could do Company of Heroes and Fear timedemos this weekend (but those would be GTX only).
Artillery. Its what's for dinner.
User avatar
Ace Pace
Hardware Lover
Posts: 8456
Joined: 2002-07-07 03:04am
Location: Wasting time instead of money
Contact:

Post by Ace Pace »

FEAR timedemos.
Ah, the ROP hardware. In terms of base pixel quality it's able to perform 8x multisampling using rotated or jittered subsamples laid over a 4-bit subpixel grid, looping through the ROP taking 4 multisamples per cycle. It can multisample from all backbuffer formats too, NVIDIA providing full orthogonality, including sampling from pixels maintained in a non-linear colour space or in floating point surface formats. Thus the misunderstood holy grail of "HDR+AA" is achieved by the hardware with no real developer effort. Further, it can natively blend pixels in integer and floating point formats, including FP32, at rates that somewhat diminish with bandwidth available through the ROP (INT8 and FP16 full speed (measured) and FP32 half speed). Each pair of ROPs share a single blender (so 12 blends per cycle) from testing empirically.

Sample rates in the ROP are hugely increased over previous generation NVIDIA GPUs, a G80 ROP able to empirically sample Z 8 times per clock (4x higher than G7x ever could), with that value scaling for every discrete subsample position, per pixel, bandwidth permitting of course. Concluding 'free' 4xMSAA from the ROP with enough bandwidth is therefore an easy stretch of the imagination to make, and the advantages to certain rendering algorithms become very clear. The 24 ROPs in a full G80 are divided into partitions of 4 each, each partition connecting to a 64-bit memory channel out to the DRAM pool for intermediary or final storage.
Am I reading this right? Is this basicly free AA aslong as the ROPs arn't busy actully writing data to the display?
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
User avatar
Ace Pace
Hardware Lover
Posts: 8456
Joined: 2002-07-07 03:04am
Location: Wasting time instead of money
Contact:

Post by Ace Pace »

To add, free AF.
anisotropic filtering is effectively pretty much free, depending on your game and target resolution of course.
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

AA's not free. It depends on what else is going on, but it certainly is less of hit than before.

Now, since Nvidia has the CUDA engine now, I need to see if I can get my boss to get on for my work machine for *cough* DSP work *cough*.
Artillery. Its what's for dinner.
User avatar
InnocentBystander
The Russian Circus
Posts: 3466
Joined: 2004-04-10 06:05am
Location: Just across the mighty Hudson

Post by InnocentBystander »

So Nvidia brings some cool shit to the table. What has ATI got?
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

Right now? Nothing.

R600, which is should be a February launch, will be unified as well (although supposedly only 64 shaders) and possibly a 512-bit bus with 1GB of RAM. If that last part is true, the R600 will probably be an ultra expensive card, and the extra memory and bandwidth probably won't help it much against the G80. On the flip side, no one knew what the G80 was going to be until a month ago, so ATI may well have some surprises in store.
Artillery. Its what's for dinner.
User avatar
Jaepheth
Jedi Master
Posts: 1055
Joined: 2004-03-18 02:13am
Location: between epsilon and zero

Post by Jaepheth »

Arrow wrote: [...] no one knew what the G80 was going to be until a month ago, so ATI may well have some surprises in store.
I would suspect ATI and NVidia have enough spies in each other's operations or at the very least a good enough team of reverse engineers to compete with anything the other throws out.
Children of the Ancients
I'm sorry, but the number you have dialed is imaginary. Please rotate the phone by 90 degrees and try again.
User avatar
Fingolfin_Noldor
Emperor's Hand
Posts: 11834
Joined: 2006-05-15 10:36am
Location: At the Helm of the HAB Star Dreadnaught Star Fist

Post by Fingolfin_Noldor »

I'm sorely tempted to get the 8800GTS, but I just upgraded a few months back to a 1900XT 512mb from a 6600GT. So it's really too early to upgrade.

I doubt ATI is that stupid to launch a card that can't match NVidia's best.
User avatar
Captain tycho
Has Elected to Receive
Posts: 5039
Joined: 2002-12-04 06:35pm
Location: Jewy McJew Land

Post by Captain tycho »

I really want one of these, but I simply can't justify spending 500+ dollars on a video card when the one I have now does everything I need it to do at a very acceptable speed, and when DX10 cards aren't that far off.
Captain Tycho!
The worst fucker ever!
The Best reciever ever!
User avatar
NRS Guardian
Jedi Knight
Posts: 531
Joined: 2004-09-11 09:11pm
Location: Colorado

Post by NRS Guardian »

Captain tycho wrote:I really want one of these, but I simply can't justify spending 500+ dollars on a video card when the one I have now does everything I need it to do at a very acceptable speed, and when DX10 cards aren't that far off.
Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
"It is not necessary to hope in order to persevere."
-William of Nassau, Prince of Orange

Economic Left/Right: 0.88
Social Libertarian/Authoritarian: 2.10
User avatar
phongn
Rebel Leader
Posts: 18487
Joined: 2002-07-03 11:11pm

Post by phongn »

NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)
User avatar
ThatGuyFromThatPlace
Jedi Knight
Posts: 691
Joined: 2006-08-21 12:52am

Post by ThatGuyFromThatPlace »

I on the other hand haven't upgraded since 6 series and have been waiting and saving for the DX10 cards for months.
Can't wait till february.
[img=right]http://www.geocities.com/jamealbeluvien/revolution.jpg[/img]"Nothing here is what it seems. You are not the plucky hero, the Alliance is not an evil empire, and this is not the grand arena."
- The Operative, Serenity
"Everything they've ever "known" has been proven to be wrong. A thousand years ago everybody knew as a fact, that the earth was the center of the universe. Five hundred years ago, they knew it was flat. Fifteen minutes ago, you knew we humans were alone on it. Imagine what you'll know tomorrow."
-Agent Kay, Men In Black
User avatar
Ace Pace
Hardware Lover
Posts: 8456
Joined: 2002-07-07 03:04am
Location: Wasting time instead of money
Contact:

Post by Ace Pace »

phongn wrote:
NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)
Architecting such a massive GPU has taken NVIDIA a great deal of time and money, four years and $475M to be exact.
:shock:
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

phongn wrote:
NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)
The 8800 GTX isn't that much more expensive than the 7800 GTX when it launched (I paid $579 for an OC'ed version when it launched). In fact, you could possibly argue that its cheaper, since I needed two 7800 GTXs to drive my 24" LCD (1920x1200), while it looks I'll only need one 8800 GTX for the same display.
Artillery. Its what's for dinner.
User avatar
Ace Pace
Hardware Lover
Posts: 8456
Joined: 2002-07-07 03:04am
Location: Wasting time instead of money
Contact:

Post by Ace Pace »

phongn wrote:
NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)
No kidding.
Nvidia's isn't handing out exact die size measurements, but they claim to get about 80 chips gross per wafer.
Eighty. From a 300mm wafer. And thats just gross numbers...

:shock:
Last edited by Ace Pace on 2006-11-09 09:54am, edited 1 time in total.
Brotherhood of the Bear | HAB | Mess | SDnet archivist |
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

For those of you that don't want to go through 15-page review after 15-page review, Ars has a great one page summary of the new architecture and what it means.
Artillery. Its what's for dinner.
User avatar
phongn
Rebel Leader
Posts: 18487
Joined: 2002-07-03 11:11pm

Post by phongn »

Arrow wrote:The 8800 GTX isn't that much more expensive than the 7800 GTX when it launched (I paid $579 for an OC'ed version when it launched). In fact, you could possibly argue that its cheaper, since I needed two 7800 GTXs to drive my 24" LCD (1920x1200), while it looks I'll only need one 8800 GTX for the same display.
That's comparing apples and oranges. We're comparing stock board to stock-board, not pre-overclocked variants or SLI.
User avatar
The Kernel
Emperor's Hand
Posts: 7438
Joined: 2003-09-17 02:31am
Location: Kweh?!

Post by The Kernel »

Ace Pace wrote:
phongn wrote:
NRS Guardian wrote:Uh... this is a DX10 card, the first DX10 card to be released. That's why it's so expensive.
Well, that and a 680 million transistor GPU at 90nm probably has rather low yields. By comparison, the newest x86 CPUs have less than 300 million at 65 nm. RAM for GPUs is also more expensive as well (being specialized, high-performance parts)
Architecting such a massive GPU has taken NVIDIA a great deal of time and money, four years and $475M to be exact.
:shock:
Mostly this was due to higher costs associated with doing transistor level design apparently.
User avatar
Arthur_Tuxedo
Sith Acolyte
Posts: 5637
Joined: 2002-07-23 03:28am
Location: San Francisco, California

Post by Arthur_Tuxedo »

phongn wrote:
Arrow wrote:The 8800 GTX isn't that much more expensive than the 7800 GTX when it launched (I paid $579 for an OC'ed version when it launched). In fact, you could possibly argue that its cheaper, since I needed two 7800 GTXs to drive my 24" LCD (1920x1200), while it looks I'll only need one 8800 GTX for the same display.
That's comparing apples and oranges. We're comparing stock board to stock-board, not pre-overclocked variants or SLI.
Still, the 7800 GTX was a somewhat anomalous case because there were so many available on launch and priced dropped by like $75 under MSRP scant weeks after release. Top end cards have been pretty consistently priced at $600 or so at release, and this is no exception. The GeForce 3 came out at $600, after all.
"I'm so fast that last night I turned off the light switch in my hotel room and was in bed before the room was dark." - Muhammad Ali

"Dating is not supposed to be easy. It's supposed to be a heart-pounding, stomach-wrenching, gut-churning exercise in pitting your fear of rejection and public humiliation against your desire to find a mate. Enjoy." - Darth Wong
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

Ok, to hell with the price debate. I have my card. I'll post pics and screen shots tomorrow (sorry Ace, but this is just too much!). First impressions:

1) This is EVGA's best packaging yet. It reminds of the same packaging used by a vendor I deal with at work for their $25,000 and (way, way) up FPGA boards. Your card will not be damaged during shipping.

2) Easy install, as expected. Only problem was that I managed to unseat my X-Fi.

3) Ok, I think I'm done tormenting Ace.

4) Maybe not! ;)

5) The first thing you notice with this card over all previous Nvidia cards is the image quality. It is literally a night and day difference. When I fired up 3DMark06 (default settings), I was completely blown away. For a second I wasn't sure it was the same program. The screenshots in the reviews simply do not do this new card justice. Everything looks more real and life like.

6) The Medieval 2 demo absolutely LOVES this card. Using the same settings as the GX2, the game ran two to three times faster (15-20 FPS to 35 to 50 FPS). The G80 has so much power, that the pussy French broke and ran without my noble Englishmen so much as lifting a finger! So I maxed everything out (but the AA - I set it to 4x because I want to play with the 8x CSAA override later) and it was still butter smooth at 1920x1200.

Ok, now its time to try a few other games, but I think I'm going to need to borrow a OMGWTFPWNED image and the Samual L Jackson pic to do a proper critique.
Artillery. Its what's for dinner.
User avatar
Arrow
Jedi Council Member
Posts: 2283
Joined: 2003-01-12 09:14pm

Post by Arrow »

Ok, I actually got some benchmarking done.

First up, my system configuration:

Image
Image
Image

Here's the 8800GTX 3DMark06 score:
Image
Right in line with everyone else's stock single card results.

Ace asked to do a Shadermark (v2.1) comparison between the 7950GX2 and the 8800GTX.

Here's the GX2 results

Here's the 8800GTX results

The 8800 pretty much blows the GX2 out of the water on that one.

Ok, now time for Oblivion. I currently have a couple dozen mods install, the ones that affect the following results are the "Low-Poly Grass" mod, the "Unique Landscapes - Ancient Redwoods 1.3" mod and an LOD texture mod. The low-poly looks 99% as good as the stock grass, and it runs better. The Ancient Redwoods is probably the single most demanding area I've ever seen in Oblivion, mod or stock. This test run started at Gottlesfont Priory, and I ran due north until I hit the Black Road, which pretty much goes straight the heart of the Redwoods mod. I stuck the same path as much as humanly possible for these runs.

7950GX2: 2xAF, 1/3 grass, 1/4 item, 1/3 shadows, HDR, all else max
Frames: 1973 - Time: 73047ms - Avg: 27.010 - Min: 15 - Max: 44

8800GTX: Default AF, 1/3 grass, 1/4 item, 1/3 shadows, HDR, all else max
Frames: 3179 - Time: 71654ms - Avg: 44.366 - Min: 15 - Max: 63

8800GTX: 8x CSAA, 16xAF, HDR, EVERYTHING maxed!
Frames: 2178 - Time: 71695ms - Avg: 30.379 - Min: 18 - Max: 45

Uh, yeah, I'll play with everything cranked, thankyouverymuch!

And here's some screenshots of Oblivion with the maxed out settings. Also, you should note that I turned on all reflections in the .ini file, and each screenshot has the FPS stamp in the upper left. 56k download warning applies as well.

Redwoods
Crowd at War

The above screen is should be very representative of the worst case with the max settings, and the game was still playable.

Here's a nice HDR shot
Lighting on armor
The lake from the Ancient Yews mod
Swamps near Fort Teleman

And last but not least, Hardware p0rn (all in 7.2 megapixel glory!):
The Box
The Card
The Computer

And fuck, I'm out of webspace...
Artillery. Its what's for dinner.
User avatar
Uraniun235
Emperor's Hand
Posts: 13772
Joined: 2002-09-12 12:47am
Location: OREGON
Contact:

Post by Uraniun235 »

What's that card next to the vid card?
"There is no "taboo" on using nuclear weapons." -Julhelm
Image
What is Project Zohar?
"On a serious note (well not really) I did sometimes jump in and rate nBSG episodes a '5' before the episode even aired or I saw it." - RogueIce explaining that episode ratings on SDN tv show threads are bunk
Post Reply