It’s been 7 months since the launch of the first Fermi cards, and at long last we’re here: we’ve reached the end of the road on the Fermi launch. Today NVIDIA is launching the final GPU in the first-generation Fermi stack into the add-in card market, launching the GeForce GT 430 and the GF108 GPU that powers it. After months of launches and quite a bit of anticipation we have the complete picture of Fermi, from the massive GTX 480 to today’s tiny GT 430.

For the GT 430, NVIDIA is taking an interesting position. AMD and NVIDIA like to talk up their cheaper cards’ capabilities in HTPC environments but this is normally in the guise of an added feature. Rarely do we see a card launched on one or two features and today is one of those launches. NVIDIA believes that they’ve made the ultimate HTPC card, and that’s the line they’re going to be using to sell it; gamers need not apply. So just what is NVIDIA up to, and do they really have the new king of the HTPC cards? Let’s find out.

  GTX 480 GTX 460 768MB GTS 450 GT 430 GT 240 (DDR3)
Stream Processors 480 336 192 96 96
Texture Address / Filtering 60/60 56/56 32/32 16/16 16/16
ROPs 48 24 16 4 8
Core Clock 700MHz 675MHz 783MHz 700MHz 550MHz
Shader Clock 1401MHz 1350MHz 1566MHz 1400MHz 1340MHz
Memory Clock 924MHz (3696MHz data rate) GDDR5 900MHz (3.6GHz data rate) GDDR5 902MHz (3.608GHz data rate) GDDR5 900MHz (1800MHz data rate) DDR3 790MHz (1580MHz data rate) DDR3
Memory Bus Width 384-bit 192-bit 128-bit 128-bit 128-bit
Frame Buffer 1.5GB 768MB 1GB 1GB 1GB
FP64 1/8 FP32 1/12 FP32 1/12 FP32 1/12 FP32 N/A
Transistor Count 3B 1.95B 1.17B 585M 727M
Manufacturing Process TSMC 40nm TSMC 40nm TSMC 40nm TSMC 40nm TSMC 40nm
Price Point $449 $169 $129 $79 $75

The GT 430 is based on NVIDIA’s GF108 GPU, which like the GT21x GPUs before it, is coming to retail cards last rather than first. It’s already shipping in notebooks and prebuilt HTPCs, but this is the first time we’ve had a chance to look at just the complete card. And it really is a complete card – unlike all of NVIDIA’s other desktop launches which had GPUs with disabled functional units, the GT 430 uses a fully enabled GF108 GPU. For once with Fermi, we’ll be able to look at the complete capabilities of the GPU.

On the shader side of things, NVIDIA has taken GF106 and nearly cut it in half. We still have 1 GPC, but now it houses 2 SMs instead of 4. Each SM still contains 48 shaders, 8 texture units, and has FP64 capabilities, fulfilling NVIDIA’s commitment to FP64 capabilities (no matter how slow) on all Fermi GPUs. So yes Virginia, you can write and debug FP64 CUDA code on GF108. Attached to the shader block is 2 64bit memory controllers providing a 128bit memory bus, along with 128KB of L2 cache and a block of 4 ROPs.

For the memory NVIDIA is using DDR3, which is still common for cards under $100 given the price premium of GDDR5. Much like the GT 240 we believe this puts the GT 430 at a memory bandwidth disadvantage, and NVIDIA is already talking about working with partners on a GDDR5 version of the card in the future. We suspect that such a card will appear once 2Gbit GDDR5 is available in sufficient volume, as NVIDIA and their partners would seem to be fixated on having 1GB of RAM for now. In practice we usually find that 512MB of GDDR5 is better than 1GB of DDR3 in most cases.

Based on what we originally saw with GF104, we had expected GF108 to be a near-perfect quarter of the GF104 die: one-quarter the shaders, one-quarter the memory controllers, one quarter the ROPs. Even though GF108 has been available for some time now in mobile, OEM, and professional parts, we’ve never really taken a look at it beyond the fact that it had 96 shaders. If we had, we would have noticed something very important much sooner: it only has 4 ROPs.

For GF100-GF106, NVIDIA paired a block of 8 ROPS with a single 64bit memory controller. At the top this gave GF100 a 384bit memory bus, and down at GF106 it had a 192bit memory bus (with the GTS 450 shipping with 2 of those 3 64bit controllers active for a 128bit bus). For GF108 NVIDIA went with 2 64bit controllers to make a 128bit memory bus, which itself is not surprising since 64bit buses have extremely limited bandwidth, and that’s only suitable for bottom-tier ultra-cheap parts of which GF108 is not. So imagine our surprise when we were looking at the final spec sheet for GF108 and noticed that it didn’t have the 16 ROPs that logic dictates would be paired with a 128bit memory bus. And imagine our further surprise when that wasn’t even 8 ROPs, which is the size of a single block of ROPs or what GT214/GT216 had.

Instead NVIDIA’s thrown us another curveball here (they keep doing that) and gone with a new ROP configuration we haven’t seen before. For GF108 there are 4 ROPs in a single block, and that block is attached to the two 64bit memory controllers that make up the GPU’s 128bit memory bus. This is a surprising change since it would have required additional work on their part to make the necessary changes to their ROPs – sticking to the standard ratio would have been much easier. In any case it’s a notable change that makes GF108 something very different than we were initially expecting.

It’s the story of 4 ROPs that also sets up the positioning of the GT 430, the only retail card being released based on GF108 at the moment. With GT 430, NVIDIA has basically surrendered to AMD on performance. In a very unusual manner, you won’t find NVIDIA extoling the virtues of the card’s performance over AMD’s lineup. Even in our press briefing there was little said about gaming performance beyond the fact that it’s faster than the GT 220 and that NVIDIA believes it’s a meaningful upgrade over Intel’s IGP products due to their greater compatibility with games. Instead NVIDIA is largely selling this card upon its virtues as an HTPC card, or as NVIDIA likes to call the broader market segment: Digital Media PCs.

NVIDIA’s ace in the hole is that they have 1 thing right now that AMD doesn’t: a complete 3D stereoscopy strategy. On the hardware side this is due to the fact that GF104/106/108 all have support for HDMI 1.4a, which is necessary for full resolution 3D television/Blu-Ray and is an advantage afforded to them by the fact that AMD’s products are too old to incorporate support for HDMI 1.4a. On the other side NVIDIA has a coherent 3D strategy, with 3D Vision hardware for PC monitors, and thanks to the HDMI support for sending Blu-Ray 3D to TVs (and later this year, 3D gaming through 3D TV Play). And of course NVIDIA has bitstreaming audio capabilities for compressed lossless audio formats, bringing them up to par with AMD's audio offerings and a step up over the GT 200 series which could only support LPCM.

The long and the short of matters is that for gaming performance NVIDIA is already beat; as we’ll see even a Radeon HD 5570 DDR3 can confidently beat the GT 430, never mind the Radeon HD 5670 which can currently be found for even cheaper than the GT 430 once you factor in rebates. Even the GT 240, a card we panned hard last year is faster for the same price. With the hardware design NVIDIA has gone for, for the enthusiasts that are reading this article the GT 430 lives and dies by its HTPC capabilities; that’s NVIDIA’s position.

Moving on, since we’re looking at a low-end 40nm DirectX 11 part, this is a good time to talk about the “DX11 tax” – the fact that adding DX11 support to hardware takes up die space. We’ve already seen AMD go through this process with the 5000 series, where AMD had to spend most of their gains from the 40nm die shrink on transistors for DX11 features as opposed to transistors for overall performance. This is why cards like the Radeon HD 5450 end up being at best only as fast as the 4000 series cards they replace, and why the only major performance gains we saw when jumping generations were with GPUs that grew larger in size such as RV870. NVIDIA has to pay the same tax, and unlike AMD it’s even more obvious thanks to the fact that NVIDIA’s last-gen sub-$100 products were already 40nm.

At 585 million transistors, GF108 has outright half as many transistors as GF106, and compared to the GT21x series it has 100mil more transistors than GT216 (card: GT 220) and about 150mil fewer than GT 214 (card: GT 240). Lest we forget that Fermi was a major architecture revision for NVIDIA, they still had to pay the DX11 tax – if they were to add DX11 functionality to the GT21x series, all of their products would get larger and less profitable. How they chose to pay this tax was why we’re talking about GF108 only having 4 ROPs.

On the diagrams we use to showcase GPU architectures ROPs are drawn fairly small, but in reality they’re disproportionally drawn to make it easier to fit everything else in. Modern ROPs are actually pretty big, especially 16 of them as NVIDIA’s normal ratios would call for. They aren’t a massive fraction of the GPU, but they’re big enough that they can’t hide from GPU designers looking for ways to make an economical GPU. NVIDIA needs shader cores for gaming and compute alike, they need memory controllers and cache to feed those shaders, and they need fixed function units to glue the GPU together. ROPs are necessary as arbitrators to the memory controllers and for graphics rendering, but ultimately they’re one of the few graphics-only functional units next to the texture units.

And thus NVIDIA chose to pay the DX11 tax by cutting down on ROPs. This redesign no doubt added some extra work for them, but the end result was that it allowed them to get GF108 down to the die size and power consumption of the product we see today. NVIDIA’s rational for choosing the ROPs comes down to the belief that they’re one of the least important parts of a GPU for a value product, as features such as anti-aliasing are not usually used with these products, and as we’ve seen even additional ROPs won’t make these cards playable at higher resolutions such as 1680x1050. As a result the ROPs got the axe so that NVIDIA could bring a small die GPU with DX11, low power consumption, and better than GMA/5450 performance to market.

Funny enough NVIDIA won’t give us the actual size of GF108 (they haven’t done this for any Fermi parts), but using a ruler we’d estimate the size to be 11mm x 10.5mm, or roughly 116mm2. As with transistor count this puts it between GT214 and GT216 (being closer to GT216), or roughly the same size as AMD’s Redwood GPU in the 5500/5600 families. The card’s TDP is rated at 49W, which should actually be lower than the GT216-based GT 220, even though GT216 was a smaller GPU using fewer transistors.

As for today’s launch, as is common with low-end products NVIDIA is leaving the launch in the hands of its partners. NVIDIA did produce a reference design for internal use, but it’s not being used by any shipping cards, nor were reviewers given the reference design. Instead all the cards you’ll see today, including the Asus card we’re reviewing today, are vendor custom cards.

We’ve attached a small gallery of card designs that will be shipping. There’s a complete range of cards, including full and half-height cards, and of course the obligatory passively cooled cards. There’s even a half-height passive card from Sparkle listed if you have that specific itch to scratch.

NVIDIA put the MSRP at $80, and while there aren’t any reference designs to tie this to we expect the first wave of cards to stick to this price point. At $80 the GT 430 goes directly up against the Radeon HD 5570 (DDR3 & GDDR5), sale-priced Radeon 5670s, the GT 240, and even the venerable 9800GT. All of these cards are faster in gaming performance, which is why NVIDIA’s strengths need to lie in HTPC performance and overall power consumption.

October 2010 Video Card MSRPs
  $700 Radeon HD 5970
$360 Radeon HD 5870
$270 Radeon HD 5850
$170 Radeon HD 5830
$130 Radeon HD 5770
GeForce GT 430
$80 Radeon HD 5670/5570
Meet the Asus ENGT430
Comments Locked


View All Comments

  • Mumrik - Monday, October 11, 2010 - link

    From the Final Words:
    "We opened up this article talking about how NVIDIA is FORGING performance in the name of (...)"

    Surely you don't mean "forging"...
  • knutjb - Monday, October 11, 2010 - link

    I can see this and similar under performing cards showing up in consumer HTPCs further turning the general public off computer based products giving yet another reprieve to blue ray set top boxes.

    It seems every time the PC comes closer to taking a key role in the entertainment center a company releases a pos product like this.
  • manno - Monday, October 11, 2010 - link

    I can not agree with your conclusions and I'm an ATI/AMD fan. I will not touch fermi with a 10' pole, but to say from your benchmarks that the 430 is not competitive with the 5570 is just plain wrong. Let's do the numbers

    Benchmark : AMD 5570 | Nvidia 430
    Image Quality : +1 | 0
    Crysis : +1 | 0
    BattleForge : +1 | 0
    HAWX : 0 | +1
    Left 4 Dead : 0 | +1
    Battlefield: BC2 : +1 | 0
    STALKER : 0 | +1
    DiRT 2 : +1 | 0
    Mass Effect 2 : 0 | +1
    Wolfenstein : +1 | 0

    On power consumption: I personally feel that a lower idle power consumption is more important than load as my system would sit idle way more frequently that it would be under load, but that depends on the user: Draw or +1 Nvidia

    Noise : 0 | +1

    AMD: 5 | Nvidia: 5

    Looks dead even to me both cards would make great HPC cards. I find a lot of image quality dbenchmarks are highly subjective and fail under a double blind test. It's like asking an audiophile for advice on weather you should use MP3's or FLAC's very few people can tell the difference between the two and unless you're highly tuned to it it's not noticable. Not to mention I haven't even gotten to the fact that it's an AMD made denchmark, I wouldn't put any weight in an Nvidia benchmark for an ATI card. I've made 0 purchases based on 3D Mark Vantage scores that handicap Nvidia a rediculous amount due to the GPU compute portion of the benchmark, otherwise I'd own no AMD cards.


    Full disclosure:
    10% of my portfolio is AMD
    0.0% of my portfolio is Nvidia
  • cknobman - Monday, October 11, 2010 - link

    Please read the article correctly. It clearly states that it is on par with 5570 but fails every time against the 5670.

    Given the pricing of this card it is not out of the question to compare it against the 5670 in gaming tests.

    This card at its current price point is a huge failure for Nvidia.

    The companies banking on 3D tech are going to find hard time in the near future because consumers arent buying in on it. Heck Ive already read about GIGANTIC price cuts in the LCD market coming in the next month because supply is through the roof and no one is buying.
  • manno - Monday, October 11, 2010 - link

    looks like their priced evenly...
  • manno - Monday, October 11, 2010 - link

    "GT 430 simply isn’t competitive with AMD’s 5570 and 5670 in gaming performance"

    Does not read as "on par"
  • Stuka87 - Monday, October 11, 2010 - link

    Well, that would be true if those points were all equal. However they are not. Especially in regards to image quality where the nVidia lost horribly in a test for the market that this card is specifically aimed at.

    As for gaming, you can pick up 5670's for the same price as the 430, and the 5570 for less. So the 5670 is really the card that nVidia is up against. Which easily beat it in every single performance test.

    The lower idle power/temps of the 430 are nice to have, but not if it means significantly worse performance in other areas.
  • manno - Tuesday, October 12, 2010 - link

    You're correct, and I want to nake this clear, my issue isn't with your post it's with this articles conclusion which states:

    "GT 430 simply isn’t competitive with AMD’s 5570 and 5670 in gaming performance"

    in which it states that Nvidia is in an unenviable position HTPC wise because the 5570 is superior to the 430. Which is not true based on their own benchmarks. Of course the 5670 is better but that's not the card this Nvidia is positioning against it that will be the GT 440 or maybe 435's job.

    The 5770 and 430 are the exact same card benchmark and price wise:

    As for AMD's Image Quality test, that is not the best test/benchmark they had at their disposal. A double blind comparison where you play two clips one after the other to random people and give them an similar, better, worse question is the best test for something as subjective as image quality. Using an AMD test to judge any card is inherently biased in AMD's favor. Their conclusion is wrong.
  • esc923 - Monday, October 11, 2010 - link

    Lol. You seem to think highly of yourselves asking others to defend their position when you can't even defend your own.

    You've thrown a lot of opinions (e.g. "I personally feel..." or "I find..."). All I can say is let's ignore your personal feelings as they're irelevant, ok?

    That just leaves your basic reasoning, and you've failed even more miserably here. Weighing the performance of something like "Battlefield 2" equally to that of video "image quality" to get your conclusion (i.e. "looks dead even to me") is just plain idiotic considering we're talking about the HTPC market where image quality is more imporant than video game frame rates. Which is pretty frigging obvious.... Note that I'm not calling you idiotic, just your reasoning.

    You see a lot of crap on the internet that you ignore, but the thing is, your memo was written in such a serious way that even includes a disclaimer as well as a DEMAND in caps that you've just got to laugh because the end result comes across as someone trying too hard to look smart, and failing.
  • manno - Monday, October 11, 2010 - link

    From my previous post:

    The 5570 and the 430 are priced exactly the same their performance for all intents and purposes is exactly the same. ATI and Nvidia are both taking the same position in this category, why is ATI's position superior? Because an Nvidia card doesn't perform as well as an ATI card on an ATI created benchmark. That is a poor reason to state one card is superior to another.

    As I stated previously a metric like image quality is purely subjective unlike a FPS score which is objective. What looks like garbage to one person looks great to another.

    Regarding my intelligence, my argument is correct regardless of how smart I am. Furthermore I could care less how smart anyone here thinks I am, for all I care your can hold a mental image of me squatting in mud smiling as I shove berries up my nose.

Log in

Don't have an account? Sign up now