NVIDIA's GeForce GT 430: The Next HTPC King?by Ryan Smith & Ganesh T S on October 11, 2010 9:00 AM EST
It’s been 7 months since the launch of the first Fermi cards, and at long last we’re here: we’ve reached the end of the road on the Fermi launch. Today NVIDIA is launching the final GPU in the first-generation Fermi stack into the add-in card market, launching the GeForce GT 430 and the GF108 GPU that powers it. After months of launches and quite a bit of anticipation we have the complete picture of Fermi, from the massive GTX 480 to today’s tiny GT 430.
For the GT 430, NVIDIA is taking an interesting position. AMD and NVIDIA like to talk up their cheaper cards’ capabilities in HTPC environments but this is normally in the guise of an added feature. Rarely do we see a card launched on one or two features and today is one of those launches. NVIDIA believes that they’ve made the ultimate HTPC card, and that’s the line they’re going to be using to sell it; gamers need not apply. So just what is NVIDIA up to, and do they really have the new king of the HTPC cards? Let’s find out.
|GTX 480||GTX 460 768MB||GTS 450||GT 430||GT 240 (DDR3)|
|Texture Address / Filtering||60/60||56/56||32/32||16/16||16/16|
|Memory Clock||924MHz (3696MHz data rate) GDDR5||900MHz (3.6GHz data rate) GDDR5||902MHz (3.608GHz data rate) GDDR5||900MHz (1800MHz data rate) DDR3||790MHz (1580MHz data rate) DDR3|
|Memory Bus Width||384-bit||192-bit||128-bit||128-bit||128-bit|
|FP64||1/8 FP32||1/12 FP32||1/12 FP32||1/12 FP32||N/A|
|Manufacturing Process||TSMC 40nm||TSMC 40nm||TSMC 40nm||TSMC 40nm||TSMC 40nm|
The GT 430 is based on NVIDIA’s GF108 GPU, which like the GT21x GPUs before it, is coming to retail cards last rather than first. It’s already shipping in notebooks and prebuilt HTPCs, but this is the first time we’ve had a chance to look at just the complete card. And it really is a complete card – unlike all of NVIDIA’s other desktop launches which had GPUs with disabled functional units, the GT 430 uses a fully enabled GF108 GPU. For once with Fermi, we’ll be able to look at the complete capabilities of the GPU.
On the shader side of things, NVIDIA has taken GF106 and nearly cut it in half. We still have 1 GPC, but now it houses 2 SMs instead of 4. Each SM still contains 48 shaders, 8 texture units, and has FP64 capabilities, fulfilling NVIDIA’s commitment to FP64 capabilities (no matter how slow) on all Fermi GPUs. So yes Virginia, you can write and debug FP64 CUDA code on GF108. Attached to the shader block is 2 64bit memory controllers providing a 128bit memory bus, along with 128KB of L2 cache and a block of 4 ROPs.
For the memory NVIDIA is using DDR3, which is still common for cards under $100 given the price premium of GDDR5. Much like the GT 240 we believe this puts the GT 430 at a memory bandwidth disadvantage, and NVIDIA is already talking about working with partners on a GDDR5 version of the card in the future. We suspect that such a card will appear once 2Gbit GDDR5 is available in sufficient volume, as NVIDIA and their partners would seem to be fixated on having 1GB of RAM for now. In practice we usually find that 512MB of GDDR5 is better than 1GB of DDR3 in most cases.
Based on what we originally saw with GF104, we had expected GF108 to be a near-perfect quarter of the GF104 die: one-quarter the shaders, one-quarter the memory controllers, one quarter the ROPs. Even though GF108 has been available for some time now in mobile, OEM, and professional parts, we’ve never really taken a look at it beyond the fact that it had 96 shaders. If we had, we would have noticed something very important much sooner: it only has 4 ROPs.
For GF100-GF106, NVIDIA paired a block of 8 ROPS with a single 64bit memory controller. At the top this gave GF100 a 384bit memory bus, and down at GF106 it had a 192bit memory bus (with the GTS 450 shipping with 2 of those 3 64bit controllers active for a 128bit bus). For GF108 NVIDIA went with 2 64bit controllers to make a 128bit memory bus, which itself is not surprising since 64bit buses have extremely limited bandwidth, and that’s only suitable for bottom-tier ultra-cheap parts of which GF108 is not. So imagine our surprise when we were looking at the final spec sheet for GF108 and noticed that it didn’t have the 16 ROPs that logic dictates would be paired with a 128bit memory bus. And imagine our further surprise when that wasn’t even 8 ROPs, which is the size of a single block of ROPs or what GT214/GT216 had.
Instead NVIDIA’s thrown us another curveball here (they keep doing that) and gone with a new ROP configuration we haven’t seen before. For GF108 there are 4 ROPs in a single block, and that block is attached to the two 64bit memory controllers that make up the GPU’s 128bit memory bus. This is a surprising change since it would have required additional work on their part to make the necessary changes to their ROPs – sticking to the standard ratio would have been much easier. In any case it’s a notable change that makes GF108 something very different than we were initially expecting.
It’s the story of 4 ROPs that also sets up the positioning of the GT 430, the only retail card being released based on GF108 at the moment. With GT 430, NVIDIA has basically surrendered to AMD on performance. In a very unusual manner, you won’t find NVIDIA extoling the virtues of the card’s performance over AMD’s lineup. Even in our press briefing there was little said about gaming performance beyond the fact that it’s faster than the GT 220 and that NVIDIA believes it’s a meaningful upgrade over Intel’s IGP products due to their greater compatibility with games. Instead NVIDIA is largely selling this card upon its virtues as an HTPC card, or as NVIDIA likes to call the broader market segment: Digital Media PCs.
NVIDIA’s ace in the hole is that they have 1 thing right now that AMD doesn’t: a complete 3D stereoscopy strategy. On the hardware side this is due to the fact that GF104/106/108 all have support for HDMI 1.4a, which is necessary for full resolution 3D television/Blu-Ray and is an advantage afforded to them by the fact that AMD’s products are too old to incorporate support for HDMI 1.4a. On the other side NVIDIA has a coherent 3D strategy, with 3D Vision hardware for PC monitors, and thanks to the HDMI support for sending Blu-Ray 3D to TVs (and later this year, 3D gaming through 3D TV Play). And of course NVIDIA has bitstreaming audio capabilities for compressed lossless audio formats, bringing them up to par with AMD's audio offerings and a step up over the GT 200 series which could only support LPCM.
The long and the short of matters is that for gaming performance NVIDIA is already beat; as we’ll see even a Radeon HD 5570 DDR3 can confidently beat the GT 430, never mind the Radeon HD 5670 which can currently be found for even cheaper than the GT 430 once you factor in rebates. Even the GT 240, a card we panned hard last year is faster for the same price. With the hardware design NVIDIA has gone for, for the enthusiasts that are reading this article the GT 430 lives and dies by its HTPC capabilities; that’s NVIDIA’s position.
Moving on, since we’re looking at a low-end 40nm DirectX 11 part, this is a good time to talk about the “DX11 tax” – the fact that adding DX11 support to hardware takes up die space. We’ve already seen AMD go through this process with the 5000 series, where AMD had to spend most of their gains from the 40nm die shrink on transistors for DX11 features as opposed to transistors for overall performance. This is why cards like the Radeon HD 5450 end up being at best only as fast as the 4000 series cards they replace, and why the only major performance gains we saw when jumping generations were with GPUs that grew larger in size such as RV870. NVIDIA has to pay the same tax, and unlike AMD it’s even more obvious thanks to the fact that NVIDIA’s last-gen sub-$100 products were already 40nm.
At 585 million transistors, GF108 has outright half as many transistors as GF106, and compared to the GT21x series it has 100mil more transistors than GT216 (card: GT 220) and about 150mil fewer than GT 214 (card: GT 240). Lest we forget that Fermi was a major architecture revision for NVIDIA, they still had to pay the DX11 tax – if they were to add DX11 functionality to the GT21x series, all of their products would get larger and less profitable. How they chose to pay this tax was why we’re talking about GF108 only having 4 ROPs.
On the diagrams we use to showcase GPU architectures ROPs are drawn fairly small, but in reality they’re disproportionally drawn to make it easier to fit everything else in. Modern ROPs are actually pretty big, especially 16 of them as NVIDIA’s normal ratios would call for. They aren’t a massive fraction of the GPU, but they’re big enough that they can’t hide from GPU designers looking for ways to make an economical GPU. NVIDIA needs shader cores for gaming and compute alike, they need memory controllers and cache to feed those shaders, and they need fixed function units to glue the GPU together. ROPs are necessary as arbitrators to the memory controllers and for graphics rendering, but ultimately they’re one of the few graphics-only functional units next to the texture units.
And thus NVIDIA chose to pay the DX11 tax by cutting down on ROPs. This redesign no doubt added some extra work for them, but the end result was that it allowed them to get GF108 down to the die size and power consumption of the product we see today. NVIDIA’s rational for choosing the ROPs comes down to the belief that they’re one of the least important parts of a GPU for a value product, as features such as anti-aliasing are not usually used with these products, and as we’ve seen even additional ROPs won’t make these cards playable at higher resolutions such as 1680x1050. As a result the ROPs got the axe so that NVIDIA could bring a small die GPU with DX11, low power consumption, and better than GMA/5450 performance to market.
Funny enough NVIDIA won’t give us the actual size of GF108 (they haven’t done this for any Fermi parts), but using a ruler we’d estimate the size to be 11mm x 10.5mm, or roughly 116mm2. As with transistor count this puts it between GT214 and GT216 (being closer to GT216), or roughly the same size as AMD’s Redwood GPU in the 5500/5600 families. The card’s TDP is rated at 49W, which should actually be lower than the GT216-based GT 220, even though GT216 was a smaller GPU using fewer transistors.
As for today’s launch, as is common with low-end products NVIDIA is leaving the launch in the hands of its partners. NVIDIA did produce a reference design for internal use, but it’s not being used by any shipping cards, nor were reviewers given the reference design. Instead all the cards you’ll see today, including the Asus card we’re reviewing today, are vendor custom cards.
We’ve attached a small gallery of card designs that will be shipping. There’s a complete range of cards, including full and half-height cards, and of course the obligatory passively cooled cards. There’s even a half-height passive card from Sparkle listed if you have that specific itch to scratch.
NVIDIA put the MSRP at $80, and while there aren’t any reference designs to tie this to we expect the first wave of cards to stick to this price point. At $80 the GT 430 goes directly up against the Radeon HD 5570 (DDR3 & GDDR5), sale-priced Radeon 5670s, the GT 240, and even the venerable 9800GT. All of these cards are faster in gaming performance, which is why NVIDIA’s strengths need to lie in HTPC performance and overall power consumption.
|October 2010 Video Card MSRPs|
|$700||Radeon HD 5970|
||$360||Radeon HD 5870|
||$270||Radeon HD 5850|
|$170||Radeon HD 5830|
|$130||Radeon HD 5770|
GeForce GT 430
|$80||Radeon HD 5670/5570|
Post Your CommentPlease log in or sign up to comment.
View All Comments
heflys - Monday, October 11, 2010 - linkSeriously?
Belard - Tuesday, October 12, 2010 - linkOverall, this card isn't impressive at all... the PRO's are there, and AMD does need 3D and physics abilities.
But at $80, it goes against the 5650 cards and easily loses.
About HDMI 1.4b... it doesn't really matter. HDMI is dead... faster than it should be, but there is no future in it. CAT6-A/V will start replacing HDMI in 2011.... all the big TV players are on board - they don't have to pay licensing fees or use special expensive connectors or cabling of HDMI.
And HTPC's will not get very popular until the Cable companies loosen up about people access channels like HBO, SHO, etc. Windows7 Media player is nice, but the interface is still rather weak for power users compare to some of the others out there. For example, the program grid is HORRIBLE... when others allow 2~4hours of blocks and around 20 channels at a time... none of this 1.5hr / 6 channel junk. Oh, and the DRM of Media player makes archiving your shows near impossible. Like if you have to reinstall the OS or do a system upgrade.
heflys - Tuesday, October 12, 2010 - linkAccording to most review sites, things like PhysX and 3d vision are nothing but gimmicks that contribute little to actual performance. Instead, most view them as pointless system hogs.
Belard - Tuesday, October 12, 2010 - linker... PhysX and 3D has never been about improving performance. It was about adding to the visual experience. Like Avatar looks great in 2D and 3D... but 3D sucks you in a bit more.
Games like Mirror's Edge come more realistic with PhysX, even thou it doesn't improve game play one bit.
Those technologies are new, and until PhysX becomes shared/standard on all video cards - it will be more gimmick then a standard. But who knows...
Hmmm... back around 1988 when computers were 8~16mhz, only Mac and Amigas pretty much had a native GUI OS, MS was horrible MS-DOS with 8.3 file names, no multi-tasking, horrible graphics and forget about sound. Someone from the DOS camp said "Who needs graphics and sound, those are for toys. PC are REAL Computers".
Uh huh. And now we have 1000Mhz cell phones with 16GB of RAM.
The 1986 vintage Amiga had Graphics, sound and Multi-tasking... was it a gimmick?
heflys - Tuesday, October 12, 2010 - link"Performance" was a typo on my part, since I clearly indicated that it was a system hogs. Physx, in most cases-as displayed titles such as Mafia II- contribute little to nothing (in some games) towards graphics. Most players won't even notice such things as enhanced physics or improved decals. In fact, the most noticeable thing displayed in Mafia II was the presence of debris. Players will, however, notice the impressive amount of lag brought on by such features.
3d Vision, as displayed in one review, rendered the GPU (a gtx 460 1gb) to unplayable frame-rates. It essentially required the player go to SLI. Which brings me to another point.....Why are you bringing up Physx or 3d vision in regards to this product? You seriously think this cheap HTPC card could handle any of the above features, particularly when a 1gb 460 struggles to?
And are we seriously comparing the Amiga to such an insignificant thing as cheesy video game effects? You can't be serious. Particularly when there are other physics engines (Havok being one of the most prominent) doing some of the same things.
However, please tell me how Physx made Mirror's Edge a more realistic experience. Particularly since that game, like Mafia II, only added physics to debris.
Belard - Tuesday, October 12, 2010 - linkI agree with you on the first paragraph. We want constant visual abilities, but without the cost of general performance.
This was one of the arguments of 3Dfx's Voodoo3 vs TNT cards -
Performance with 16bit graphics vs nVidia's 24bit.
When I played the JSF game around 1999-2000, the 16bit limitation was noticed BIG time on my Voodoo1, but the frame rate murdered the ATI I had. It was a trade off. This is always a constant battle with out GPUs... remember when AA was added? Even today, AA effects the performance of every single video card - but unlike 8 years ago, it no longer renders most cards useless.
Yeah, 3d Vision & PhysX is useless on the GF430... pretty much like the ATI Eyefinity's tech doesn't belong on every ATI card (reduce the cost by $10, improve airflow) - especially for the low-end, but its very handy for business users.
You said: "And are we seriously comparing the Amiga to such an insignificant thing as cheesy video game effects?"
Yes, in that PhysX and 3D tech is still baby tech. In a few years, we'll be start seeing 3D TV's that don't require glasses. PhysX or Havok or other becomes more standard - or perhaps MS adds it to DX12. It's going to be years before we see results of the latest technology. Just like the PC folk's of the 80's who said the Amiga was a toy and computers didn't need graphics and sound. And yes, my Amigas still work.
"please tell me how Physx made Mirror's Edge a more realistic experience." Look up the various side-by side videos. It adds cloth effects, broken glass and yes, debris. A side by same example: http://www.youtube.com/watch?v=w0xRJt8rcmY and check out batman too.
Of course, that didn't help to actually POPULATE the city of Mirror's edge with people... funny, a huge modern city with only a few people and police, with all that construction - where are the workers? Another example. A burger that is just meat and bread is bland... but add some tomatoes, lettuce, cheese and it becomes a better meal.
heflys - Tuesday, October 12, 2010 - linkThanks for the civil discussion....I half expected you to call me an idiot for some reason......Don't know why.....
I think ATI's just going to bide its time with the 3d/Physics display, since at this point, they don't really need to invest in that platform. Maybe in the future.
Belard - Wednesday, October 13, 2010 - linkWould it make you feel better if I did? :)
I've been into computers for a long long time - and I do my best to NOT be a fanboy. Give credit where credit is due... Apple, Intel, MS, AMD, Nvidia, Opera, FireFox, etc.
What gives me/us the best deal at the time of purchase.
In 2015, our graphics on consoles (don't know about computers) will make todays GPUs look like GeForce 5900/ATI 9700 in terms of performance and abilities.
We'll see. Perhaps Archive this page?
drjonz - Tuesday, October 12, 2010 - linkWhy no comparison to integrated Intel Clarksdale? Many of us with HTPC went with that since we're not gamers. I've been really happy with it. Maybe once per Blu-Ray watching, I'll get a stutter. Not sure if it's because I'm underpowered or what. Would be cool to see what more I'd get for $100.
ganeshts - Tuesday, October 12, 2010 - linkWe have mentioned the HQV score for Clarkdale (Intel HD Graphics) as 133, much lower than 5570 and slightly lower than the 430.
Please take a look at the Core 100 review we carried a few months back. It reviewed the Arrandale platform for HTPCs and it is quite good for casual HTPC users.