GF100 will be NVIDIA's hottest GPU ever, with TDP in excess of 200W (RV870's TDP is 188W)...
http://www.fudzilla.com/content/view/17356/1/
http://www.fudzilla.com/content/view/17356/1/
Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature may not be available in some browsers.
Lol, no, not at all. Either some of you are noobs and only saw this less than stellar release or are very forgetful. For instance, the FX5800 was staggeringly worse. The HD 2900 XT wasn't a bad card, it just wasn't great. It couldn't compete with the 8800GTX and Ultra in high resolution gaming, but, for instance, it was the king of benchmarking. It was a new architecture and AMD was testing it's strengths and weaknesses.Wasn't the 2900XT TDP 215W? It was one of the greatest failures ever made.
All in all, we believe that end users won’t mind, as long as it stays faster than ATI’s RV870-powered HD 5870 card.
the FX5800 was staggeringly worse
What it all boils down to in the end.
not true...if Fermi is only slightly faster then a 5870 then it will matter as people will stick with ATI...if it's a lot faster then people won't care as much...heat/power consumption are big issues for a lot of people and ATI is gonna hammer this point in the next few months
You would think they would, but AMD's marketing has been "meh" at best. Here's a golden opportunity for them. Personally, considering my setup, Fermi's power consumption/heat output is enough for me to pass it by.not true...if Fermi is only slightly faster then a 5870 then it will matter as people will stick with ATI...if it's a lot faster then people won't care as much...heat/power consumption are big issues for a lot of people and ATI is gonna hammer this point in the next few months
IMO if your dealing with enthusiast level cards then you should have already accounted for power / heat and made the proper arrangements.
The price vs heat vs power etc. comparison will be true to some on here but most of the time people just want the biggest stick.
Oh how wrong you are. The way most people get started down the road of enthusiast level hardware is by upgrading their store bought brand name pre-assembled PC's. Ours is a very small niche compared to the mainstream market that brings the actual bulk of earnings for add-in board vendors like Nvidia and AMD(ATI).
We can only hope NVIDIA has not.I've completely blocked the FX series out of my memories...
Why are you arguing over the piece of unreleased hardware?![]()
The HD 2900 XT wasn't a bad card, it just wasn't great. It couldn't compete with the 8800GTX and Ultra in high resolution gaming, but, for instance, it was the king of benchmarking. It was a new architecture and AMD was testing it's strengths and weaknesses.
How are we supposed to overclock this thing, much less cool it down when overclocked
I mean if TDP is that sky high, tri slot coolers![]()
And they will be. And if it so happens that it doesn't compare so well to the 5870 it doesn't mean it is "one of the greatest failures ever made" either. There are plenty of other examples of that, the phrase was used incorrectly there.Hahaha... come on guy. By that rationale, nVidia is just testing Fermi's strengths and weaknesses.
And they will be. And if it so happens that it doesn't compare so well to the 5870 it doesn't mean it is "one of the greatest failures ever made" either. There are plenty of other examples of that, the phrase was used incorrectly there.
My 4870x2's each have a TDP of about 270w and I'm running them both using the stock dual-slot coolers and overclocking on them. 200w TDP Fermi should not be that big of a deal really.
Nice overclock on your Q9650 btw. Is that 24/7 and what OS are you using?
Hopefully GT100 doesn't really need any/much overclocking, though I'll admit I'm not really that optimistic at this point4870x2s are crap oc'ers generally.
Actually, NVIDIA does really well with respect to performance/watt (or at least they have historically). There are the obvious exceptions like the FX series, but for the most part, they've done pretty well in this area.NVIDIA seems to lack efficiency in their designs or something.
the Fermi release in March is not what Nvidia wants to release...it's what they feel they have to put out to compete with ATI...all the early adopters are gonna get burned (literally) when Nvidia releases their Fermi refresh 6 months down the road
Sure are a lot of opinions and statments on things we have no real clue about.
Water cooled editions. I get the feeling we are going to see a lot more of them than we normally do. And I don't really consider that a bad thing. I would love for there to be more factory installed water blocks on vid cards. It will bring competition to a market that sorely needs it. (A $400 vid card with a $50 water block tossed on it does not make a $600 product imho.)
My 4870x2's each have a TDP of about 270w and I'm running them both using the stock dual-slot coolers and overclocking on them. 200w TDP Fermi should not be that big of a deal really.
Water cooled editions. I get the feeling we are going to see a lot more of them than we normally do. And I don't really consider that a bad thing. I would love for there to be more factory installed water blocks on vid cards. It will bring competition to a market that sorely needs it. (A $400 vid card with a $50 water block tossed on it does not make a $600 product imho.)
Not trying to flame, but are you really comparing a dual gpu card to a single gpu card?
.heat/power consumption are big issues for a lot of people and ATI is gonna hammer this point in the next few months
And you think the two are separate? lol.Heat ill give someone, but power consumption?...pfft buzzwords for forum trolls. Nobody buys a high end card based on power consumption.
All in all, we believe that end users won’t mind, as long as it stays faster than ATI’s RV870-powered HD 5870 card.