Nvidia GeForce GTX 580 Reviewed in 2018, 250 watts of Fury!
Nvidia GeForce GTX 580 Reviewed in 2018, 250 watts of Fury!
2018-01-25
welcome back to our unboxed for today's
video we're turning the clock back way
back to a time when GPUs were built
using a 40 nanometer process and
featured gddr5 memory just 1.5 gigabytes
of gddr5 memory yeah anyway we're
jumping back to 2010 to revisit the
once-mighty GeForce GTX 580 now this
isn't going to be a quick blast from the
past type video where I compare the GTX
580 to some modern GPUs and a few games
laughs but it's outdated design and then
poke you in the eye before fading out to
20 seconds of cool barrel I'll still do
those last two things but before we do I
want to touch on some history show some
benchmark graphs of course and then play
a heap of games so strap yourselves in
as we go for a gentle stroll down memory
lane the GTX 580 is a venture began on
November 9th 2010 that was the first of
the GeForce 500 series to be released
and for a good reason
early that year in march and video
released their fermi microarchitecture
with the gtx 470 and 480 GPUs codenamed
GF 100 this marked a transition from 55
nanometers down to 40 nanometers for
Nvidia's high-end complex GPUs they did
start playing around the 40 nanometer
process back in late 2009 with the
entry-level price such as the geforce g2
220 230 and 240 for example then four
months later the mighty GTX 480 arrived
in gamers got their chance to buy a $500
u.s. graphics card that doubled as a
George Foreman grill with a TDP of 240
watts the GTX 480 didn't raise too many
eyebrows AMD's much smaller Radeon HD
5870 for example that was rated for 228
watts while their Joule GPU monster the
5970 smoked most power supplies with its
294 watt rating although both the Radeon
58 70 and GeForce GTX 480 were built
upon the same 40 nanometer manufacturing
process the 480 was massive the DOE is
over 50% larger Nvidia was striving to
create a GPU that packed both strong
gaming and strong computer
and that resulted in the GF 100-day
measuring and insane 529 millimeters
squared this behemoth of a die caused
problems 40 SMC's 40 nanometer process
which was still maturing at the time and
this resulted in poor yields and leaky
transistors for gamers the GTX 480 was
simply too hot to handle
literally you couldn't touch it even
after 10 minutes of shutting the PC down
that thing was still searing hot and
this also meant that it made more noise
than an old vacuum cleaner the fermi
based GF 100 architecture went down as
one of Nvidia's worst creations in
history and even at the time they knew
it was a real stinker so they quickly
got back to work and later that same
year they did relaunch Fermi with the
GTX 580 by November tsmc had ironed out
the kinks for their 40 nanometer process
and we got 7% more cos in a slightly
smaller die of course it wasn't just
tsmc that had time to work things out
and video 2 went back to the drawing
board and fixed what was wrong with the
GF 100 is on when they created the new
GF 110 otherwise known as the Fermi
refresh based on my own testing done in
2010 I noted that power consumption was
reduced by at least 5 percent and the
low temperatures for the reference
design card was 15 percent lower making
those results even more impressive was
the fact that across the dozen or so
games tested the 580 was on average 25
percent faster than the 480 that's a
significant performance uplift within
the same year particularly given that
they are both produced on the same
manufacturing process and the 580 as I
said also saved power so yeah it wasn't
really until the 580 came along that we
realised just how much the 480 sucked I
didn't like the 480 when it was first
released but yeah really wasn't until
the 580 came along eight months later
that I understood just how rubbish it
really was I concluded my 580 coverage
by saying the following the GeForce GTX
580 is unquestionably the fastest single
GPU graphics card available today
representing a better value than any
other high-end video card however it's
not the only option for those looking to
spend $500 dual Radeon HD 6870 s remain
very attractive it's just under $500
deliver more performance than a single
GeForce GTX 580 in most titles
however multi GPU technology isn't
without its pitfalls and given the
choice we will always opt for a single
high-end graphics card in videos GeForce
GTX 580 may be the current king of the
hill but this could all change next
month when AMD launches their new Radeon
HD 6900 series aimed II was originally
expected to deliver its came an XT and
Pro based Radeon HD 6970 and 6950
graphics cards sometime during November
but they have postponed their arrival
until mid-december for undisclosed
reasons if you don't mind holding off a
few weeks the weight could be worth some
savings or potentially more performance
for the same dollars depending on what
AMD has reserved for us now for those of
you wondering the Radeon HD 6970 was a
bit of a disappointment it fell short of
the GTX 518 as a result was priced to
compete with the GTX 570 this allowed
him video to hold the performance crown
without dispute for the next few years
it could be interesting to see how those
two stack up today perhaps that's
something we can look into in a future
video so we've established that the
GeForce GTX 580 was able to save face
from video in 2010 and retain the
performance crown for a few more years
before the 28 nanometer process arrived
in 2012 reigniting the GPU wore once
more what I want to know is how the
tubby 520 millimeters squared die gets
on today can the GTX 580 nits 512 CUDA
cores locked at 772 megahertz handle
today's games sure it does have gddr5
memory on a rather wide 384 bit wide bus
pumping out an impressive 192 gigabytes
per second but there's only one point 5
gigabytes of it at least for most models
there were three gigabyte variants
floating around but most people picked
up the original 1.5 gigabyte version
because that made the most sense at the
time to see how it handles I'm going to
compare it with the Radeon HD 7950 along
with the more recently released geforce
gtx 1050 1050 Ti and Radeon rx 560
so let's check out those results before
taking a gander at some actual gameplay
first up we have the battlefield 1
results using the ultra quality preset
which
as it turns out is a rather bad choice
for the GTX 580 and its measly 1.5
gigabyte frame buffer
don't worry though after we go over all
the graphs I will show you some gameplay
using some more agreeable settings for
now though we can see that here for 580
simply isn't cutting it and fails to
deliver what I would call playable
performance as a result it was a little
over 30% slower than the Radeon HD 7950
next up we have the dawn of war 3
results and this game was tested using
the more appropriate medium quality
preset here we see a reasonable 41 FPS
on average began the GTX 580 is haunted
by that limited vram buffer as the
minimum frame rate dropped down to just
23 fps and this meant that stuttering
was a bit of an issue dirt 4 was also
tested using the medium quality settings
and here the experience was quite good
with the GTX 580 certainly playable and
for the most part was very smooth at
1080p moving on we offer honor and here
we found a mostly playable experience if
that makes sense
at times we did drop down below 30 fps
and this did make the gameplay
noticeably more choppy than it was say
on the RX 560 or HD 7950 it's also
interesting to note that the GTX 1050
absolutely obliterates 2010 s flagship
part Ghost Recon wildlands isn't the
most optimized game to be released in
2017 that's for sure putting it mildly
let's say here the GTX 580 really
struggled to provide playable
performance using the lowest possible
quality settings at 1080p and there were
a number of graphical glitches as well
like dirt 4 we see decent results when
testing with Mass Effect Andromeda here
the GTX 580 Mansion average of 49 FPS at
1080p using the medium quality preset
the minimum or one percent low result of
44 FPS also meant that the gameplay was
quite smooth performance in prey wasn't
bad either though the texture quality
was horrible as most of the textures
simply didn't load so there was lots of
blurry looking surfaces the minimum
frame rate suffered a little bit because
of this but overall the gameplay itself
wasn't that bad it just looked quite
poor visually the average in Resident
Evil 7 was also quite good using the
medium quality settings they'll again
that limited frame buffer does kill
performance at times and we did see dips
down to 45 FPS which you know
is a playable frame rate we understand
but dropping from such a high average to
that kind of low point does cause
noticeable stuttering finally one test
where the GeForce GTX 580 wins hands
down is power consumption impressive
stuff unfortunately for the old Fermi
GPU lower is of course better here so
while impressive its impressively bad by
today's standards the GTX 1050 for
example was often seen delivering over
30% more frames and here we see it did
that while consuming over 50% less power
or at least allowing the system to
consume 50% less power
ok so the performance seen in a number
of those games wasn't particularly great
now the quality settings though they
weren't tuned to suit the GTX 580 those
are simply the settings that we used
previously to test modern low-end GPUs
that have larger vram buffers and those
results were recorded on our core i7 77
ok test system clocked at 4.9 gigahertz
though the CPU used isn't really a big
deal as none of the GPUs are limited by
our modern quad-core processor that said
though since I will be using more
appropriate quality settings in the
gameplay footage that you're about to
see I've also downgraded to the core i3
8100 using ddr4 2,400 memory albeit 610
gigabytes of ddr4 memory right so I
think that's pretty much everything
let's move on to see what the GTX 580
can do in a few more modern games here's
a look at our battlefield 1 runs using
the medium quality preset at Tim ep
after about 5 minutes of gameplay we saw
an average of around 60 FPS though there
were occasional dips down to 45 FPS as
you'd expect the 1.5 gigabyte via Ram
buffer was maxed out for the entire time
and as a result we were eating into
quite a bit of system memory but overall
the results weren't too bad with a
mechanical hard drive they frame dips
and stuttering will likely be much worse
so please note for testing that we were
using the crucial MX 500s SD still
overall the game looked good and played
surprisingly well now we have some Call
of Duty World War 2 results and here the
game had to be turned down to the normal
quality settings and this allowed for
reasonably good performance
the average framerate attorney peace
in the mid 70s while we never dip below
60 FPS so again that's not bad and
certainly very playable please note the
VRAM usage is not displayed in quite a
few of these tests I assume this is some
kind of glitch with the GTX 580 and
rivatuner
Assassin's Creed origins had to be
dialed down to the very low preset and
here we see a number of graphical
anomalies frame rates were also terrible
at turn DP here we saw just 24 FPS on
average with dips down to 18 fps I mean
it is technically playable but by my
standards I'm sure many of you yeah it's
it's not playable don't fall still looks
nice with the media and quality preset
and it played superbly never dropping
below 60 fps and like Call of Duty World
War 2 frame rates often hovered above 70
FPS very armed usage is quite low in
this game though again remember JUnit
isn't displaying that information
unfortunately towards the end of my test
I swapped to the high quality preset
which did improve visuals quite
noticeably but as you'd expect it also
reduced frame rates quite a participa
that's at an average of 53 FPS with dips
down to 48 fps was still very playable
f1 2017 also ran quite nicely at ADP
using the medium quality preset here the
GTX 580 averaged 56 FPS and only dropped
down to 46 fps salt was very playable
and actually very enjoyable fortnight
was also playable using the medium
quality settings I saw an average of
almost 70 FPS with dips down to 53 fps
of course this is a particularly
demanding game and VRAM usage only
peaked just over 1.1 gigabytes so the
RAM usage was quite low hell-blade
played reasonably well here the GTX 580
almost averaged 60fps after a
five-minute test and only dropped down
to 51 FPS so again very playable under
these conditions
Mass Effect and Roman are played
reasonably well with the medium quality
preset enabled in my test run saw an
average of 55 FPS with a minimum of 45
fps overwatch is the kinda game that
really runs well on pretty much anything
and this meant I was able to use the
high quality settings and still enjoy
over 60 FPS at all times at 1080p my 10
minute long tests are an average of 85
FPS for what was a smooth and
experience pray is actually a very well
optimized title but it does love to you
is plenty of vram and this has caused
some issues for the GTX 580 in the graph
seen previously we did see a 1% lower
result of just 39 fps and here we see a
42 FPS minimum which is a reasonable
drop from the 56 fps seen on average
still the game was playable and to be
honest the frame rates aren't really the
issue here rather it was the lack of
textures that spoiled the experience and
that is of course due to the fact that
we can't fit the textures in that 1.5
gigabyte v round buffer of course I have
tested with pub G and my test for this
one ran for 10 minutes and in that time
we saw almost 70 FPS on average with
dips as low as 44 fps overall though the
experience was certainly playable but I
was forced down to the lowest possible
quality things to achieve that
performance finally for testing World of
Tanks I used the built-in benchmark
using the medium quality option which
renders at 1080p here the GTX 580
averaged 142 fps by the end of the test
and never dip below 82 fps so very
playable performance in this title well
there you have it eight years later the
GeForce GTX 580 can still kind of almost
play all the latest titles in fact the
only real exception he was Assassin's
Creed origins I was able to achieve
smooth gameplay in pretty much every
other title that we tested using you
know low to medium type quality settings
so for bargain shoppers is the GTX 580
worth buying well right now you can
expect to pay anywhere from 30 to $60 us
and it's a similar story here in
Australia I've got one for $40 but they
have been selling for as little as $30
ozzie considering the performance you
get out of this card that actually seems
pretty good for a 580 not a particularly
bad option there that said for it to
really make any kind of sense you do
need to get it for about $30 and this is
really because the GTX 680 s can be
quite regularly had for $70 u.s. and
well that's just a much better product
the big issue for the GTX 580 other than
the very limited one and a half gigabyte
frame buffer is of course the power
consumption
it doubled the consumption of our core
i3 8,100 system where you can figure
with the gtx 1050 and as we noted it was
quite often up to over 30% slower so for
that factor alone I'd avoid the GTX 580
and look for something a little more
fuel-efficient on that note I'm gonna
wrap up this rather long video I assume
it's been quite long I hope you enjoyed
the revisit of the GTX 580 if you did
help us out and hit the like button I'm
your host Steve I'll catch you again
next time
We are a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for us to earn fees by linking to Amazon.com and affiliated sites.