What's new
  • Please do not post any links until you have 3 posts as they will automatically be rejected to prevent SPAM. Many words are also blocked due to being used in SPAM Messages. Thanks!

ASUS Radeon HD 5770 1GB Voltage Tweak Edition Review

Status
Not open for further replies.

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal


ASUS Radeon HD 5770 1GB Voltage Tweak Edition Review





Manufacturer's Part Number: EAH5770/2D/IS/1GD5/A
Price: Approx. $160USD
Warranty: 3-Years



Not long ago, ATI was late to market with their DX10-supporting architecture and it looked like NVIDIA was about to run away with the performance crown. Somehow even after the minor disaster that the HD 2900-series was, the boys in red were able to hold on for dear life until they had products that could compete on a level footing with the best the competition could offer. If we fast forward to the present day, the roles of each company have dramatically reversed with ATI taking its first plunge into the DX11 marketplace while NVIDIA is struggling to find a way to answer.

Usually when new architectures are released, graphics card companies are content to focus the first cards on the high end enthusiast segment (with astronomical prices to boot) and wait a few months for the market to settle before attacking the lower price brackets. ATI is now bucking that trend. They aren’t content to just go after the competition’s jugular when it comes to high-end cards but are rather aiming to establish a stranglehold on the entire market, top to bottom. Today marks the day that the DX11 generation moves into the affordable mainstream performance market with the availability of ATI’s HD 5770 and HD 5750 cards. While both cards are aimed for basically the same type of consumers, they may prove to be quite different from one another when it comes to overall performance.

The HD 5770 1GB is supposed to be the spiritual successor to the popular HD 4770 512MB card from a few months ago. Priced around $160USD, it is also aimed to bridge the gap between NVIDIA’s own GTS 250 and GTX 260 216 series of cards that have had the limelight in the sub-$225 category for some time now. However, this isn’t to say that other ATI cards will be safe either since the red team has the HD 4850 512MB, HD 4770 512MB and the now-EOL HD 4830 512MB all within spitting distance of this new card as well.

The card we will be looking at in this review is the ASUS HD 5770 1GB Voltage Tweak Edition. Even though ASUS doesn’t have the warranty length of some of their competitors and this card sticks to stock clocks, they are trying to spice things up with included voltage adjustment software that is supposed to help increase the overclockability of their cards. This should help them stand out from the crowd come launch day which is typically cluttered with reference-based cards and not much else.

All in all, the HD 5770 1GB is looking to be a great card for the market it is aimed at. We know people are anxious for it as well so it should also prove to be a boon for retailers in these somewhat unsettled times.

 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
A Look at the ATI 5000-series

A Look at the ATI 5000-series



As you can probably tell by the chart above, all of the HD 5000-series fit perfectly into ATI’s current lineup. In essence, The HD 5870 takes the place of the expensive-to-produce and comparably inefficient dual GPU HD 4870X2 as the top dog for the time being. Judging from paper specifications alone, the HD 5870 is a technological marvel considering it packs all of the rendering potential of ATI’s past flagship card and then some while not being saddled by an inefficient dual processor design. The fact that this new card could trump the performance of a HD 4890 just a few months after that card’s release is nothing short of stunning.


The HD 5850 on the other hand looks to be the purebred price / performance leader of the new ATI lineup. Barring slightly lower clock speeds for both the core and memory along with eight disabled texture units (totalling 160 stream processors), it is basically a clone of the HD 5870. This is the card ATI hopes will compete directly with the GTX 285 for the near future and then come into its own when DX11 games make their way into the market. We believe this card will appeal to the majority of early adopters since it allows them to buy class-leading DX9 and DX10 performance now without gambling $400 on unproven DX11 potential.

We can also see that ATI did some careful price cutting prior to launch since even though the HD 4890 looks to offer significantly less performance than a HD 5850, it is actually priced accordingly. As such, this previously high end card will stick around for the next few months in the $200 price bracket but that isn’t to say that it will stay there indefinitely...


In short order, ATI will have a full range of DX11 cards on the market; all of which have been talked about in rumours over the last quarter. To begin with we will see the two “Cypress” series cards which are the HD 5870 and HD 5850 followed before the new year by the dual GPU Hemlock card which will make use of two Cypress processors. The Hemlock sticks to ATI’s mantra of never releasing a card that retails for above $500 but it will nonetheless take over the premier position of this DX11 lineup.

Meanwhile, we now have the HD 5700-series of code-named Juniper cards as well with the HD 5770 and HD 5750. The HD 5770 1GB is one of the first sub-$200 cards which will come stock with a 1GB framebuffer and along with the GDDR5 memory, comes with some hefty clock speeds as well. However, even though upon first glance the HD 5770 looks like it can compete with the HD 4890, this isn’t the case. According to ATI, the 128-bit memory interface will limit this card’s performance so it lies right within its stated price range. We should also mention that ATI won’t be replacing the HD 4890 until at least the first quarter of 2010 even though the HD 5770 is looking to take over from the HD 4850.

The HD 5750 on the other hand is simply a cut down HD 5770 with lower clocks, less SPs and a cut down number of Texture Units. It is this card that ATI sees going head to head with the NVIDIA GTS 250 and 9800 GT. It uses GDDR5 memory but there will be both 512MB and 1GB versions released to cater to the $100 market along with those looking for a little jump in performance.

So there you have it. In the high stakes game of poker that is the GPU industry, ATI has shown its hand. All that is left is for the competition to respond.
 
Last edited:

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
Focusing on DX11

Focusing on DX11


It has been a hair under three years since the release of Windows Vista and with it the DirectX 10 API. In that amount of time, a mere 33 DX10 games were released. That isn’t exactly a resounding success considering the hundreds of titles released in that same time. Let’s hope DX11 does a bit better than that.


DX11 is focused on taking the lessons learned from the somewhat inefficient DX10 and shaping them into a much more efficient API which will demand less system resources while being easier to develop for. In addition to the usual 3D acceleration, it will also be used to speed up other applications which in the past have not been associated with the DirectX runtime. This may be a tall order but with the features we will be discussing here, developers have already started using DX11 to expand the PC gaming experience. It is an integral component in Windows 7 and according to Microsoft, will also be adopted into Windows Vista through a software update.

Let’s scratch the surface of what DX11 can bring to the table.


Unlike past DirectX versions, DX11 endeavours to move past the purely graphics-based uses of the API and push it towards being the lynchpin of an entire processing ecosystem. This all begins with the power which DirectX Compute will bring into the fold. Not only can it increase the efficiency of physics processing and in-game NPC intelligence within games by transferring those operations to the GPU but it can also be used to accelerate non-3D applications.




Through the use of Compute Shader programs in Shader Model 5.0, developers are able to use additional graphical features such as order independent transparency, ray tracing, and advanced post-processing effects. This should add a new depth of realism to tomorrow’s games and as mentioned before, also allow for programs requiring parallel processing to be accelerated on the GPU.


For the majority of you reading this review, it is the advances in graphics processing and quality that will interest you the most. As games move slowly towards photo-realistic rendering quality, new technologies must be developed in order to improve efficiency while adding new effects.


Some of the technologies that ATI is championing are DX11’s new Depth of Field, OIT (or Order Independent Transparency) and Detail Tessellation. While the pictures above do a good job of showing you how each of these works, it is tessellation which ATI seems most excited about. They have been including hardware tessellation units in their GPUs for years now and finally with the dawn of DX11 will these units be finally put to their full use. OIT on the other hand allows for true transparency to be added to an object in a way that will be more efficient resource-wise than the standard alpha blending method currently used.


Let’s talk about DX11 games. As you would expect, due to the ease of programming for this new API and the advanced tools it gives developers, many studios have been quite vocal in their support. Even though some of the titles listed above may not be high on your list of must have games, A-list titles like the upcoming Aliens vs. Predator from Rebellion and DiRT 2 are sure to get people interested. What we like see is at least three DX11 games being available before the Christmas buying season even though BattleForge is already available and will have DX11 support added through a patch.

Another exciting addition to the list is EA DICE’s FrostBite 2 Engine which will power upcoming Battlefield games. Considering the popularity of this series, the inclusion of DX11 should open up this API to a huge market.

 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
OpenCL: The Next Big Thing?

OpenCL: The Next Big Thing?



As consumers, we have all heard of the inroads GPUs have been making towards offering stunning performance in compute-intensive applications. There have been attempts to harness this power by engines such as NVIDIA’s Compute Unified Device Architecture (CUDA) and ATI’s Stream SDK (which in v2.0 supports OpenCL).


“Build it and the will come” says the old mantra but industry adoption of CUDA and Stream was anything but quick since there were two standards being pushed for the same market. CUDA in particular is having a hard time of it since it is vendor-specific without hardware support from any other vendor. The industry needed a language that was universal and available across multiple platforms. That’s were OpenCL (Open Computing Language) along with DirectX Compute come into play. It is completely open-source and managed by a non-profit organization called the Khronos Group which also has control over OpenGL and OpenAL


At its most basic level, OpenCL is able to be executed across multiple mediums such as GPUs, CPUs and other types of processors. This makes it possible to prioritize workloads to the processor that will handle them most efficiently. For example, a GPU is extremely good at crunching through data-heavy parallel workloads while an x86 CPU is much more efficient at serial and task-specific This also allows developers to write their programs for heterogeneous platforms instead of making them specific to one type of processor.


So what does this mean for gamers? First of all, AMD has teamed up with Bullet and PixeLux in order to achieve more realistic environments for players. The Bullet Physics is an open-source physics engine which has an ever-expanding library for soft body, 3D collision detection and other calculations. Meanwhile, PixeLux uses their DMM (Digital Molecular Matter) engine which uses the Finite Element Analysis Method of calculating physics within a game. In past applications, it has been used to calculate actions which have an impact on the game’s environment such as tumbling rubble or debris movement.


With Stream moving to OpenCL, ATI is truly moving towards an open platform for developers which they are hoping will lead to broader developer and market adoption than the competition’s solutions. At this point it looks like we will soon see ATI’s GPUs accelerating engines from Havok, PixeLux and Bullet through the use of OpenCL. Considering these are three of the most popular physics engines on the market, ATI is well placed to make PhysX a thing of the past.
 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
ATI’s Eyefinity Technology

ATI’s Eyefinity Technology



The term Surround Gaming may not mean much to many of you who are reading this article but with the advent of ATI’s new Eyefinity technology, now is a good time to educate yourself. Basically, Eyefinity will give users the ability to use multiple monitors all running from the same graphics card. In the past, simple dual monitor setups have been used by many graphics, CAD or other industry professionals in order to increase their productivity but gaming on more than one monitor was always a bit of a clunky affair. Granted, some products like Matrox’s TripleHead2Go were able to move multi monitor setups into the public’s perception but there were always limitations (resolution and otherwise) associated with them. ATI is aiming to make the implementation of two or even more monitors as seamless as possible within games and productivity environments while offering the ability to use extreme resolutions.


While the price of two or even three new monitors may be a bit daunting at first for many of you, but good 20” and even 22” LCDs have come down in price to the point where some are retailing below the $200 mark. ATI figures that less than $600 for three monitors will allow plenty of people to make the jump into a true surround gaming setup. Indeed, with three or even six monitors, the level of immersion could be out of this world.


The reason that main in the professional field are familiar with multi monitor setups is for one simple matter: they increase productivity exponentially. Imagine watching a dozen stocks without having to minimize windows all the time or using Photoshop on one screen while watching a sports broadcast on another and using the third screen for Photoshop’s tooltips. The possibilities are virtually limitless if it is implemented properly.


When it comes to a purely gaming perspective, the thought of a massive view of the battlefield or the ability to see additional enemies in your peripheral vision is enough to make most gamers go weak in the knees. Unfortunately, the additional monitors will naturally mean decreased performance considering the massive amount of real-estate that would need rendering. This will mean tradeoffs may have to be made in terms of image quality if you want to use Eyefinity.


According to ATI, all of the new HD 5800-series graphics cards will have the ability to run up to three monitors simultaneously. This is done by having a pair of DVI connectors as well as a DisplayPort and HDMI connector located on the back of the card. It should be noted that ATI will be releasing a special Eyefinity version of the HD 5870 in the coming months which features six DisplayPort connectors for those of you who want to drive six monitors from a single card.


This technology is all made possible through the use of DisplayPort connectors but this also provides a bit of a limitation as well. Above we can see that a number of 3-screen output combinations which the current HD5800-series support and one thing is constant: you will need at least one monitor which supports DisplayPort. Unfortunately, at this time DP-supporting monitors tend to carry a price premium over standard screens which will increase the overall cost of an Eyefinity setup. Luckily the other two monitors can either use DVI or a combination of DVI and HDMI for connectivity.
 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
Packaging and Accessories

Packaging and Accessories



When buying a sub-$200 card, customers don’t want to spend a fortune on shipping so ASUS drastically cut down the usual gargantuan-sized box to something a little more palatable. Even though it is short and narrow, ASUS was able to fit an incredible amount of information onto the exterior cardboard sleeve. Along with the Voltage Tweak sticker, there is mention of the DisplayPort, HDMI, 1GB GDDR5 and the included Smart Doctor software.


We have come to expect good protection within the ASUS cards’ boxes and this one doesn’t disappoint. The card itself is nestled within a cocoon of cardboard while being additionally wrapped in an anti-static bag and the accessories are pushed off to the side in their own separate compartment.


The accessory package is a bit bare but you get all of the necessities including a Crossfire bridge and DVI to VGA adaptor. However, there wasn’t an included Molex to PCI-E adaptor so we are guessing ASUS is assuming that everyone who buys this card will have a compatible PSU. When it comes to software, there is a driver CD that includes the Smart Doctor software as well as a CD with a multi-lingual manual.
 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
A Closer Look at the ASUS HD 5770 Voltage Tweak Edition

A Closer Look at the ASUS HD 5770 Voltage Tweak Edition



The ASUS HD 5770 1GB doesn’t differ from any of the other reference-based HD 5770 cards you will see come launch day even though it incorporates Voltage Tweak within the software package. It is a very clean looking card whose heatsink resembles that of the HD 5850 and HD 5870 with a predominantly black color along with a few red accents. It actually looks like a high performance card wrapped into a small package.


ASUS seems to be doing things a bit differently with their HD 5770 since the heatsink shroud is devoid of any extravagant stickers that shout out tackiness from every corner. Instead, ASUS leaves the shroud as is and instead applies a simple logo and fan hub sticker. We love this understated minimalism since we all know that good taste doesn’t have to make itself heard.


The side of the card houses a pair of Crossfire connectors as well as a long piece of red plastic with the ATI Radeon logo. This area not only provides an extra section where the accent color is used but it also dubs as a vent for any hot air that can’t be exhausted through the backplate.


The rearmost portion of the HD 5770 1GB features a pair of gaudy red-fringed vent holes that are used by the fan to suck cool air over the VRMs. Unfortunately, since the plastic heatsink shroud projects over the edge of the PCB, attaching the PCI-E 6-pin connector to the card can be a challenge, especially in a cramped case.

Meanwhile, the backplate makes use of its dual slot design to give the user a pair of DVI connectors as well as ports for DisplayPort and HDMI.


The back of the card shows us a black PCB with a quartet of GDDR5 memory modules. These ICs are H5GQ1H24AFR units from Hynix and are rated for 1.25Ghz (5Ghz QDR) speed at 1.5V.


Length-wise, the HD 5770 is about 8 ¾” in length which is slightly shorter than the HD 5850 but slightly longer than the pint-sized HD 5750.
 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
Test System & Setup

Test System & Setup

Processor: Intel Core i7 920(ES) @ 4.0Ghz (Turbo Mode Enabled)
Memory: Corsair 3x2GB Dominator DDR3 1600Mhz
Motherboard: Gigabyte EX58-UD5
Cooling: CoolIT Boreas mTEC + Scythe Fan Controller
Disk Drive: Pioneer DVD Writer
Hard Drive: Western Digital Caviar Black 640GB
Power Supply: Corsair HX1000W
Monitor: Samsung 305T 30” widescreen LCD
OS: Windows Vista Ultimate x64 SP1


Graphics Cards:

XFX HD 5770
XFX HD 5750
XFX HD 5850 (Reference)
ATI HD 4890 (Reference)
Sapphire HD 4850 (Reference)
Diamond HD 4770 (Reference)
EVGA GTX 260 216 (Reference)
EVGA GTS 250 1GB (Reference)
EVGA GTS 250 512MB (Reference)
9800 GT 512MB (Reference)


Drivers:

ATI 8.66 RC7 Beta (HD 5000-series)
ATI 9.9 WHQL
NVIDIA 191.07 WHQL


Applications Used:

Call of Duty: World at War
Call of Juarez: Bound in Blood
Crysis: Warhead
Dawn of War II
Fallout 3
Far Cry 2
Left 4 Dead
Tom Clancy’s HawX


*Notes:

- All games tested have been patched to their latest version

- The OS has had all the latest hotfixes and updates installed

- All scores you see are the averages after 4 benchmark runs

All game-specific methodologies are explained above the graphs for each game

All IQ settings were adjusted in-game
 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
Call of Duty: World at War

Call of Duty: World at War



To benchmark this game, we played through 10 minutes of the third mission (Hard Landing) starting from when the player first enters the swamp, through the first bunker until the final push onto the airfield. This was benchmarked using FRAPS.


1680 x 1050





1920 x 1200





2560 x 1600





Lately, it seems like Call of Duty: World at War has been particularly hard on the performance of ATI cards but the HD 5770 actually does quite well. In the majority of the tests, it stays behind the GTX 260 216 and gradually pulls away from the GTS 250 1GB as the resolution increases. We can also see that the gap between it and the HD 5750 is significant to say the least.
 

SKYMTL

HardwareCanuck Review Editor
Staff member
Joined
Feb 26, 2007
Messages
13,421
Location
Montreal
Call of Juarez II: Bound in Blood

Call of Juarez II: Bound in Blood



CoJ is a bit of an oddity without any in-game AA options but nonetheless, it looks incredible. For this benchmark we used a 10 minute gameplay sequence which included panoramic views of a town and gun battles. FRAPS was used to record the framerates.

1680 x 1050



1920 x 1200



2560 x 1600



Call of Juarez is an NVIDIA card’s worst nightmare and the HD 5770 is able to manhandle even the more expensive GTX 260 216 at every resolution. Once again, it is also lightyears ahead of the HD 5750 1GB.
 
Status
Not open for further replies.

Latest posts

Twitter

Top