views:

140

answers:

8

Is Harddrive rpms really significant for Visual Studio development experience?

A: 

VS2003 with Visual Source Safe was crazy sluggish with lots of disk hits. I currently use vs2005 which still seems to be a little more sluggish than I would like but a marked improvement. I have also noticed better performance since we switched from Visual Source Safe to Team Foundation Server. Hope I don't sound like a M$ hater here, but I have come to expect bloat from Microsoft.

J.Hendrix
+1  A: 

Depends... if you have a really slow harddrive, it will be a problem.

Take a look at: http://weblogs.asp.net/scottgu/archive/2007/11/01/tip-trick-hard-drive-speed-and-visual-studio-performance.aspx

CodeJoust
I have a 7500rpm SAMSUNG HM251JJ
I have a 7500rpm SAMSUNG
Anything Scott Guthrie has to say on the subject must be taken as gospel...
gmagana
+3  A: 

Check to see where your bottleneck lies. If you hit CTRL-SHIFT-ESC, it will bring up task manager and you can view the Performance TAB. When you compile, if your computer is very slow but the CPU usage is low, you are probably HD (IO) bound. If the CPU is at 100%, you are likely CPU bound.

Your two most likely problems for performance if your HD is the bottleneck in VS development is either HD speed itself or lack of memory (and excess paging / bad HD caching).

For better HD performance, you can acquire a fast reliable SSD like one of the Intel Drives or a Western Digital Velociraptor. The Velociraptor is about $230 for 300GB and the Intel SSD is around $500 for 160GB. Either one should be large enough to handle most development projects.

For RAM issues, buy as much RAM as you can afford without buying more than your OS can see (Win32 can only see 3GB usually). Windows will use the extra RAM for HD caching and it will also avoid any paging to memory that would happen on a RAM limited machine. If possible, run a 64-bit version of Windows with 6 or 8 GB of RAM. You can still use VS to compile and build 32-bit projects under Win64 but it will do a much better job of HD caching.

Adisak
+1  A: 

Get an SSD. You will not be disappointed.

Jarrett Meyer
Hmm... why are people modding this down? What's wrong with the statement?
gmagana
I dunno who would mod it down... But my advice would be to get a *GOOD* SSD. There are a lot of older SSD's out there based on the early JMicron controller that have stalling and stuttering issues when doing heavy small random writes (which can happen during compiles and especially happens during links). Right now, the Intel SSD is top of the heap followed by Indilinx controller-based SSDs. Most of the other SSD's aren't worth the extra cost per GB compared to a less expensive and larger capacity 300GB WD Velociraptor.
Adisak
A: 

What matters the most is how quickly the hard drive can access the data (usually called the access time). The faster your drive can access the hundreds of tiny files that the compiler accesses as it builds your project, the better. RPMs may improve the access time a bit, but not as much as switching to drives that use different mechanisms and interfaces.

Unfortunately, most hard drive manufacturers don't publish the average access time. Instead, they promote the seek time (the time it takes to move the read/write head from track to track) and the latency (the time it takes the platters to rotate 1/2 revolution). The access time should be roughly equal to seek+latency.

Most consumer 3.5" SATA drives have a seek time of 8-12 ms, a figure which has remained relatively unchanged even as the RPMs went from 5400 to 7200.

By contrast, SCSI and SAS drives have seek times of 3.5-4.5 ms. They've always been faster than SATA/ATA drives, due to their more robust actuators and better interfaces.

SSD drives have near-zero access time.

Barry Brown
A: 

You will likely not notice a huge difference between the RPMs, but if you have the ability to install a second hard drive you can have VS and your project files running off of that so windows is more free to hit its system and paging files.

shogun
A: 

I'm sure everyone knows this already, but what the heck... fragmentation is evil. It's surprising how many developer systems have badly fragmented NTFS drives. Adding a nightly scheduled defrag task is an easy way to mitigate this, at least on XP systems.

Also, when purchasing a drive, another reason to get a really big one is to avoid filling the drive up to the point where defragging starts to fail.

MCP
A: 

Personally I'm a fan of Jeff's suggestion, that being to use an SSD for a boot drive and key applications, and use a more cost-effective drive for actual data storage. I'm going to implement this myself once I eventually find some time to backup data and pull apart hardware.

Tullo