Dense disks make for smart storage

By
Friday, 01 January, 2010


Disk drives will soon reach 10 terabytes, but the storage industry is worrying about how these new drives will complicate storage, rather than rejoicing in extra capacity. Simon Sharwood explains the new technologies pushing disk capacity higher and explores why larger disks are not entirely good news.

You can take it for granted that the technology industry will produce bigger, faster, more capable devices every year, and that users will lap up the increased power on offer. The storage industry, however, is worried about the consequences that will result from an imminent growth spurt that will accelerate disk drives’ capacity to 10 terabytes and beyond.

The spurt will come thanks to advances in a field called ‘areal density’, an umbrella term that describes the quantity of data that can be stored on a disk drive*.

Areal density has improved markedly in recent years, thanks to innovations in the magnetic materials (disk wonks jokingly call this ‘rust’) used on the surface of the platters that whirr around inside disk drives.

These materials have improved thanks to advances in the science of ‘magnetoresistance’, a quality of some metallic substances that sees them change their electrical resistance when exposed to a magnetic field. Magnetoresistance is useful in disk drives because changing the electrical resistance of disk media is a great way to represent a zero or a one.

Various researchers have, over the years, discovered materials with greater levels of magnetoresistance, which allows data to be written with greater precision so that a smaller area of a disk’s surface is required to record a one or a zero. IBM famously dubbed one such innovation, its 2001 addition of the rare metal Ruthenium to its disks, as ‘pixie dust’ and boasted of areal density up to 25.7 gigabits (around 3.2 gigabytes) per square inch.

Pixie dust boosted areal density, but was held back by the prevalent ‘longitudinal’ method of recording data that sees information smeared across a disk’s surface. A newer technology, ‘perpendicular’ recording, changes the magnetism of media by sending a charge down inside the rust instead of just changing the surface layer’s value. The deeper changes mean that information is written to smaller spaces on the disk’s surface, increasing areal density.

Perpendicular recording is today’s dominant technology, but disk-makers believe it is running out of steam and are investigating two new technologies to bring us more advances in areal density.

One of those technologies is called Patterned Media, an idea which will see disks divided into millions of tiny, discrete ‘cells’ that are smaller than the areas of disk that perpendicular recording currently requires to record data. Patterned Media will therefore be able to pack more data into less space.

The second technology has the spiffy acronym HAMR - short for heat assisted magnetic recording - and involves new materials that can store data in smaller spaces, provided they are heated before data is written. HAMR drives will therefore include a small laser which zaps the surface of a disk to warm it up before data is written.

Development of HAMR is already advanced. In early 2009, drive-maker Seagate demonstrated a 250 GB drive using the technology. TDK demonstrated HAMR technology at an October 2009 trade show in Japan and mentioned one-terabit-per-square-inch densities.

Hitachi GST, the main proponent of Patterned Media, has often predicted its debut in 2011 or 2012.

Both companies predict that by 2012 or 2013, their technologies will result in disk drives that easily reach 10 terabytes. By contrast, LTO-6 - the largest mainstream tape technology currently envisaged - will reach 3.6 terabytes (uncompressed) in 2012 or 2013. Seagate and Hitachi GST are both optimistic that they can use HAMR and/or Patterned Media to reach 50 terabyte drives before the 2010s come to an end.

Big is problematic

The storage industry, however, is not yet certain what to do with very large disk drives.

Simon Elisha, Hitachi Data Systems’ Chief Technologist for Australia and New Zealand, believes they will meet some users’ needs by occupying less space than their predecessors.

“I was talking to a customer who has decommissioned a [storage] system and archived the data,” he explains. “Only four people will ever need to see the data, but when they do they will need it ASAP, so they want it to take up a very small amount of space in the data centre.

“That is where people are seeing large drives as having some benefit.”

But size is not important for everyone, says Clive Gold, EMC’s Chief Technology Officer, Marketing.

“Today, we can give you 75 petabytes of storage in a 19 rack, and we are not even using 2.5 drives,” he says. “The problem today is speed,” he contends, as it will take longer to read data from a large drive than from a smaller device. “It could take a month to rebuild a 10 TB drive in a RAID set,” a period of time that makes RAID - the mainstay of data protection - considerably less useful!

Sean Derrington, Director of Symantec’s Storage and Availability Management Group, also has worries about speed.

“The slowest part of the input/output equation is always the mechanical drive,” Derrington says. No matter what the size, he says, “they will all have the same rotational latency”, which means that a very large hard disk may store more information, but will generally be slower at delivering it to applications as the drive will have more work to do to find data.

Faster hard disks could theoretically solve this problem, making larger capacity drives better and delivering data to applications, but faster drives consume more electricity and make more heat. Many data centres simply cannot access the electricity needed for more power-hungry appliances or for cooling equipment to ensure their continued operation, making faster drives unlikely.

Instead, it is likely that large capacity drives will find a home in tiered storage implementations. Organisations will take advantage of their size to tackle space issues, but use them alongside faster storage media like solid state disks to ensure that the data which is frequently required by an application is delivered faster than large mechanical drives can deliver it.

“You can get some flash drives in to boost input/output rates,” says EMC’s Gold. “Maybe between flash and serial ATA drives you have some fibre channel drives in a tiered model that moves data around.

“If I know which data will be hot, I move it to flash, or if it is not used a lot, push it to a lower cost tier of storage,” such as a very large disk.

But doing so, warns Symantec’s Derrington, will mean a need for greater management of storage arrays.

“Organisations will need to look at input/output all the way from the application to the individual spindle [disk] on which data resides,” he says, as only by understanding the capabilities of every storage device will it be possible to tune systems to give applications the performance they require.

“If you do not have a plan in place to manage storage utilisation you cannot optimise it, so if you do not know where the information resides - the serial number of a spindle - you cannot optimise storage.”

New opportunities

Gold also believes that the advent of very large disks may change the way we think about storage.

“We need to think about the format written to the drive,” he says. “Today, we take ones and zeros and put them onto blocks and sectors of a disk.” A new alternative is ‘binary large objects’ (BLOBs), which he says “make a drive an object store rather than a block store. Today, one of the issues is that metadata is buried in file structure. But if we take a metadata layer and abstract it, we can search an index before we get into the object itself.” The result is faster data retrieval, a possible boost to the performance of a very large drive and therefore a potential route to their wider use.

Another possible change large drives could force, according to NetApp’s Martin, is for “backup as we know it to disappear”.

“It is going to become impossible to use RAID because the rebuild times will be so long,” he opines. “Most file level backups are random, so you will see that style of backup disappear in favour of image-style backups.”

* The industry measures areal density in bits per square inch and has settled on this measure because while drives are named after their sizes - 2.5 and 3.5″ drives are the main form factors on sale today - the disks inside a drive vary in size depending on vendors’ whims.

Related Articles

Seven predictions that will shape this year

Pete Murray, Managing Director ANZ for Veritas Technologies, predicts trends that will have a...

ARENA jointly funds Vic's first large-scale battery storage

Two large-scale, grid-connected batteries are to be built in Victoria with the help of the...

Protecting next-gen storage infrastructures

Companies looking to modernise their overall IT infrastructure cannot afford to take a relaxed...


  • All content Copyright © 2024 Westwick-Farrow Pty Ltd