Studies show that the amount of data being recorded is increasing at 30 to 40
percent per year. At the same time, the capacity of modern hard drives,
which are used to store most of this, is increasing at less than half
that rate. Fortunately, much of this information doesn’t need to be
accessed instantly. And for such things, magnetic tape is the perfect
solution.
Seriously? Tape? The very idea may evoke images of reels rotating fitfully next to a bulky mainframe in an old movie like Desk Set or Dr. Strangelove. So, a quick reality check: Tape has never gone away!
Indeed,
much of the world’s data is still kept on tape, including data for
basic science, such as particle physics and radio astronomy, human
heritage and national archives, major motion pictures,
banking, insurance, oil exploration, and more. There is even a cadre of
people (including me, trained in materials science, engineering, or
physics) whose job it is to keep improving tape storage.
Tape has
been around for a long while, yes, but the technology hasn’t been frozen
in time. Quite the contrary. Like the hard disk and the transistor, magnetic tape has advanced enormously over the decades.
The first commercial digital-tape storage system, IBM’s Model 726,
could store about 1.1 megabytes on one reel of tape. Today, a modern
tape cartridge can hold 15 terabytes. And a single robotic tape library
can contain up to 278 petabytes of data. Storing that much data on
compact discs would require more than 397 million of them, which if
stacked would form a tower more than 476 kilometers high.
It’s
true that tape doesn’t offer the fast access speeds of hard disks or
semiconductor memories. Still, the medium’s advantages are many. To
begin with, tape storage is more energy efficient: Once all the data has
been recorded, a tape cartridge simply sits quietly in a slot in a
robotic library and doesn’t consume any power at all. Tape is also
exceedingly reliable, with error rates that are four to five orders of
magnitude lower than those of hard drives. And tape is very secure, with
built-in, on-the-fly encryption and additional security provided by the
nature of the medium itself. After all, if a cartridge isn’t mounted in
a drive, the data cannot be accessed or modified. This “air gap” is
particularly attractive in light of the growing rate of data theft
through cyberattacks.
The offline nature of tape also provides an
additional line of defense against buggy software. For example, in 2011,
a flaw in a software update caused Google to accidentally delete
the saved email messages in about 40,000 Gmail accounts. That loss
occurred despite there being several copies of the data stored on hard
drives across multiple data centers. Fortunately, the data was also
recorded on tape, and Google could eventually restore all the lost data from that backup.
The
2011 Gmail incident was one of the first disclosures that a
cloud-service provider was using tape for its operations. More recently,
Microsoft let it be known that its Azure Archive Storage uses IBM tape storage equipment.
All these pluses notwithstanding, the main reason why companies use
tape is usually simple economics. Tape storage costs one-sixth the
amount you’d have to pay to keep the same amount of data on disks, which
is why you find tape systems almost anyplace where massive amounts of
data are being stored. But because tape has now disappeared completely
from consumer-level products, most people are unaware of its existence,
let alone of the tremendous advances that tape recording technology has
made in recent years and will continue to make for the foreseeable
future.
All this is to say that tape has been with us for decades and will be here for decades to come. How can I be so sure? Read on.
Tape has survived
for as long as it has for one fundamental reason: It’s cheap. And it’s
getting cheaper all the time. But will that always be the case?
You
might expect that if the ability to cram ever more data onto magnetic
disks is diminishing, so too must this be true for tape, which uses the
same basic technology but is even older. The surprising reality is that
for tape, this scaling up in capacity is showing no signs of slowing.
Indeed, it should continue for many more years at its historical rate of
about 33 percent per year, meaning that you can expect a doubling in
capacity roughly every two to three years. Think of it as a Moore’s Law
for magnetic tape.
That’s great news for anyone who has to deal
with the explosion in data on a storage budget that remains flat. To
understand why tape still has so much potential relative to hard drives,
consider the way tape and hard drives evolved.
Both rely on the
same basic physical mechanisms to store digital data. They do so in the
form of narrow tracks in a thin film of magnetic material in which the
magnetism switches between two states of polarity. The information is
encoded as a series of bits, represented by the presence or absence of a
magnetic-polarity transition at specific points along a track. Since
the introduction of tape and hard drives in the 1950s, the manufacturers
of both have been driven by the mantra “denser, faster, cheaper.” As a
result, the cost of both, in terms of dollars per gigabyte of capacity,
has fallen by many orders of magnitude.
These cost reductions are
the result of exponential increases in the density of information that
can be recorded on each square millimeter of the magnetic substrate.
That areal density is the product of the recording density along the
data tracks and the density of those tracks in the perpendicular
direction.
Early on, the areal densities of tapes and hard drives
were similar. But the much greater market size and revenue from the sale
of hard drives provided funding for a much larger R&D effort, which
enabled their makers to scale up more aggressively. As a result, the
current areal density of high-capacity hard drives is about 100 times
that of the most recent tape drives.
Nevertheless, because they
have a much larger surface area available for recording,
state-of-the-art tape systems provide a native cartridge capacity of up to 15 TB—greater
than the highest-capacity hard drives on the market. That’s true even
though both kinds of equipment take up about the same amount of space.
With the exception of capacity, the performance characteristics of
tape and hard drives are, of course, very different. The long length of
the tape held in a cartridge—normally hundreds of meters—results in
average data-access times of 50 to 60 seconds compared with just 5 to 10
milliseconds for hard drives. But the rate at which data can be written
to tape is, surprisingly enough, more than twice the rate of writing to
disk.
Over the past few years, the areal density scaling of data
on hard disks has slowed from its historical average of around 40
percent a year to between 10 and 15 percent. The reason has to do with
some fundamental physics: To record more data in a given area, you need
to allot a smaller region to each bit. That in turn reduces the signal
you can get when you read it. And if you reduce the signal too much, it
gets lost in the noise that arises from the granular nature of the
magnetic grains coating the disk.
It’s possible to reduce that
background noise by making those grains smaller. But it’s difficult to
shrink the magnetic grains beyond a certain size without compromising
their ability to maintain a magnetic state in a stable way. The smallest
size that’s practical to use for magnetic recording is known in this
business as the superparamagnetic limit. And disk manufacturers have reached it.
Until
recently, this slowdown was not obvious to consumers, because
disk-drive manufacturers were able to compensate by adding more heads
and platters to each unit, enabling a higher capacity in the same size
package. But now both the available space and the cost of adding more
heads and platters are limiting the gains that drive manufacturers can
make, and the plateau is starting to become apparent.
There are a
few technologies under development that could enable hard-drive scaling
beyond today’s superparamagnetic limit. These include heat-assisted magnetic recording (HAMR) and microwave-assisted magnetic recording
(MAMR), techniques that enable the use of smaller grains and hence
allow smaller regions of the disk to be magnetized. But these approaches
add cost and introduce vexing engineering challenges. And even if they
are successful, the scaling they provide is, according to manufacturers,
likely to remain limited. Western Digital Corp., for example, which
recently announced that it will probably begin shipping MAMR hard drives in 2019, expects that this technology will enable areal density scaling of only about 15 percent per year.
In
contrast, tape storage equipment currently operates at areal densities
that are well below the superparamagnetic limit. So tape’s Moore’s Law
can go on for a decade or more without running into such roadblocks from
fundamental physics.
Still, tape is a tricky technology. Its
removable nature, the use of a thin polymer substrate rather than a
rigid disk, and the simultaneous recording of up to 32 tracks in
parallel create significant hurdles for designers. That’s why my
research team at the IBM Research–Zurich lab
has been working hard to find ways to enable the continued scaling of
tape, either by adapting hard-drive technologies or by inventing
completely new approaches.
In 2015, we and our collaborators at FujiFilm Corp. showed that by using ultrasmall barium ferrite particles oriented perpendicular to the tape,
it’s possible to record data at more than 12 times the density
achievable with today’s commercial technology. And more recently, in collaboration with Sony Storage Media Solutions,
we demonstrated the possibility of recording data at an areal density
that is about 20 times the current figure for state-of-the-art tape
drives. To put this in perspective, if this technology were to be
commercialized, a movie studio, which now might need a dozen tape
cartridges to archive all the digital components of a big-budget
feature, would be able to fit all of them on a single tape.
To enable this degree of scaling, we had to make a
bunch of technical advances. For one, we improved the ability of the
read and write heads to follow the slender tracks on the tape, which
were just 100 or so nanometers wide in our latest demo.
We also
had to reduce the width of the data reader—a magnetoresistive sensor
used to read back the recorded data tracks—from its current micrometer
size to less than 50 nm. As a result, the signal we could pick up with
such a tiny reader got very noisy. We compensated by increasing the
signal-to-noise ratio inherent to the media, which is a function of the
size and orientation of the magnetic particles as well as their
composition and the smoothness and slickness of the tape surface. To
help further, we improved the signal processing and error-correction
schemes our equipment employed.
To ensure that our new prototype
media can retain recorded data for decades, we changed the nature of the
magnetic particles in the recording layer, making them more stable. But
that change made it harder to record the data in the first place, to
the extent that a normal tape transducer could not reliably write to the
new media. So we used a special write head that produces magnetic
fields much stronger than a conventional head could provide.
Combining
these technologies, we were able to read and write data in our
laboratory system at a linear density of 818,000 bits per inch. (For
historical reasons, tape engineers around the world measure data density
in inches.) In combination with the 246,200 tracks per inch that the
new technology can handle, our prototype unit achieved an areal density
of 201 gigabits per square inch. Assuming that one cartridge can hold
1,140 meters of tape—a reasonable assumption, based on the reduced
thickness of the new tape media we used—this areal density corresponds
to a cartridge capacity of a whopping 330 TB. That means that a single
tape cartridge could record as much data as a wheelbarrow full of hard
drives.
In 2015,the Information Storage Industry Consortium,
an organization that includes HP Enterprise, IBM, Oracle, and Quantum,
along with a slew of academic research groups, released what it called
the “International Magnetic Tape Storage Roadmap.” That forecast
predicted that the areal density of tape storage would reach 91 Gb per
square inch by 2025. Extrapolating the trend suggests that it will
surpass 200 Gb per square inch by 2028.
The authors of that road
map each had an interest in the future of tape storage. But you needn’t
worry that they were being too optimistic. The laboratory experiments
that my colleagues and I have recently carried out demonstrate that 200
Gb per square inch is perfectly possible. So the feasibility of keeping
tape on the growth path it’s had for at least another decade is, to my
mind, well assured.
Indeed, tape may be one of the last
information technologies to follow a Moore’s Law–like scaling,
maintaining that for the next decade, if not beyond. And that streak in
turn will only increase the cost advantage of tape over hard drives and
other storage technologies. So even though you may rarely see it outside
of a black-and-white movie, magnetic tape, old as it is, will be here
for years to come.