FUJIFILM INSIGHTS BLOG

Data Storage

The Advantages of an Active Archive in Today’s Data-Flooded World

Reading Time: 4 minutes

The vast volumes of data created daily, coupled with the opportunity to derive value from that data, is making active archives an increasingly important part of organizations’ data management game plans across the globe.

In this Q&A, Active Archive Alliance Chairman, Peter Faulhaber, FUJIFILM Recording Media, U.S.A., Inc., shares his perspective on the role of active archives in managing the data deluge.

Q: What are some of the key trends driving the shift to active archive?

A: I would say the relentless rate of data growth and how to manage it. The answer lies in proper data classification and moving data to the right tier of storage at the right time. Analysts say that 60% of data becomes archival after 90 days or less. So there is a need to cost-effectively store, search for and retrieve enormous volumes of rapidly growing archival content.

Q: So what exactly does an active archive enable?

A: An active archive enables online access to data throughout its lifecycle regardless of which tier it resides in the storage hierarchy. Active archive file systems span all pools of storage, whether they are SSDs, HDDs, tape or cloud. But tape is a key enabler. Since tape has the lowest total cost of ownership for long term data retention, you can cost-effectively maintain on-line access to all of your data in an active archive.

Q: Speaking of tape, is cloud killing tape?

A: That’s a misconception as cloud storage providers such as Microsoft Azure have publicly stated their use of tape as part of their deep archive service offerings. The main reason is economics, which is supported by tape’s high reliability, long life and future areal density roadmap. I also think the industry is settling on a sensible balance of on-premises and off-site storage where tape has a role in both. So no, cloud is not killing tape, rather it’s an opportunity.

Q: How is an archive different than a backup?

A: Backup and archive are entirely different processes with different objectives. Think of the backup process as a “copy” of your data for recovery purposes. Backups are cycled and updated frequently to account for and protect the latest versions of important data assets.  As for archiving, think of this as a “move” of your fixed data to a new, more cost-effective tier for long-term retention. But if you ask end users, they don’t want their data sitting on a shelf off-line. They want it online, searchable and readily available and that’s what active archive provides.

Q: What is the market opportunity for active archives?

A: The market opportunity is significant due to the volume of archival data, the value of that data and the velocity or speed of access that’s required today. A recent ESG Research survey indicated that less than 40% of corporations have a dedicated archive strategy in place, yet every organization has archival data! The market is ready for modern, leading-edge archiving concepts like active archive.

Q: How is an active archive implemented?

A: There are numerous software and hardware solutions ranging from stand-alone active archive appliances to intelligent data management software that includes active archiving among other capabilities. End users canleveragetheir existing storage systems to implement an active archive strategy. Most active archive solutions allow customers to repurpose existing tape libraries to create an active archive partition that looks to users like another disk volume. When this is combined with open standard LTFS, or Linear Tape File System, the active archive is free from vendor lock-in, and ensures data portability and copy management for long-term archives.

Q: What are some of the advantages of an active archive?

A: An active archive enables users to easily find and utilize archived data, while also removing complexity and operational load on IT administrators. By automating archiving so that data doesn’t get stranded on an inaccessible “shelf” somewhere, the data value increases without tying up the most expensive storage resources. This improves storage performance, lowers total cost of ownership, and reduces risk of non-compliance and data loss.

Q: You mention simplified data storage and ease of use, how is that achieved?

A: Active archive solves complexity by leveraging an intelligent data management layer. Access and management of data is getting more complex so we need modern strategies with intelligent data management techniques that are automated and policy based.  Classifying data upon its creation by its value, and automatically updating performance and capacity requirements over its lifecycle will enable the right data to be in the right place at the right time.

Q: Can active archives be implemented in the cloud?

A: Yes. An active archive can combine onsite, offsite, and cloud environments. Most, if not all, cloud providers are offeringarchival data services including active archiving. Active archive brings the same benefits to public clouds as it does to on-premises solutions.

Q: What’s in store for active archives in the future?

A: As organizations fully embrace digital transformation, they are quickly learning the value of analyzing large amounts of previously dormant archival data, and that makes having quick and affordable access to that data so important. New tools and use cases such as artificial intelligence, machine learning, big data, IoT, and video surveillance for example will drive increased demand for active archives by organizations that need to effectively manage data from terabytes to exabytes across multiple storage tiers.

Finally, I would say that organizations archive their data because they either want to preserve the value of the content or because they have to, such as for compliance — but either way, the magnitude of archival storage requirements will be a major challenge. With the amount of archival data exploding with no end in sight, active archives will play a vital and necessary role in optimizing data storage to reduce costs, but also to ensure archived data is accessible and protected. That’s an attractive value proposition, so the future is bright for active archives.

Originally published in Storage Newsletter, January 14, 2019.

 

Read More

Active Archive Alliance Members Share Top Data Storage and Active Archive Predictions for 2019

Reading Time: 3 minutes

Explosive data growth continues to be a top challenge for today’s organizations and this growth is only going to increase in the future. In fact, according to analyst firm IDC, by 2025 worldwide data will grow 61% to 175 zettabytes, with as much of the data residing in the cloud as in data centers.

New technologies and approaches are continually being created to help address this data storage deluge. Members of the Active Archive Alliance from Fujifilm Recording Media, U.S.A., Inc, Spectra Logic, StrongBox Data and Quantum recently shared their insights into what the future looks like for active archives and data storage in 2019. Here are some of their top predictions:

Artificial Intelligence Creates Demand for Active Archives
The evolution of deep learning, machine learning and artificial intelligence will continue to expand in 2019 across every industry as the digital transformation wave produces an explosion of big data. With these AI tools, organizations will be able to extract more value from their data than ever before giving rise to an insatiable demand for more data, more analytics…more competitive advantage. A dramatic increase in storage and specifically active archive will be required to cost effectively and efficiently provide accessibility to big data at scale.

Flash Will Gain Wide-Scale Adoption, But a Need to Store Everything Will Make Secondary Storage More Important Than Ever
In the coming year we will see wide-scale adoption of flash storage. Organizations of all sizes will include solid-state drive (SSD) for greater performance, energy savings, space efficiency, and reduced management. New technologies like integrated data protection, storage federation/automation, policy-based provisioning, tiered data movement, and public cloud integration will be built on top of this flash foundation.

With the increased adoption of flash, organizations will also face the challenge of how to affordably store the data that is not mission critical, but still has value and therefore cannot be deleted. With the move to flash organizations will utilize a secondary storage tier to affordably manage all the organizations data, and this will happen through intelligent data management software designed to move data to a more affordable tier, without sacrificing access and searchability of the data.

Shift From Managing Your Storage to Managing Your Data
Data, not the underlying physical storage, is what matters. However, traditional storage systems are “big dumb buckets” that provide precious little insight into what data is growing, what applications or users are accessing it, or what is consuming storage performance and why.

Next-generation storage systems are “data-aware,” with real-time analytics built directly into the storage itself, delivering real-time information on data and performance at massive scale providing insight into data and storage.  As organizations better understand their data (how it is being generated, at what pace, by who, for what project) they are more informed as to how to plan and budget for the future growth of their data, and better understand how to move data to different tiers based on customized policies.

Cross-platform Storage Automation Reduces Costs, Increases Productivity
The reality is that there is not a “one-size-fits-all” storage solution that addresses the multiple requirements faced by most organizations.  The result is that large environments typically rely on multiple storage vendors and point solutions to address the different performance and cost profiles needed for their data. The problem is this adds complexity for IT managers, requiring them to do more with static or shrinking operational budgets. This trend is driving a demand in the industry for solutions that provide automation of data and storage resource management across any storage type from any vendor. Such solutions leverage policy engines and management tools that are driven by multiple types of metadata about the files and their business value as they evolve over time. Such automation tools help data managers know what they have, and gives them control of cross-platform data migration, tiering, active archiving, and protection, without interrupting users. This type of metadata-driven automation will be an increasing trend over the next few years, because it provides demonstrable ROI by reducing OPEX and complexity for IT, breaking storage vendor lock-in, while increasing storage utilization efficiency and user productivity.

Rich Media Content Will Grow Exponentially, Across Many Industries
Video now constitutes 50% of all data. Rich media comprises our video surveillance; consumer images, voice and video; medical imagery, IoT, entertainment and social media. Large and unstructured data is often 50 times or larger than the average corporate database. Video is unique, and it is not typically a good fit for traditional backup; it cannot be compressed or deduplicated, it doesn’t work well with replication, snaps or clones, and ingest speed is critical. Rich media is projected to surpass 100 Zetabytes worldwide by 2020. Expect enterprise data services to be increasingly optimized for large or rich media data sets, with infrastructure optimized for ingest processing and the full life cycle management of forms of rich media.

Read More

Leveraging Artificial Intelligence in an Active Archive

Reading Time: < 1 minute

In this Fujifilm Summit presentation, Molly Presley, founder of the Active Archive Alliance, explains how an active archive can provide visibility into your applications and machine generated data with actively assigned metadata no matter what tier it is stored on.

Watch the video to learn more:

 

Read More

Breaking Down Data Silos — Highlights From SC18

Reading Time: 3 minutes

By Kevin Benitez

I had the opportunity to attend SC18 last month in Dallas. Every year the Supercomputing Conference brings together the latest in supercomputing technology and the most brilliant minds in HPC. People from all over the world and different backgrounds converged this year for the 30thSupercomputing Conference.

As you can imagine, some of the demonstrations were absolutelymind-blowing and worth sharing. For starters, power consumption in data centers is becoming more of a challenge as data rates continue to surge. Fortunately, 3M was live on the trade show floor tackling this issue by demonstrating immersion cooling for data centerswhich has the potential to slash energy use and cost by up to 97%. As this technology continues to evolve,we could see huge gains in performance and in reducing environmental impacts.

The race to dominate quantum computing continues! IBM’s 50-Qubit quantum computer made an appearance at this year’s show. What does it mean to have a computer with 50 qubits working perfectly? (Side note, in quantum computing a qubitis the basic unit of quantum information). According to Robert Schoelkopf, a Yale professor, if you had 50 or 100 qubitsyou could “do unfathomable calculations that can’t be replicated on any classical machine, now or ever.” Although the quantum computer churns out enough computational power to rank within the top ten supercomputers in the world,the device can only compute for 100 milliseconds due to a short-lived power supply.

StrongBox Data’s flagship product, StrongLink, was demonstrated on the show floor as a way to store and contain the vast amount of data that research universities and laboratories are producing. StrongLinkis a software solution that simplifies and reduces the cost of managing multi-vendor storage environments. StrongLink provides multi-protocol access across any file system, object storage, tape and cloud in a global namespace. Users maintain a constant view of files regardless of where they arestored, which maximizes their storage environment for performance and cost.

Recently the University of Southampton’s Supercomputer Iridis 5 teamed up with StrongLink to get more value out of its data. Oz Parchment, Director of the University’s iSolutions IT support division, commented in March saying: “One wayStrongLink interested us was its cognitive component, the ability to look at and match up metadata at scale, which gets interesting when you combine that with different data infrastructures. Our set up currently includes large-scale tape stores, large-scale disc stores, some of that being active data, some of that being nearline data, some being effectively offline data. But then, by linking these into the [Iridis] framework, which StrongLink allows us to do, we can connect these various data lakes that we have across the research side of the organization, and begin to create an open data space for our community where people in one discipline can look through data and see what kinds of data are available in other communities.“

Never has HPC been more crucial. As we say here at Fujifilm “Never Stop Transforming Ourselves and the World.”

Read More

Flash, HDDs and Tape Slay Data Challenges

Reading Time: 3 minutes

By Rich Gadomski

At Storage Visions 2018, held in Santa Clara this past October, I had the opportunity to talk about the future outlook for tape as attendees wanted to know how they were going to store all the data that’s being created. The session I participated in was entitled “Epic Battles with Classic Heros – Flash, HDDs and Tape Slay Data Challenges.” As the title suggests, battling exponential data growth takes more than one storage media type to effectively handle the deluge of data that’s being created (now estimated to be 33 ZB in 2018 and growing to 175 ZB by 2025, according to IDC).

As our session moderator, Jean Bozeman from Hurwitz & Associates pointed out in her introduction, a variety of storage workloads create the need for a spectrum of storage technologies. Certainly the need for speed at the top of the storage pyramid is being addressed by performance HDD and increasingly by ever evolving solid state drives.

The need for longer term storage at scale is the domain of capacity HDD and of course, tape. Managing the data deluge is all about having the right data on the right storage medium at the right time. Not everything can or should be stored on expensive high performance flash. You need high capacity optimized media for long term data retention and that’s where HDD and tape come in to play (often in a user friendly active archive environment).

When it comes to the future of capacity in the domain of HDD, current perpendicular magnetic recording technology has reached  ‘super paramagnetic” limitations where increasing areal density to increase capacity is not a viable option. With helium filled HDDs, more platters can fit in the same form factor as air filled HDDs but this has not allowed a significant growth in capacity.  New technology concepts such as Heat Assisted Magnetic Recording (HAMR) and Microwave Assisted Magnetic Recording (MAMR) are on the horizon but market availability has been delayed. There is also the potential of vacuum sealed HDDs with better operating characteristics than helium that could help HAMR and MAMR HDDs get up to 40 – 50 TB at some point in the future.

But fundamentally, increasing capacity of a storage medium and ultimately reducing its cost is best achieved by increasing areal density. This is where magnetic tape technology really shines as today’s modern tape with per cartridge capacities already as high as 20 TB having very low areal densities compared to HDD.

Therefore, tape has a long runway before facing areal density limits and as a result, future tape roadmaps have the potential to achieve up to 220 TB on a standard form factor cartridge using Barium Ferrite (BaFe) magnetic particles and up to 400 TB using next generation Strontium Ferrite (SrFe). At the same time, both BaFe and SrFe can maintain magnetic signal strength integrity for at least 30 years making them ideal not only for high capacity but for cost effective long term data retention as well.

“No wonder the cloud guys are all using tape now,” exclaimed an attendee in the audience during the Q&A. They certainly also use a lot of flash and a lot of disk too. It is an epic battle and it takes a spectrum of storage technologies to slay the data challenges.

Read More

YES Network Says “Yes” to Migrating to a Virtual Environment

Reading Time: 2 minutes

Since its launch in March 2002, the YES Network has been the number 1 rated regional sports network in the USA. From its inception, YES has striven for the highest-quality picture and sound and to be at the leading-edge of broadcast technology. To this end, YES was constructed as an all-digital facility.

To manage its growing library, the network launched a digital archive project. Initially, the plan was to find a way to convert HD content into a file format that could be stored in a system so that producers and directors could easily find and retrieve selected plays to be integrated into classic game and other shoulder programmes. Avid had provided the YES editing systems from the outset, and the original five Avid editing systems were connected to an Avid Omega JBOD array for storage.

This paper provides a deep dive into the pros and cons of local, cloud, solid-state and linear tape-open storage solutions. It opens with YES Network Director of Engineering and Technology John McKenna’s account of the YES Network’s digital transformation, and is followed by YES Network Engineering Department Project Manager of Broadcast Systems Jason Marshall’s summary of modular to virtual technology migration. This paper details ratios on high-performance broadcast data systems, as well as power consumption and solution trade-offs. This paper aims to gain the reader’s confidence in virtualising a media asset system as well as converting linear systems to packet-based media technologies including transcoding, Active Archive and future SMPTE 2110 solutions.

Read the full paper here: Migrating to a Virtual Environment

Read More

Used / Recertified / Reconditioned Tape – Is It Worth the Risk?

Reading Time: 3 minutes

By Ken Kajikawa 

“If it’s too good to be true, it probably is.” I’m not sure who first coined that old adage, but it certainly applies to used data tape regardless of whether it’s called “recertified” or “reconditioned.” Let’s review some of the facts.

Recertified? Is there such a thing as legitimately recertified tape? The answer is no and here’s why. The equipment and procedures to fully certify and control the quality of tape performance are available only to licensed manufacturers. No one else has it–so used tape can’t be “re certified” by third parties.  Fujifilm does not recertify tape.  Fujifilm only provides new/unused product to the marketplace.

Reconditioned? Okay, how about reconditioned tape? A tape cartridge cannot be reconditioned either. Once a tape is scratched, creased,  edge damaged, degraded–it can’t be restored to its original factory-new condition. Additionally, a recertifier’s equipment and practices such as data erasure could damage the tape.

Data Erasure? Do the so-called recertifiers actually erase the data from the previous owner? Not exactly. Typically a table of contents overwrite is all that is performed, if anything. To actually overwrite the data on a common LTO data tape would take hours and degaussing is not a quick fix as it would destroy the servo tracks and render the tape useless. A few years back, Fujifilm had acquired 50 “recertified” LTO data tapes from recertified tape resellers. It was determined through expert analysis that 48 out of 50 tapes still contained original user data including highly confidential customer data.

So my advice here is: don’t be a buyer or seller of used tape!

Still considering used tape? Read on for more details about some of the potential hazards! 

  • Storage Environment: Tape must be properly stored and cared for in controlled environments (preferably cool, dry conditions for archive storage 16°C to 25°C and 20% to 50% relative humidity). Used tapes may have been stored for extended periods under poor environmental controls. Tape media degradation and damage are all possibilities that will not be readily apparent to the end user.
  • Care and Handling: Tape must be properly handled. Poor transportation and handling practices, (dropped tapes) could result in internal damage, poor tape pack alignment, and/or tape edge damage.
  • Proper Tape Operating Environment During Prior Usage: Airborne contaminants or dust can get wound into the tape-pack and damage the tape. In addition, excessive heat at the tape head interface can damage tape. This can be a result of drives that were running above maximum operating temperature specification due to integration of inside units lacking sufficient ventilation. Or, a combination of ambient room temperature being too hot and the drive being inside a unit or rack with marginal allowance for thermal transfer (not enough cooling capability under higher ambient temperature conditions).
  • Drive Maintenance: A previous user’s improperly maintained or malfunctioning tape drive could have damaged the tape, or the mechanical functionality of the cartridge.
  • Risk of Damage to Existing Drives and Tapes: Many tapes share the same drives in a typical usage environment. Debris left behind by used tape that is scratched or otherwise physically damaged will certainly contaminate good tapes that follow on those same drives.

Fujifilm always recommends against used/recertified media because the customer can never be assured of the quality, performance, and reliability in several key areas as discussed above.

If you are taking the time and resources to back up your data, why risk your data to a cartridge with unknown history?

Fujifilm high capacity data cartridges are consistently manufactured to the highest specifications and standards and fully supported and backed by a Lifetime Warranty against manufactures defects.

 

Read More

The Cost Viability of Tape for Data Protection and Archive

Reading Time: < 1 minute

The most efficient data protection utilizes proper archiving, and with the data growth rate almost doubling, tape storage is growing from an archiving standpoint. In this Fujifilm Summit video, Dr. James Cates, SVP of Archive Development at Oracle, discusses the advantages of tape for archiving. Watch it here:

 

Read More

How Does Google Do Backups?

Reading Time: < 1 minute

In this Fujifilm Summit video, Raymond Blum, Staff Site Reliability Engineer at Google, explains how Google handles its backups and the importance of diversity when it comes to storage. Watch it here:

Read More

LET’S DISCUSS YOUR NEEDS

We can help you reduce cost, decrease vendor lock-in, and increase productivity of storage staff while ensuring accessibility and longevity of data.

Contact Us >