FUJIFILM INSIGHTS BLOG

Data Storage

Active Archive Alliance Members Share Top Data Storage and Active Archive Predictions for 2019

Reading Time: 3 minutes

Explosive data growth continues to be a top challenge for today’s organizations and this growth is only going to increase in the future. In fact, according to analyst firm IDC, by 2025 worldwide data will grow 61% to 175 zettabytes, with as much of the data residing in the cloud as in data centers.

New technologies and approaches are continually being created to help address this data storage deluge. Members of the Active Archive Alliance from Fujifilm Recording Media, U.S.A., Inc, Spectra Logic, StrongBox Data and Quantum recently shared their insights into what the future looks like for active archives and data storage in 2019. Here are some of their top predictions:

Artificial Intelligence Creates Demand for Active Archives
The evolution of deep learning, machine learning and artificial intelligence will continue to expand in 2019 across every industry as the digital transformation wave produces an explosion of big data. With these AI tools, organizations will be able to extract more value from their data than ever before giving rise to an insatiable demand for more data, more analytics…more competitive advantage. A dramatic increase in storage and specifically active archive will be required to cost effectively and efficiently provide accessibility to big data at scale.

Flash Will Gain Wide-Scale Adoption, But a Need to Store Everything Will Make Secondary Storage More Important Than Ever
In the coming year we will see wide-scale adoption of flash storage. Organizations of all sizes will include solid-state drive (SSD) for greater performance, energy savings, space efficiency, and reduced management. New technologies like integrated data protection, storage federation/automation, policy-based provisioning, tiered data movement, and public cloud integration will be built on top of this flash foundation.

With the increased adoption of flash, organizations will also face the challenge of how to affordably store the data that is not mission critical, but still has value and therefore cannot be deleted. With the move to flash organizations will utilize a secondary storage tier to affordably manage all the organizations data, and this will happen through intelligent data management software designed to move data to a more affordable tier, without sacrificing access and searchability of the data.

Shift From Managing Your Storage to Managing Your Data
Data, not the underlying physical storage, is what matters. However, traditional storage systems are “big dumb buckets” that provide precious little insight into what data is growing, what applications or users are accessing it, or what is consuming storage performance and why.

Next-generation storage systems are “data-aware,” with real-time analytics built directly into the storage itself, delivering real-time information on data and performance at massive scale providing insight into data and storage.  As organizations better understand their data (how it is being generated, at what pace, by who, for what project) they are more informed as to how to plan and budget for the future growth of their data, and better understand how to move data to different tiers based on customized policies.

Cross-platform Storage Automation Reduces Costs, Increases Productivity
The reality is that there is not a “one-size-fits-all” storage solution that addresses the multiple requirements faced by most organizations.  The result is that large environments typically rely on multiple storage vendors and point solutions to address the different performance and cost profiles needed for their data. The problem is this adds complexity for IT managers, requiring them to do more with static or shrinking operational budgets. This trend is driving a demand in the industry for solutions that provide automation of data and storage resource management across any storage type from any vendor. Such solutions leverage policy engines and management tools that are driven by multiple types of metadata about the files and their business value as they evolve over time. Such automation tools help data managers know what they have, and gives them control of cross-platform data migration, tiering, active archiving, and protection, without interrupting users. This type of metadata-driven automation will be an increasing trend over the next few years, because it provides demonstrable ROI by reducing OPEX and complexity for IT, breaking storage vendor lock-in, while increasing storage utilization efficiency and user productivity.

Rich Media Content Will Grow Exponentially, Across Many Industries
Video now constitutes 50% of all data. Rich media comprises our video surveillance; consumer images, voice and video; medical imagery, IoT, entertainment and social media. Large and unstructured data is often 50 times or larger than the average corporate database. Video is unique, and it is not typically a good fit for traditional backup; it cannot be compressed or deduplicated, it doesn’t work well with replication, snaps or clones, and ingest speed is critical. Rich media is projected to surpass 100 Zetabytes worldwide by 2020. Expect enterprise data services to be increasingly optimized for large or rich media data sets, with infrastructure optimized for ingest processing and the full life cycle management of forms of rich media.

Read More

Leveraging Artificial Intelligence in an Active Archive

Reading Time: < 1 minute

In this Fujifilm Summit presentation, Molly Presley, founder of the Active Archive Alliance, explains how an active archive can provide visibility into your applications and machine generated data with actively assigned metadata no matter what tier it is stored on.

Watch the video to learn more:

 

Read More

Breaking Down Data Silos — Highlights From SC18

Reading Time: 3 minutes

By Kevin Benitez

I had the opportunity to attend SC18 last month in Dallas. Every year the Supercomputing Conference brings together the latest in supercomputing technology and the most brilliant minds in HPC. People from all over the world and different backgrounds converged this year for the 30thSupercomputing Conference.

As you can imagine, some of the demonstrations were absolutelymind-blowing and worth sharing. For starters, power consumption in data centers is becoming more of a challenge as data rates continue to surge. Fortunately, 3M was live on the trade show floor tackling this issue by demonstrating immersion cooling for data centerswhich has the potential to slash energy use and cost by up to 97%. As this technology continues to evolve,we could see huge gains in performance and in reducing environmental impacts.

The race to dominate quantum computing continues! IBM’s 50-Qubit quantum computer made an appearance at this year’s show. What does it mean to have a computer with 50 qubits working perfectly? (Side note, in quantum computing a qubitis the basic unit of quantum information). According to Robert Schoelkopf, a Yale professor, if you had 50 or 100 qubitsyou could “do unfathomable calculations that can’t be replicated on any classical machine, now or ever.” Although the quantum computer churns out enough computational power to rank within the top ten supercomputers in the world,the device can only compute for 100 milliseconds due to a short-lived power supply.

StrongBox Data’s flagship product, StrongLink, was demonstrated on the show floor as a way to store and contain the vast amount of data that research universities and laboratories are producing. StrongLinkis a software solution that simplifies and reduces the cost of managing multi-vendor storage environments. StrongLink provides multi-protocol access across any file system, object storage, tape and cloud in a global namespace. Users maintain a constant view of files regardless of where they arestored, which maximizes their storage environment for performance and cost.

Recently the University of Southampton’s Supercomputer Iridis 5 teamed up with StrongLink to get more value out of its data. Oz Parchment, Director of the University’s iSolutions IT support division, commented in March saying: “One wayStrongLink interested us was its cognitive component, the ability to look at and match up metadata at scale, which gets interesting when you combine that with different data infrastructures. Our set up currently includes large-scale tape stores, large-scale disc stores, some of that being active data, some of that being nearline data, some being effectively offline data. But then, by linking these into the [Iridis] framework, which StrongLink allows us to do, we can connect these various data lakes that we have across the research side of the organization, and begin to create an open data space for our community where people in one discipline can look through data and see what kinds of data are available in other communities.“

Never has HPC been more crucial. As we say here at Fujifilm “Never Stop Transforming Ourselves and the World.”

Read More

Flash, HDDs and Tape Slay Data Challenges

Reading Time: 3 minutes

By Rich Gadomski

At Storage Visions 2018, held in Santa Clara this past October, I had the opportunity to talk about the future outlook for tape as attendees wanted to know how they were going to store all the data that’s being created. The session I participated in was entitled “Epic Battles with Classic Heros – Flash, HDDs and Tape Slay Data Challenges.” As the title suggests, battling exponential data growth takes more than one storage media type to effectively handle the deluge of data that’s being created (now estimated to be 33 ZB in 2018 and growing to 175 ZB by 2025, according to IDC).

As our session moderator, Jean Bozeman from Hurwitz & Associates pointed out in her introduction, a variety of storage workloads create the need for a spectrum of storage technologies. Certainly the need for speed at the top of the storage pyramid is being addressed by performance HDD and increasingly by ever evolving solid state drives.

The need for longer term storage at scale is the domain of capacity HDD and of course, tape. Managing the data deluge is all about having the right data on the right storage medium at the right time. Not everything can or should be stored on expensive high performance flash. You need high capacity optimized media for long term data retention and that’s where HDD and tape come in to play (often in a user friendly active archive environment).

When it comes to the future of capacity in the domain of HDD, current perpendicular magnetic recording technology has reached  ‘super paramagnetic” limitations where increasing areal density to increase capacity is not a viable option. With helium filled HDDs, more platters can fit in the same form factor as air filled HDDs but this has not allowed a significant growth in capacity.  New technology concepts such as Heat Assisted Magnetic Recording (HAMR) and Microwave Assisted Magnetic Recording (MAMR) are on the horizon but market availability has been delayed. There is also the potential of vacuum sealed HDDs with better operating characteristics than helium that could help HAMR and MAMR HDDs get up to 40 – 50 TB at some point in the future.

But fundamentally, increasing capacity of a storage medium and ultimately reducing its cost is best achieved by increasing areal density. This is where magnetic tape technology really shines as today’s modern tape with per cartridge capacities already as high as 20 TB having very low areal densities compared to HDD.

Therefore, tape has a long runway before facing areal density limits and as a result, future tape roadmaps have the potential to achieve up to 220 TB on a standard form factor cartridge using Barium Ferrite (BaFe) magnetic particles and up to 400 TB using next generation Strontium Ferrite (SrFe). At the same time, both BaFe and SrFe can maintain magnetic signal strength integrity for at least 30 years making them ideal not only for high capacity but for cost effective long term data retention as well.

“No wonder the cloud guys are all using tape now,” exclaimed an attendee in the audience during the Q&A. They certainly also use a lot of flash and a lot of disk too. It is an epic battle and it takes a spectrum of storage technologies to slay the data challenges.

Read More

YES Network Says “Yes” to Migrating to a Virtual Environment

Reading Time: 2 minutes

Since its launch in March 2002, the YES Network has been the number 1 rated regional sports network in the USA. From its inception, YES has striven for the highest-quality picture and sound and to be at the leading-edge of broadcast technology. To this end, YES was constructed as an all-digital facility.

To manage its growing library, the network launched a digital archive project. Initially, the plan was to find a way to convert HD content into a file format that could be stored in a system so that producers and directors could easily find and retrieve selected plays to be integrated into classic game and other shoulder programmes. Avid had provided the YES editing systems from the outset, and the original five Avid editing systems were connected to an Avid Omega JBOD array for storage.

This paper provides a deep dive into the pros and cons of local, cloud, solid-state and linear tape-open storage solutions. It opens with YES Network Director of Engineering and Technology John McKenna’s account of the YES Network’s digital transformation, and is followed by YES Network Engineering Department Project Manager of Broadcast Systems Jason Marshall’s summary of modular to virtual technology migration. This paper details ratios on high-performance broadcast data systems, as well as power consumption and solution trade-offs. This paper aims to gain the reader’s confidence in virtualising a media asset system as well as converting linear systems to packet-based media technologies including transcoding, Active Archive and future SMPTE 2110 solutions.

Read the full paper here: Migrating to a Virtual Environment

Read More

Used / Recertified / Reconditioned Tape – Is It Worth the Risk?

Reading Time: 3 minutes

By Ken Kajikawa 

“If it’s too good to be true, it probably is.” I’m not sure who first coined that old adage, but it certainly applies to used data tape regardless of whether it’s called “recertified” or “reconditioned.” Let’s review some of the facts.

Recertified? Is there such a thing as legitimately recertified tape? The answer is no and here’s why. The equipment and procedures to fully certify and control the quality of tape performance are available only to licensed manufacturers. No one else has it–so used tape can’t be “re certified” by third parties.  Fujifilm does not recertify tape.  Fujifilm only provides new/unused product to the marketplace.

Reconditioned? Okay, how about reconditioned tape? A tape cartridge cannot be reconditioned either. Once a tape is scratched, creased,  edge damaged, degraded–it can’t be restored to its original factory-new condition. Additionally, a recertifier’s equipment and practices such as data erasure could damage the tape.

Data Erasure? Do the so-called recertifiers actually erase the data from the previous owner? Not exactly. Typically a table of contents overwrite is all that is performed, if anything. To actually overwrite the data on a common LTO data tape would take hours and degaussing is not a quick fix as it would destroy the servo tracks and render the tape useless. A few years back, Fujifilm had acquired 50 “recertified” LTO data tapes from recertified tape resellers. It was determined through expert analysis that 48 out of 50 tapes still contained original user data including highly confidential customer data.

So my advice here is: don’t be a buyer or seller of used tape!

Still considering used tape? Read on for more details about some of the potential hazards! 

  • Storage Environment: Tape must be properly stored and cared for in controlled environments (preferably cool, dry conditions for archive storage 16°C to 25°C and 20% to 50% relative humidity). Used tapes may have been stored for extended periods under poor environmental controls. Tape media degradation and damage are all possibilities that will not be readily apparent to the end user.
  • Care and Handling: Tape must be properly handled. Poor transportation and handling practices, (dropped tapes) could result in internal damage, poor tape pack alignment, and/or tape edge damage.
  • Proper Tape Operating Environment During Prior Usage: Airborne contaminants or dust can get wound into the tape-pack and damage the tape. In addition, excessive heat at the tape head interface can damage tape. This can be a result of drives that were running above maximum operating temperature specification due to integration of inside units lacking sufficient ventilation. Or, a combination of ambient room temperature being too hot and the drive being inside a unit or rack with marginal allowance for thermal transfer (not enough cooling capability under higher ambient temperature conditions).
  • Drive Maintenance: A previous user’s improperly maintained or malfunctioning tape drive could have damaged the tape, or the mechanical functionality of the cartridge.
  • Risk of Damage to Existing Drives and Tapes: Many tapes share the same drives in a typical usage environment. Debris left behind by used tape that is scratched or otherwise physically damaged will certainly contaminate good tapes that follow on those same drives.

Fujifilm always recommends against used/recertified media because the customer can never be assured of the quality, performance, and reliability in several key areas as discussed above.

If you are taking the time and resources to back up your data, why risk your data to a cartridge with unknown history?

Fujifilm high capacity data cartridges are consistently manufactured to the highest specifications and standards and fully supported and backed by a Lifetime Warranty against manufactures defects.

 

Read More

The Cost Viability of Tape for Data Protection and Archive

Reading Time: < 1 minute

The most efficient data protection utilizes proper archiving, and with the data growth rate almost doubling, tape storage is growing from an archiving standpoint. In this Fujifilm Summit video, Dr. James Cates, SVP of Archive Development at Oracle, discusses the advantages of tape for archiving. Watch it here:

 

Read More

How Does Google Do Backups?

Reading Time: < 1 minute

In this Fujifilm Summit video, Raymond Blum, Staff Site Reliability Engineer at Google, explains how Google handles its backups and the importance of diversity when it comes to storage. Watch it here:

Read More

How Do You Get Renewables to Power Data Centers?

Reading Time: < 1 minute

By diversifying your renewable energy mix, you can achieve energy efficiency gains even with data centers which typically carry large power loads.  In this Fujifilm video, Craig Lewis, Executive Director of Clean Coalition talks about how tape storage allows us to do more work with more data storage using a lot less energy.  Watch it here:

 

 

Read More

How to Store a Zettabyte

Reading Time: < 1 minute

According to Aaron Ogus, partner development manager for Microsoft Azure Storage, storing a zettabyte of storage will be financially feasible in 2020. Data growth will always exceed expectations, and tape has a more credible road map and one that is easier to get to with not as much investment. Learn more in this video blog:

Read More

LET’S DISCUSS YOUR NEEDS

We can help you reduce cost, decrease vendor lock-in, and increase productivity of storage staff while ensuring accessibility and longevity of data.

Contact Us >