Entries Tagged 'Archiving' ↓

Bridging the “storage-information” gap

When Nick first unveiled this blog last month he rightly noted ‘storage’ as one of the many categories that falls into a capacious bucket we term ‘information management.’ With this in mind he reminded me that it would be appropriate for the 451 Group’s storage research team to contribute to the debate, so here it is!

For the uninitiated, storage can appear to be either a bit of a black hole, or just a lot of spinning rust, so I’m not going to start with a storage 101 (although if you have a 451 password you can peruse our recent research here). Suffice to say that storage is just one element of the information management infrastructure, but its role is certainly evolving.

Storage systems and associated software traditionally have provided applications and users with the data they need, when they need it, along with the required levels of protection. Clearly, storage has had to become smarter (not to mention cheaper) to deal with issues like data growth; technologies such as data deduplication help firms grapple with the “too much” part of information management. But up until now the lines of demarcation between “storage” (and data management) and “information” management have been fairly clear. Even though larger “portfolio” vendors such as EMC and IBM have feet in both camps, the reality is that such products and services are organized, managed and sold separately.

That said, there’s no doubt these worlds are coming together. The issues we as analysts are grappling with relate to where and why this taking place, how it manifests itself, the role of technology, and the impact of this on vendor, investor and end-user strategies. At the very least there is a demand for technologies that help organizations bridge the gap – and the juxtaposition – between the fairly closeted, back-end storage “silo” and the more, shall we say, liberated, front-end interface where information meets its consumers.

Here, a number of competing forces are challenging, even forcing, organizations to become smarter about understanding what “information” they have in their storage infrastructure; data retention vs data disposition, regulated vs unregulated data and public vs private data being just three. Armed with such intelligence, firms can, in theory, make better decisions about how (and how long) data is stored, protected, retained and made available to support changing business requirements.

“Hang on a minute,” I hear you cry. “Isn’t this what Information Lifecycle Management (ILM) was supposed to be about?” Well, yes, I’m afraid it was. And one thing that covering the storage industry for almost a decade has told me is that it moves at a glacial pace. In the case of ILM, the iceberg has probably lapped it by now. The hows and whys of ILM’s failure to capture the imagination of the industry is probably best left for another day, but I believe that at least one aim of ILM – helping organizations better understand their data so it can better support the business — still makes perfect sense.

What we are now seeing is the emergence of some real business drivers that are compelling a variety of stakeholders – from CIOs to General Counsel — to take an active interest in better understanding their data. This, in turn, is driving industry consolidation as larger vendors in particular move to fill out their product portfolios; the latest example of this is the news of HP’s acquisition of Australia-based records management specialist Tower Software. Over the next few weeks I’ll be exploring in more detail three areas where we think this storage-information gap is being bridged; in eDiscovery, archiving and security. Stay tuned for our deeper thoughts and perspectives in this fast-moving space.

Legal – IT backwater no more

I’ve been attending LegalTech here in New York for the past few years, but this year things seemed to be different.Firstly, and most noticeably, every inch of available space at the New York Hilton on 6th Avenue was taken, spread across three floors. The corridors, which in less busy shows simply lead you to rooms, were lined with stands as were the exhibition spaces. It reminded me of the annual SIFMA Technology Management conference, which is a bit of a zoo and in the same location. But unlike the financial services industry, the legal industry and general counsel offices of corporations haven’t traditionally been seen as major buyers of IT, let alone cutting edge stuff.But there’s nothing like regulations to fuel a surge in the market. The changes the Federal Rules of Civil Procedure (FRCP), which took effect in December 2006 and mandated that all electronic records were discoverable and that parties needed to be ready within 120 days of the start of a lawsuit to discuss their eDiscovery terms. This made eDiscovery a very hot market in 2007 (and helped Stratify to a nice valuation when it was bought by Iron Mountain in July 2007 for $158m).

But one of the messages I picked up pretty loud and clear is that law firms and legal departments have their eye on a much bigger problem, currently being done largely manually, but ripe for automation: document review. Figures of a $15bn market for document review now and a bill of $40bn by 2011 for overall review expense raised more than a few eyebrows among some prospective customers of document review vendors (many of which are also eDiscovery vendors, a market pegged at about $3bn). Jay Brudz, senior counsel, Legal Technology at GE, put it bluntly, “you know how many freaking lightbulbs we’ve gotta sell to pay for that?,” before making it clear that he had no intention in paying what vendors are asking.

The other point of tension I’m picking up is the one between intelligent archiving and search – the battle of ideas between those that think it’s better to do all the tagging at archive time and do some culling at that point (to avoid storing dupes and garbage) and those that think you should store everything and develop smarter search engines.

It’s clear – admittedly without any empirical evidence to hand – that protagonists in this space, be they general counsel departments, outside law firms or the vendors feel the rate is increasing so fast, their ability to cull the data at archiving time to make it more easily discoverable later can’t keep pace. There’s clearly somethig to that, given how rapidly talk has moved from gigabytes, to terabytes to petabytes to something an IBMer who handles data governance strategy for the company told me his clients call Goog-bytes – a generic term to mean so much data they can’t get their heads around it. After all, at this rate it won’t be that long before we talk of yottabytes in this arena, and what comes after that?

Search and archiving is something we at 451 Group have spent a lot of time on already and that is sure to continue in 2008.