Outlook 2013

Big data in 2013: Not just for big projects

Bill Cull

No organization in the world has more data than the U.S. government. From real-time weather information to the tracking of swipes of Common Access Cards at federal buildings, the government has vast amounts of data at its disposal that can be collected, sorted and analyzed for purposes that include economic development and cybersecurity.

The presidential memorandum “Building a 21st Century Digital Government” released in May calls for a larger movement to “unlock the power of government data to spur innovation across our nation and improve the quality of services for the American people.” However, with tightening budgets and an uncertain fiscal future, agencies might be reluctant to hop on the big-data bandwagon, and understandably so.

With the start of a new year, several agencies will face increasing pressure to incorporate big data into their larger IT strategies. Investing in big data can sound like a tall order, but the trick is to dip your toe in the water before jumping in headfirst.

There is no specific formula for tackling big-data challenges in government. Each agency is unique and has its own set of challenges to solve. Rather than take on big data for big data’s sake, agencies should be strategic in their approach.

As the TechAmerica Foundation Big Data Commission points out in its recent report “Demystifying Big Data: A Practical Guide to Transforming the Business of Government,” agencies should identify a few key business or mission requirements that big data can address but understand the “art of the possible” before making any substantial investments. There are often several opportunities to deploy small-scale big-data projects and build additional larger-scale investments after the success of those pilot programs.

Agencies should ask themselves a few key questions as part of the cost/benefit analysis process: Will this big-data project help achieve one of our critical business objectives? Do we have the resources and capabilities needed to take on a project of this scale? Do we have access to the datasets necessary to make the most of this project? Can it be completed in a timely manner?

Although big-data projects will eventually pay for themselves by saving time and resources, federal, state and local CIOs should also consider technologies that reap benefits beyond their agencies’ bottom line.

Big data is not a fad technology. It’s transformational.

For example, one of the most significant challenges the National Security Agency faces is the ability to provide vast amounts of real-time data gathered from intelligence agencies, military branches and other sources to authorized users based on different access privileges.

Since July 2009, NSA has used the open-source Apache Hadoop platform for a massive, nationwide system for sharing and analyzing data. Hadoop allows agencies to tap huge, distributed data sources and is leading to the adoption of advanced big-data solutions.

NSA and other agencies, including the CIA, are now turning to data warehousing, mining and visualization tools that integrate with Hadoop. Those solutions manage and analyze data for the entire Hadoop environment. By integrating Hadoop with other big-data technologies, agencies are better equipped to make critical decisions in real time.

Big data is not a fad technology. It’s transformational. Instead of waiting to see how big data settles into the federal IT community, agencies should embark on small big-data projects to test the technology themselves.

As the TechAmerica commission notes in its report, agencies should not think of big data as an IT solution that solves reporting and analytical problems but rather as a strategic asset that can help achieve missions. There are several efficiencies to be gained as budgets continue to shrink, and investing in big-data projects will serve as an added layer to executing an agency’s strategic goals.

About the Author

Bill Cull is vice president of the public sector at Splunk.

FCW in Print

In the latest issue: Looking back on three decades of big stories in federal IT.


  • Anne Rung -- Commerce Department Photo

    Exit interview with Anne Rung

    The government's departing top acquisition official said she leaves behind a solid foundation on which to build more effective and efficient federal IT.

  • Charles Phalen

    Administration appoints first head of NBIB

    The National Background Investigations Bureau announced the appointment of its first director as the agency prepares to take over processing government background checks.

  • Sen. James Lankford (R-Okla.)

    Senator: Rigid hiring process pushes millennials from federal work

    Sen. James Lankford (R-Okla.) said agencies are missing out on younger workers because of the government's rigidity, particularly its protracted hiring process.

  • FCW @ 30 GPS

    FCW @ 30

    Since 1987, FCW has covered it all -- the major contracts, the disruptive technologies, the picayune scandals and the many, many people who make federal IT function. Here's a look back at six of the most significant stories.

  • Shutterstock image.

    A 'minibus' appropriations package could be in the cards

    A short-term funding bill is expected by Sept. 30 to keep the federal government operating through early December, but after that the options get more complicated.

  • Defense Secretary Ash Carter speaks at the TechCrunch Disrupt conference in San Francisco

    DOD launches new tech hub in Austin

    The DOD is opening a new Defense Innovation Unit Experimental office in Austin, Texas, while Congress debates legislation that could defund DIUx.

Reader comments

Please post your comments here. Comments are moderated, so they may not appear immediately after submitting. We will not post comments that we consider abusive or off-topic.

Please type the letters/numbers you see above

More from 1105 Public Sector Media Group