Welcome!

@ThingsExpo Authors: Zakia Bouachraoui, Yeshim Deniz, Liz McMillan, Elizabeth White, Pat Romanski

Related Topics: @DXWorldExpo, @CloudExpo, @ThingsExpo

@DXWorldExpo: Article

The Big Data Divide | @CloudExpo #BigData #IoT #IIoT #M2M #IternetOfThings

Business users must be able to quickly and easily access the data they need to make more informed business decisions

Within enterprises - and even small businesses - there are two distinct groups with differing goals, needs and objectives when it comes to business intelligence and analytics strategies. Business users are tasked with analyzing Big Data to help their companies make timely and more meaningful decisions, and as such, require immediate access to a wide variety of sources, including structured, semi-structured, unstructured and streaming data. On the other side of the house, IT professionals are tasked with storing and securing massive data stores, as well as ensuring regulatory compliance of corporate information. Therefore, they prefer to make information available on an as-needed basis, rather than creating an environment of open access. So the tug-of-war game begins.

In the self-service analytics world we live in, business users must be able to quickly and easily access the data they need to make more informed business decisions. The problem is that the data that holds the most analytical value is often locked away in data stores protected by IT or housed within multi-structured documents, such as PDF files, POs and third-party reports that require IT intervention. Consequently, business users have no choice but to rely on IT to get the data they need for analysis, and the lag time between request and delivery often results in decisions being made based on old or incomplete data.

Big Data access and real-time analysis are essential for reliable company operations. With business users and IT departments at odds, many organizations are left wondering how to bridge the gap between the ease of use and agility that business users demand and the automation, scalability and governance required by IT.

The changing relationship between IT and business users
The rise of Big Data has changed how IT and business users interact and work together. Business intelligence used to be a centralized, IT-led initiative, but business users quickly became frustrated because they couldn't access the information they needed fast enough.

To remain competitive, organizations must enable business users and citizen analysts to quickly and easily access and prep data for analysis, visualization or for other business processes. Data users can no longer afford to wait hours, days or weeks for their requests to be fulfilled.

Many organizations now realize that an IT-managed, centralized approach to business intelligence and analytics does not provide the agility necessary for business users. But providing business users with self-service analytics capabilities can introduce serious governance risks - especially given the fact that half of the data business users access for analysis comes from sources that aren't managed by IT.

While most organizations have well-defined strategies for governing data that lives in managed systems like enterprise applications or data warehouses, analysts often need to pull data from non-managed sources as well, like CSV or text extracts from transactional systems, personal spreadsheets, third-party reports or semi-structured content. Without proper governance in place, this can create big headaches for IT. IT must be able to properly track and secure these non-managed data sources and ensure regulatory compliance.

What's to be done? Business users must be self-sufficient, but in a way that brings management and governance to the process to satisfy IT requirements. This is the role self-service data preparation solutions play today, and it is why this technology has become a key component in Big Data and analytics strategies.

Bridging the gap
Self-service data preparation tools are rapidly being recognized as a necessary component of any data discovery or advanced analytics implementation. More advanced, bi-modal data preparation solutions empower business users to retrieve, blend and prepare various data sources, but do so within a central content repository that allows for secure storage, management and access control to all source content as well as any reusable models for data extraction and prep routines.

Further easing IT's anxiety, self-service data prep solutions offer the following governance capabilities:

  • Data Retention - Documents version control for consistency. Additionally, to meet regulatory and business requirements, relevant source data and documents should be archived.
  • Data Masking - Prevents data breaches, especially those caused by internal employees, by hiding or obfuscating original data with random characters. The data is still usable for analytics, while the underlying data is visible only to authorized users.
  • Data Lineage - Drills down into any source or document for data reconciliation or auditing.
  • Role-based Access - Segments prepared data sets based on user roles to ensure the right subset of data is delivered to authorized users.
  • Auditing - Tracks information access for complete audit logging and reporting.
  • Data curation - Share frequently used data sources or automated data preparation routines with other analysts allowing for consistency in the data used to make important business decisions.

Telling the whole story - the right way
To yield maximum return on investment from business intelligence and analytics investments, organizations must empower business users with the speed and agility they require to access all the data they need to tell the whole story. But they must do so in a way that satisfies IT's need for security, governance and compliance. Fortunately, self-service data prep technology has turned these two previously mutually exclusive requirements into a reality for organizations of all sizes.

It's estimated that only 12 percent of data1 is being used today to make decisions. Self-service data prep enables businesses to tap into the remaining 88 percent of unused data to make timely, more meaningful decisions. When you have good - and complete - data, you can then begin to predict what is likely to happen and respond faster when it does.

1 The Forrester Wave™: Big Data Hadoop Solutions, Q1 2014, http://ibm.biz/Bd4DJC

More Stories By Dan Potter

Dan Potter is chief marketing officer at Datawatch Corporation. Prior to Datawatch, he held senior roles at IBM, Oracle, Progress and Attunity where he was responsible for identifying and launching solutions across a variety of markets including data analytics, cloud computing, real-time data streaming, federated data and e-commerce.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


IoT & Smart Cities Stories
The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-c...
Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by ...
The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...
René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions. René is a member of the Society of Women Engineers (SWE) and a m...
Poor data quality and analytics drive down business value. In fact, Gartner estimated that the average financial impact of poor data quality on organizations is $9.7 million per year. But bad data is much more than a cost center. By eroding trust in information, analytics and the business decisions based on these, it is a serious impediment to digital transformation.
Digital Transformation: Preparing Cloud & IoT Security for the Age of Artificial Intelligence. As automation and artificial intelligence (AI) power solution development and delivery, many businesses need to build backend cloud capabilities. Well-poised organizations, marketing smart devices with AI and BlockChain capabilities prepare to refine compliance and regulatory capabilities in 2018. Volumes of health, financial, technical and privacy data, along with tightening compliance requirements by...
Predicting the future has never been more challenging - not because of the lack of data but because of the flood of ungoverned and risk laden information. Microsoft states that 2.5 exabytes of data are created every day. Expectations and reliance on data are being pushed to the limits, as demands around hybrid options continue to grow.
Digital Transformation and Disruption, Amazon Style - What You Can Learn. Chris Kocher is a co-founder of Grey Heron, a management and strategic marketing consulting firm. He has 25+ years in both strategic and hands-on operating experience helping executives and investors build revenues and shareholder value. He has consulted with over 130 companies on innovating with new business models, product strategies and monetization. Chris has held management positions at HP and Symantec in addition to ...
Enterprises have taken advantage of IoT to achieve important revenue and cost advantages. What is less apparent is how incumbent enterprises operating at scale have, following success with IoT, built analytic, operations management and software development capabilities - ranging from autonomous vehicles to manageable robotics installations. They have embraced these capabilities as if they were Silicon Valley startups.
As IoT continues to increase momentum, so does the associated risk. Secure Device Lifecycle Management (DLM) is ranked as one of the most important technology areas of IoT. Driving this trend is the realization that secure support for IoT devices provides companies the ability to deliver high-quality, reliable, secure offerings faster, create new revenue streams, and reduce support costs, all while building a competitive advantage in their markets. In this session, we will use customer use cases...