Welcome!

@DXWorldExpo Authors: William Schmarzo, Rostyslav Demush, Jason Bloomberg, Greg Pierce, Stackify Blog

Related Topics: @DXWorldExpo, Cognitive Computing , @ThingsExpo

@DXWorldExpo: Blog Feed Post

Data Is a New Currency | @CloudExpo #BigData #BI #AI #ML #Analytics

Data is the new Oil

This guest blog is provided by Brandon Kaier (@bkaier). Brandon has more than two decades of experience in the IT industry as a transformational leader.  Kaier is responsible for setting the strategy, defining the service line offerings and capabilities as the Field CTO for the North America. These responsibilities include bringing net new products and solutions to market with a focus on the deployment of solutions enabling analytics and agile application development. Most recently Kaier held positions on Technical Architecture team and Strategy and Direction team for Dell EMC’s Data Lake solutions. Kaier also brings experienced based, Dell EMC IT best practices, Big Data and Transformational strategies to Dell EMC customers.

“Data is the new Oil.”

Has anyone not heard this phrase yet? This analogy was first presented by Clive Humby at the Association of National Advertisers (ANA), Senior Marketers Summit at Kellogg School in 2006.

If you have been following and or are active in the analytics and data science community, I am sure that you have, and honestly I would hope so. This was not the first time we heard the phrase “data is the new oil,” and it certainly will not be the last. Interestingly enough this exclamation is actually at least ten years old. For example, marketing commentator Michael Palmer blogged back in 2006: “Data is just like crude. It’s valuable, but if unrefined it cannot really be used. It has to be changed into gas, plastic, chemicals, etc., to create a valuable entity that drives profitable activity; so must data be broken down, analyzed for it to have value.”

I really like Michael’s extended comment here because I think that it is much needed clarity. The difference between a barrel of oil and highway traffic is one of energy. A barrel of oil had a ton of Potential Energy and burning that oil as fuel creating the movement of the cars is its conversion to Kinetic Energy. (Definition: Kinetic Energy is the energy of motion; the movement of objects. Objects that are not in motion possess Potential Energy, which is converted to kinetic energy when some force acts upon the object to set it in motion.) This is the essential challenge of analytics. Data sitting on your data center floor is generally valueless except for the time at which it was created. Let me be clear, this is not to say that the data isn’t important. It is, you need it for all sorts of things; GRC, management and compliance reporting, order tracking, and what not, but it is no longer creating value. Data is like oil in that it retains a ton of Potential Value. However, it requires analytics to create motion in the business to find something actionable or decisions to optimize in order to realize that value.

Oil Data
Oil has Potential Energy Data has Potential Value
Gas or refined Oil has more Potential Energy than Oil by volume Data Science or refined Data increases the Potential Value of Data
Burning Gas to create motion converts Potential Energy to Kinetic Energy Analyzing Data to create Analytic Insights converts Potential Value to Kinetic Value
Oil is raw and of little direct use Data is raw and of little direct use
Oil is refined into fuel; powers the economy and drove the 3rd Industrial Revolution Data is refined by analytics to power the business and shall fuel 4th Ind. Rev.

So the really interesting question to ask now is “What is the economic value of data?”

There is this idea that currency, once spent does not lose value, it retains it. It retains its potential energy or potential value, in other words…..

Think of a $10 bill.

You may take it and give it to Starbucks and while, yes, you are out that $10 the $10 bill is still worth well, $10.

Starbucks is going to use that $10 to buy coffee from a distributor.

Who in turn buys coffee from a grower, who might then use it on food or maybe even  a Starbucks.

When you insert something, a new demand, into a circular flow like this then you create an economic concept called the Multiplier Effect. This is a concept that countries use to consider how they invest money and how that investment, by having it distribute though a supply chain, like the example above, will impact the economy of their country.

Multiplier Effect Definition: “An effect in economics in which an increase in spending produces an increase in national income and consumption greater than the initial amount spent.”

Multiplier Effect is a well-established economical principal first postulated in 1939 and there are standard formulas to apply this effect based on different kind of investments.

To overly simplify the process it basically equates to 4:1 (Note: the actual multiplier is dependent upon the marginal propensity to consumer (MPC). See here for more details. For every dollar put into the supply chain you are creating $4 of economic value. Whoever they invest in or award a contract to has to buy supplies from someone else and they buy other materials from a third company and so on down the line.

So with that in mind, how does Multiplier Effect impact Data in a modern organization? Bill Schmarzo is noted as saying that, “Data exhibits a Network Effect, where data can be used at the same time across multiple use cases thereby increasing its value to the organization.” I would contend that this network effect is in fact the same thing principally as the Multiplier Effect.

We have seen this Network Effect – the Multiplier Effect – play out within our customer base. Let’s assume that I have three use cases, A, B, and C. Let’s also assume that independently those three use cases have a value of 5, 3, and 2 (high to low) respectively. Now we have one of the data scientists apply the results of use case B to use case C and we find that the value of C goes from 2 to 4. This is what I mean by the Network Effect is similar if not identical to the Multiplier Effect; investment on one side of the value chain has cascading impacts to other parts of the value chain, just like my $10 bill.

I have found some companies who have recognized this inherent value in data sharing and are starting to form ecosystems where the only cost to get into the ecosystem is the contribution of your data. Take a home hospice care company. Their employees have very stressful jobs with a lot of time in their car going form patient to patient. Would it be valuable to work with other companies to share information that would make the lives of their nurses and home care employees better? We are seeing one such case starting to come together in the UK.

If seemingly disparate companies can find value to their business through the sharing of data what is to stop one company from asking to exchange hard tangible goods for data.

I think that companies are going to have to make a decision and I think that it is going to be one of three choices. Fundamentally, they can continue to treat data as a cost to be minimized and as a consequence will relinquish the value in the data to more aggressive partners and competitors at the risk of market share. They will buy into the notion that data is the new oil and as a result treat data like a tangible asset with all of the associated processes; supply chain management, asset capitalization or depreciation one would normal use with a physical asset. This approach will yield results and positive outcomes to the business but will fail to take real advantage of the asset.

The final option, consider data to be in fact a new currency, Leverage the Multiplier Effect internally, then expand it to include value chain partners, and then one day soon, I think, exchange or barter data and the associated analytics for goods and services. Conceptually, this isn’t a new idea. Back in the 80’s, when Point of Sale (POS) scanners were first making their way into retail stores, in many cases those scanners were being given to the retailers, “for free”, in exchange for all the data from those POS devices. If you were to consider what GE knows about jet engines how much longer will it be before GE uses or gives away its data to the supplier of fuel valves for the engine in exchange for those valves?

The post Data is a New Currency appeared first on InFocus Blog | Dell EMC Services.

More Stories By William Schmarzo

Bill Schmarzo, author of “Big Data: Understanding How Data Powers Big Business” and “Big Data MBA: Driving Business Strategies with Data Science”, is responsible for setting strategy and defining the Big Data service offerings for Dell EMC’s Big Data Practice. As a CTO within Dell EMC’s 2,000+ person consulting organization, he works with organizations to identify where and how to start their big data journeys. He’s written white papers, is an avid blogger and is a frequent speaker on the use of Big Data and data science to power an organization’s key business initiatives. He is a University of San Francisco School of Management (SOM) Executive Fellow where he teaches the “Big Data MBA” course. Bill also just completed a research paper on “Determining The Economic Value of Data”. Onalytica recently ranked Bill as #4 Big Data Influencer worldwide. Bill has over three decades of experience in data warehousing, BI and analytics. Bill authored the Vision Workshop methodology that links an organization’s strategic business initiatives with their supporting data and analytic requirements. Bill serves on the City of San Jose’s Technology Innovation Board, and on the faculties of The Data Warehouse Institute and Strata. Previously, Bill was vice president of Analytics at Yahoo where he was responsible for the development of Yahoo’s Advertiser and Website analytics products, including the delivery of “actionable insights” through a holistic user experience. Before that, Bill oversaw the Analytic Applications business unit at Business Objects, including the development, marketing and sales of their industry-defining analytic applications. Bill holds a Masters Business Administration from University of Iowa and a Bachelor of Science degree in Mathematics, Computer Science and Business Administration from Coe College.

@BigDataExpo Stories
DX World EXPO, LLC, a Lighthouse Point, Florida-based startup trade show producer and the creator of "DXWorldEXPO® - Digital Transformation Conference & Expo" has announced its executive management team. The team is headed by Levent Selamoglu, who has been named CEO. "Now is the time for a truly global DX event, to bring together the leading minds from the technology world in a conversation about Digital Transformation," he said in making the announcement.
"Space Monkey by Vivent Smart Home is a product that is a distributed cloud-based edge storage network. Vivent Smart Home, our parent company, is a smart home provider that places a lot of hard drives across homes in North America," explained JT Olds, Director of Engineering, and Brandon Crowfeather, Product Manager, at Vivint Smart Home, in this SYS-CON.tv interview at @ThingsExpo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that Conference Guru has been named “Media Sponsor” of the 22nd International Cloud Expo, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. A valuable conference experience generates new contacts, sales leads, potential strategic partners and potential investors; helps gather competitive intelligence and even provides inspiration for new products and services. Conference Guru works with conference organizers to pass great deals to gre...
In his Opening Keynote at 21st Cloud Expo, John Considine, General Manager of IBM Cloud Infrastructure, led attendees through the exciting evolution of the cloud. He looked at this major disruption from the perspective of technology, business models, and what this means for enterprises of all sizes. John Considine is General Manager of Cloud Infrastructure Services at IBM. In that role he is responsible for leading IBM’s public cloud infrastructure including strategy, development, and offering m...
"Evatronix provides design services to companies that need to integrate the IoT technology in their products but they don't necessarily have the expertise, knowledge and design team to do so," explained Adam Morawiec, VP of Business Development at Evatronix, in this SYS-CON.tv interview at @ThingsExpo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
To get the most out of their data, successful companies are not focusing on queries and data lakes, they are actively integrating analytics into their operations with a data-first application development approach. Real-time adjustments to improve revenues, reduce costs, or mitigate risk rely on applications that minimize latency on a variety of data sources. In his session at @BigDataExpo, Jack Norris, Senior Vice President, Data and Applications at MapR Technologies, reviewed best practices to ...
Widespread fragmentation is stalling the growth of the IIoT and making it difficult for partners to work together. The number of software platforms, apps, hardware and connectivity standards is creating paralysis among businesses that are afraid of being locked into a solution. EdgeX Foundry is unifying the community around a common IoT edge framework and an ecosystem of interoperable components.
"ZeroStack is a startup in Silicon Valley. We're solving a very interesting problem around bringing public cloud convenience with private cloud control for enterprises and mid-size companies," explained Kamesh Pemmaraju, VP of Product Management at ZeroStack, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Large industrial manufacturing organizations are adopting the agile principles of cloud software companies. The industrial manufacturing development process has not scaled over time. Now that design CAD teams are geographically distributed, centralizing their work is key. With large multi-gigabyte projects, outdated tools have stifled industrial team agility, time-to-market milestones, and impacted P&L stakeholders.
"Akvelon is a software development company and we also provide consultancy services to folks who are looking to scale or accelerate their engineering roadmaps," explained Jeremiah Mothersell, Marketing Manager at Akvelon, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"Infoblox does DNS, DHCP and IP address management for not only enterprise networks but cloud networks as well. Customers are looking for a single platform that can extend not only in their private enterprise environment but private cloud, public cloud, tracking all the IP space and everything that is going on in that environment," explained Steve Salo, Principal Systems Engineer at Infoblox, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Conventio...
"IBM is really all in on blockchain. We take a look at sort of the history of blockchain ledger technologies. It started out with bitcoin, Ethereum, and IBM evaluated these particular blockchain technologies and found they were anonymous and permissionless and that many companies were looking for permissioned blockchain," stated René Bostic, Technical VP of the IBM Cloud Unit in North America, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Conventi...
In his session at 21st Cloud Expo, Carl J. Levine, Senior Technical Evangelist for NS1, will objectively discuss how DNS is used to solve Digital Transformation challenges in large SaaS applications, CDNs, AdTech platforms, and other demanding use cases. Carl J. Levine is the Senior Technical Evangelist for NS1. A veteran of the Internet Infrastructure space, he has over a decade of experience with startups, networking protocols and Internet infrastructure, combined with the unique ability to it...
22nd International Cloud Expo, taking place June 5-7, 2018, at the Javits Center in New York City, NY, and co-located with the 1st DXWorld Expo will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud ...
Gemini is Yahoo’s native and search advertising platform. To ensure the quality of a complex distributed system that spans multiple products and components and across various desktop websites and mobile app and web experiences – both Yahoo owned and operated and third-party syndication (supply), with complex interaction with more than a billion users and numerous advertisers globally (demand) – it becomes imperative to automate a set of end-to-end tests 24x7 to detect bugs and regression. In th...
Agile has finally jumped the technology shark, expanding outside the software world. Enterprises are now increasingly adopting Agile practices across their organizations in order to successfully navigate the disruptive waters that threaten to drown them. In our quest for establishing change as a core competency in our organizations, this business-centric notion of Agile is an essential component of Agile Digital Transformation. In the years since the publication of the Agile Manifesto, the conn...
"Codigm is based on the cloud and we are here to explore marketing opportunities in America. Our mission is to make an ecosystem of the SW environment that anyone can understand, learn, teach, and develop the SW on the cloud," explained Sung Tae Ryu, CEO of Codigm, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
High-velocity engineering teams are applying not only continuous delivery processes, but also lessons in experimentation from established leaders like Amazon, Netflix, and Facebook. These companies have made experimentation a foundation for their release processes, allowing them to try out major feature releases and redesigns within smaller groups before making them broadly available. In his session at 21st Cloud Expo, Brian Lucas, Senior Staff Engineer at Optimizely, discussed how by using ne...
Vulnerability management is vital for large companies that need to secure containers across thousands of hosts, but many struggle to understand how exposed they are when they discover a new high security vulnerability. In his session at 21st Cloud Expo, John Morello, CTO of Twistlock, addressed this pressing concern by introducing the concept of the “Vulnerability Risk Tree API,” which brings all the data together in a simple REST endpoint, allowing companies to easily grasp the severity of the ...
Coca-Cola’s Google powered digital signage system lays the groundwork for a more valuable connection between Coke and its customers. Digital signs pair software with high-resolution displays so that a message can be changed instantly based on what the operator wants to communicate or sell. In their Day 3 Keynote at 21st Cloud Expo, Greg Chambers, Global Group Director, Digital Innovation, Coca-Cola, and Vidya Nagarajan, a Senior Product Manager at Google, discussed how from store operations and ...