Skip to main content

Data is the New oil of Industry?

Let's go back to 18th century ,when development was taking its first footstep.The time when oil was considered to be the subset of industrial revolution. Oil than tends to be the most valuable asset in those time. Now let's come back in present. In 21st century, data is vigorously called the foundation of information revolution. But the question that arises is why are we really calling data as the new oil. Well for it's explanation


Now we are going to compare Data Vs Oil

  1. Data is an essential resource that powers the information economy in much the way that oil has fueled the industrial economy.
  2. Once upon a time, the wealthiest were those with most natural resources, now it’s knowledge economy, where the more you know is proportional to more data that you have.
  3. Information can be extracted from data just as energy can be extracted from oil.
  4. Traditional Oil powered the transportation era, in the same way that Data as the new oil is also powering the emerging transportation options like driverless cars and hyperloop (1200km/hr) which are based on advanced synthesis of data inform of algorithms and cognitive knowledge without use of fossil fuel.
  5. Traditional oil is finite, Data availability seems infinite.
  6. Data flows like oil but we must “drill down” into data to extract value from it. Data promises a plethora of new uses — diagnosis of diseases, direction of traffic patterns, etc. — just as oil has produced useful plastics, petrochemicals, lubricants, gasoline, and home heating.
  7. Oil is a scarce resource. Data isn’t just abundant, it is a cumulative resource.
  8. If Oil is being used, then the same oil cannot be used somewhere else because it’s a rival good. This results in a natural tension about who controls oil. If Data is being used, the same Data can be used elsewhere because it’s a non-rival good.
  9. As a tangible product, Oil faces high friction, transportation and storage costs. As an intangible product, Data has much lower friction, transportation and storage costs.
  10. The life cycle of Oil is defined by process: extraction, refining, distribution. The life cycle of Data is defined by relationships: with other data, with context and with itself via feedback loops.
Data is valuable, and can be ‘mined’ and refined, like oil. But there are many differences where the analogy breaks down:
  • Oil is a finite resource that that we are drawing down on. Data is growing at an exponential rate.
  • Oil is consumed when it is used. Data is not. We can make copies of data.
  • Oil is stored physically and is not easily replicable. Data is stored digitally and is readily replicated.
  • Oil is a commodity. Data is highly context dependent.
  • There are lots of other analogies for data as well. For example:
  1. Data is like currency (a medium for exchange, when we exchange our data for ‘free’ services)
  2. Data is like water (abundant and essential for our survival, but requiring cleaning)
  3. Data is a weapon (dormant, but with the potential to cause harm)
  4. However, for all of these, they only show some aspects of data while editing out the others. Ultimately, all analogies break down and it may be futile looking for a single phrase to capture the multi-faceted nature of data.
As Per me this is a Subjective , everyone has own explanation ðŸ˜‰

Happy Learning...!!

Comments

  1. Very Informative and creative contents. This concept is a good way to enhance the knowledge. thanks for sharing. Continue to share your knowledge through articles like these, and keep posting more blogs. visit below for

    Data Engineering Solutions 

    AI & ML Service

    Data Analytics Solutions

    Data Modernization Solutions

    ReplyDelete

Post a Comment

Popular posts from this blog

Statistics Interview Questions Part-1

Q1. What is the difference between “long” and “wide” format data? In the  wide-format , a subject’s repeated responses will be in a single row, and each response is in a separate column. In the  long-format , each row is a one-time point per subject. You can recognize data in wide format by the fact that columns generally represent groups. Q2. What do you understand by the term Normal Distribution? Data is usually distributed in different ways with a bias to the left or to the right or it can all be jumbled up. However, there are chances that data is distributed around a central value without any bias to the left or right and reaches normal distribution in the form of a bell-shaped curve. Figure:   Normal distribution in a bell curve The random variables are distributed in the form of a symmetrical, bell-shaped curve. Properties of Normal Distribution are as follows; Unimodal -one mode Symmetrical -left and right halves are mirror image...

Scope of an Artificial Intelligence

Artificial Intelligence has grown exponentially in the past decade, and so have the career opportunities as an AI expert/specialist. But what exactly does an AI expert do? Also, is becoming an expert the only option while pursuing a career in artificial intelligence?I don’t have any programming/ coding background. Can I still work as an AI expert? And, what specialization or skill set do I need to acquire to get into this field? Skills Required to Build a Career in Artificial Intelligence 1. Sound Mathematical and Algorithmic Understanding To be an ideal candidate in AI, you need to have solid knowledge of applied mathematics and a set of algorithms. Having proficiency in problem-solving and analytical abilities will help you in performing tasks in a more efficient way. You must also have reasonable knowledge of statistics and probability. This helps in understanding various models of AI, like Naive Bayes, Gaussian Mixture Model, etc. 2. Basic Know-How of Programmin...

Data Science Interview Questions -Part 2

1) What are the differences between supervised and unsupervised learning? Supervised Learning Unsupervised Learning Uses known and labeled data as input Supervised learning has a feedback mechanism  Most commonly used supervised learning algorithms are decision trees, logistic regression, and support vector machine Uses unlabeled data as input Unsupervised learning has no feedback mechanism  Most commonly used unsupervised learning algorithms are k-means clustering, hierarchical clustering, and apriori algorithm 2) How is logistic regression done? Logistic regression measures the relationship between the dependent variable (our label of what we want to predict) and one or more independent variables (our features) by estimating probability using its underlying logistic function (sigmoid). The image shown below depicts how logistic regression works: The formula and graph for the sigmoid function is as shown: 3) Explain the steps in making a deci...