Articles

In the five-year span between 2015 and 2020, the Internet of Things is expected to grow faster than any other category of connected devices. The number of machine-to-machine connections should grow nearly 2.5-fold, from 4.9 billion in 2015 to 12.2 billion in 2020.

However, in the IoT world, being able to run analytics as close to the source as possible has become a necessity in order to reduce the amount of data being transferred over the wire. Not only that, it’s also crucial if your goal is to provide quick feedback, even if limited, to the edge device (transportation vehicle, oil rig, mobile device, etc).

Source de l’article sur DZONE

To understand the current and future state of IoT, we spoke to more than a dozen IT executives active in the space. Here’s what they told us when we asked, "How can organizations get more from IoT?"

Use Case

  • We’ve seen a lot of B2B implementations with high novelty, smart devices of every kind. There needs to be a legitimate use case. With the continued miniaturization of sensors and devices and the proliferation of 5G, there will be plenty of compelling business problems for IoT to solve. 
  • We are making the transformation from just connecting things and getting data to now figuring out the problem we’re trying to solve. There’s a data explosion. We need to determine how to manage and get value from the data. You need to think about business first. What business problem are you trying to solve? Play the “what if?” Dream ask the big question and put in business metrics. Align people from the technology with the business needs and the business partners. Quantify the business value. Identify the vision, strategies, goals, and hypotheses to validate. Have a clear destination. 
  • I think companies can get the most out of IoT if they start by looking for problems or opportunities that can be addressed by IoT technologies like a temperature monitoring system with alerts or remote control capabilities. Too often, companies look at the products or solutions on the market first and then try to think of ways they could apply them. It’s generally just faster and more productive to go after solving a well-defined issue first — and besides, building some expertise and practice in implementing IoT, it often leads to faster wins and a sense of momentum. 
  • In order to get the most out of the IoT, companies should focus on two things: leveraging most of your technology and building a true revenue model. For example, are you leveraging your data in the best ways? If not, what can you adjust in your business model to ensure you are properly leveraging that data. Companies also need to ask themselves: can I create a business model around connectivity that justifies the recurring cost incurred by connected devices? Many companies work backward by imagining the connected product first, and then the value proposition. These IoT projects are hardly ever successful because the company never took the time to fully understand the problem they were trying to solve. Simply adding an Internet connection to your widget doesn’t mean your business will make immediate profits. IoT products come with significant ongoing costs — web infrastructure, networking, and other connectivity and data-related costs. If you can’t justify the added value to your customers, those costs will eat away at your margins. The most successful IoT products are those that deliver recurring, continuous value for your customers (and recurring revenue for you). Companies are finding ways to deliver this recurring, continuous value by using IoT for preventative maintenance, asset tracking, and environmental monitoring. These are business models that not only contribute back to the business but help the customer as well.

Value of Data

  • One thing IoT companies don’t realize is how valuable their data really is. Take a home automation company. The amount of data streaming through their service is staggering — temperatures, energy usage, humidity, the list goes on. They can take that data, turn it into a data firehose, and make it consumable as a business in itself.
  • 1) Think about short-term design to feed into the long-term. Value comes from applications and services that make use of the data. You need to address a real business need and be able to generate the real value, which is to sell and make money.
    2) In the long-term, you need to be able to support and scale. That’s where standards and open source come in. Devices needed to get smart and detected aren’t real expensive. Make sure you have a support structure to manage the cost so as not to eat up benefits. Right now, every time you get a new IoT device, you get a new app for your smartphone. That doesn’t scale; hence, we have a need for standards. You need to be able to bridge to other ecosystems.
  • Connect the data to the problems people have. Five years ago, we focused on the data but didn’t gain traction until realize how data impacts the people in the industry. Business improvement and optimization software — how it impacts the people in their day-to-day life. If you don’t make the connection, you won’t get the adoption.

Other

  • The main benefit that APIs bring is the ability to stitch together IoT deployments within a wider ecosystem of other applications and capabilities across the business. When IoT assets are exposed internally as APIs, they form part of an application network, which provides a way of connecting IoT capabilities with other applications, data, and devices. In this model, these assets are reusable across the business, removing the need for IT to create point-to-point connections for every IoT deployment. As such, APIs become the ‘digital glue,’ providing a future-proof way of combining IoT with other business systems to create a rich ecosystem that gets the most benefit from IoT deployments, all within a secure-by-design approach.

    Source de l’article sur DZONE

To understand the current and future state of big data, we spoke to 31 IT executives from 28 organizations. We asked them, "What’s the future of big data ingestion, management, and analysis from your perspective – where do the greatest opportunities lie?" Here’s what they told us:

AI/ML

  • We’ll see the transition from on-prem to the cloud and subsequently see traditional Hadoop make the transition to the cloud. This will lead to higher adoption of AI/ML. 
  • Just drive the digitization agenda of the company. You have sufficient compute power and data – what could you do? Take advantage of the capability. Use AI/ML to filter through the data. Enable more people to get involved. 
  • Leverage big data and ML anomaly detection with more sensors entering the world. Cameras checking on safety helmets, ML models from city sensors early warning indicators. The entire economy becomes information driven. Understand why anomalies might happen. 
  • 1) AI/ML becoming less hype and more of a trend. ML needs big data to work. Any ML requires big data. Big data is not useful by itself. Ability to have an engine automatically see trends and make suggestions on what to look at is valuable. 2) Expect more tools for visualizing and reporting on big data. Salesforce has Einstein. Tableau has a tool. Expect thousands more we haven’t seen yet. AI/ML will become more prevalent. 
  • AI protected systems. Maintain and keep the data safer. Create ethical and moral dilemmas for humans. Protect the data because at some point it will be turned over to machines which is terrifying because you don’t know what the machine may do with it and you cannot recover. 
  • The use of AI and ML technologies, like TensorFlow, providing the greatest possible future opportunities for big data applications. With AI the computer uncovers patterns that a human is unable to see. 
  • We’re going to suffer a talent problem in organizations. Ability to make the value of the data visible to people who are not data scientists is an important factor to deal with. AI/ML will focus on making sense of data to provide answers to people. Context is also important — how can we create context and get it out of people’s heads?
  • Maturing past the Hadoop data lake world. Hadoop is a workload good for some things and not good for everything. Everyone is taking a deep breath. Hadoop is good for these things. THe same is true for the data lake. You have to go through the growing pains to figure it out. Opportunity increases as we get more into the world of AI and the system is finding things in the future, that’s the reality, we’ll get there as an industry. Huge opportunity to do across data and workloads. You have to scope that. Some use cases and workloads.

Streaming

  • Streaming for more real-time, faster ingestion, and analysis. Still in the early days of animated actions with the data.
  • How to connect real-time data to get the most out of big data. Connect data and dots to explore and predict relationships.

Tools

  • Recognition about the variety of data. Rationalize across all the different kinds of data. Aggregation of variables – credit bureau, core banking system, data on Hadoop. There’s an opportunity with the proliferation of tools you can put in the hands of the data analysts or business users rather than relying on data governance or DBAs. Give people access to the data and the tools to manipulate. 
  • More maturity and more tools with the ability to interpret. 1) More data, more types, streaming more quickly. 2) Analytical methods used to process the data. 3) Automation of an insight. 
  • The trend to make the common denominator across systems more SQL-centric through an API. SQL is how devs interact with data across different systems. Move to more open source and lower cost tooling as a visualization step. The difference between Power BI and Tableau is shrinking. Data-as-a-service makes tools for visualization less critical. Increasing role of the data steward bridge between analyst and data consumer to be more self-sufficient. 
  • There is a continued drive for standardization of data ingestion, with many companies looking to Kafka for on-premises or private cloud or Kinesis for AWS cloud. Data management and analytic tools then become sinks and sources for data for those data movement frameworks, which creates a sort of data utility grid for these companies, sort of like the electrical system in a house. If you need electricity, you just need an appliance with a standard plug and you plug in. The same is occurring with data access — and is already in place at some companies — if you need to get use of data or provide data to someone else, you just plug your application into the data grid using their standard “plug” (or interface). This will also allow for more “best of breed” components to be used, like the best BI or analytics tool or best database for a particular workload rather than having to compromise on an inferior all-in-one product since the data integration will be more standard than custom. Localization of data is a great opportunity, too. That is, having data located in the world where it is needed rather than needed to traverse long networks in order to retrieve it, process it, change it, or analyze it. That means more master-master, active-active architectures which can create application challenges for any enterprise, so the right choice of components will be important.  
  • Leading companies are increasingly standardizing on mature open source technologies like Apache Kafka, Apache Ignite, and Apache Spark to ingest, manage and analyze their big data. All of these projects have experienced major adoption growth in the past few years and it appears likely this will continue for the foreseeable future. As these technologies mature and become increasingly easier to install and use, they will create opportunities for those who know how to use and implement distributed computing technologies for an increasingly real-time world.

Other

  • Look at tagging, get the proper metadata models, ensure the context of the information. Tags and metadata draw context. Ensure proper metadata is wrapped around. Have traceability for reliability.
  • Focus on operationalization driven by the continued emergence of streaming always-on technology. Complete understanding of what’s going on. The cloud drives this home where cloud-based application architectures are always on and being updated. The same needs to happen with data architectures with automation. Customers see themselves going down a data operations path.
  • All three parts of big data can lead to a considerably successful project in terms of ROI as well as data governance. I would order them hierarchically. First, we need to be able to collect data in large amounts from many different sources. Once the data becomes available, the proper management, like the proper creation of informative KPIs, might already lead to some unexpected discoveries. Finally, after the data have been so transformed, their analysis produces even further insights that are vital for the company business. So, as you see, you already get information from step 1. But you can get step 2 without having completed step 1 first.
  • This will all become easier. Things that are challenging today will become second nature and automated in the future. See ease of accessing big data just as easy as anything we do on a computer. Handling, moving, connecting will have far less friction.  Using big data identifying the value proposition within the data is where the opportunities lie within each business.
  • Augmented analytics pulling together natural language, data, and analytics to drive answers. How do we get to analyzing based on identifying what you don’t know to query?
  • Data analysts and scientists don’t care where the data is, they just want the data and the tools they need to analyze it. Catalog and know where the data is. Next step just want data where I want it. Build a virtual catalog to access delivery. There’s a logical progression of what we’re doing.
  • Regardless of on-prem or cloud needs companies to ensure the engine keeps working so you can get value. As a service model doesn’t automatically solve the problems. Need to know and manage performance problems. Bring performance transparency. Think through security from end-to-end.
  • The future is big data analytical platforms that provide proven capabilities for ingestion, management, and analysis at the speed and scale that enables businesses to compete and win. The greatest opportunities are for businesses to no longer be constrained by the imagination of the business in getting accurate insight so that they can act on all opportunities – understand exactly which customers are likely to churn and grow your business, establish entirely new business models based on the revenue-generating capabilities of data (think pay-as-you-drive insurance, as an example). Every single industry can differentiate itself based on the valuable insight of the data. Make an investment in a proven data analytical platform that offers no compromises and prepares you for whatever the future holds in terms of deployment models (clouds, on-premises, on Hadoop, etc.) or seamless integration with emerging technologies in the data ecosystem.
  • The greatest opportunities lie in delivering true agile data engineering processes that allow companies to quickly create data pipelines to answer new business questions without requiring business people to depend on IT. This requires the automation of the end-to-end development, operationalization, and ongoing governance of big data environments in an integrated fashion. The key to success is automating away the complexity so organizations can use people with basic SQL and data management skills to fully leverage big data for competitive advantage.
  • There is a very bright future ahead for all of these. One area of great opportunity is in the IoT arena. There are over 9 billion devices deployed and the rate of deployment is speeding up as the cost of devices decreases and the sophistication of devices increases. This device data requires very high-speed ingestion and robust management. It is also ripe for advanced analytics such as machine learning for outlier detection.
  • We see three mission-critical opportunities in the future of data-driven marketing and sales. 1) Cord-Cutters — Our clients’ customers are more mobile and digital than ever. Traditional data elements and IDs such as home phone, home address, business extension, etc. have to be complemented with digital IDs such as mobile phone number, GPS coordinates, cookie ID, device ID, MAIDs, etc. 2) Predictive World — Artificial intelligence is woven throughout our everyday lives and experiences. Our phones predict the next few words in the sentence we are texting. Our thermostats predict what temperature is optimal for personal warmth and cost savings. Our cars brake for us before an accident happens. Consumers now expect marketing and sales experiences will also be predictive, using data and intelligence to improve their brand experiences in real-time. 3) B2B2C Life — There is a blending of our business and consumer selves. Research shows that approximately 43% of consumer work remotely and the number of people that spend > 50% of their time working at home has grown 115% over the past 10 years. Therefore, marketers must be able to connect the data IDs, attributes and behaviors of individuals versus siloed B2B or B2C targeting. 

Here’s who we spoke to:

Source de l’article sur DZONE