Before you start: If you don’t know what Katalon Studio is, read this first, and get to know why I picked it to build web automation.

Katalon Studio at a Glance

To be mentioned, before looking at the studio, Katalon project offers an IDE tool for Chrome, called Katalon Recorder. It’s an extension that records actions, captures web elements, plays the automated test case you recorded and speaks Selenese.

Source de l’article sur DZONE

Most companies only use 5 to 10 percent of the data they collect. So estimates Beatriz Sanz Sai, a 20-year veteran in advanced analytics and the head of Ernst and Young’s global data and analytics practice. While it’s impossible to validate such a claim, the fact is many organizations are gathering lots of data but analyzing little of it.

Legacy database management systems shoulder much of the blame for this. They hog time and resources for the sake of storing, managing, and preparing data and thus impede analytics.

Source de l’article sur DZONE

In this ever-growing network of physical devices, Internet of Things (IoT) has become much more realistic, with smart devices that allow less human input and more machine output. Today, it is very much a reality that smart home hubs, wearable devices, connected cars, industrial Internet, smart retail, and many more consider IoT for functionality. 

The desktop apps, servers, and mobile apps use some programming languages that are more or less the same. So it makes us believe that there is no difference in smart objects as they are like minicomputers. IoT comprises of a three-tier architectural environment that consists of generating the data, organizing the data by the local gateways, or hubs and centralized servers that are geographically distant where all the data ends up.

Source de l’article sur DZONE

Writing code to make sure that input parameters are not null does not make Java code "safer;" it’s exactly the opposite — it makes code less readable and less safe.

Code With Null-Checks Is Less Readable

It would be difficult to argue that null-checks are attractive. Most of the time, it’s just boilerplate code, which contributes nothing to the "logic" of the method. It is a purely technical construct. Like this one:

Source de l’article sur DZONE

Amazon Simple Storage Service, better known as AWS S3, is one of the oldest and most widely-used solutions from Amazon Web Services. S3 is designed to store terabytes of customer data with high availability and scale. Prior to AWS S3 Intelligent Tiering, Amazon Cloud users had access to four storage class options:

  1. Standard – Designed for frequently accessed data.
  2. Standard-IA – Designed for long-lived, infrequently accessed data.
  3. One Zone-IA – Designed for long-lived, infrequently accessed, non-critical data.
  4. Glacier – Designed for long-lived, infrequent accessed, archived critical data.

The issue here is that access patterns can be different for the different objects. Though customers could build lifecycle policies to move the objects across storage classes, it was challenging to predefine the most appropriate storage class(s) because of the unpredictable nature of application adoption and usage. Even in scenarios where access frequency is known, customers may forget to make the proper class adjustments to optimize their budget.

Source de l’article sur DZONE

In this article, I will explain how to return custom HTTP errors in Spring Boot. When we make an HTTP request to a resource, it is common that the request has to consider the option of returning an error.

It is the typical case that we made a RESTful request to query for a record, but it does not exist. In this case, you will usually return an HTTP code 404 (Not Found), and with this code, you also return a JSON object that with a a format defined for Spring Boot, like this:

Source de l’article sur DZONE

Marriott International has disclosed that the guest reservation database of its Starwood division has been breached, affecting as many as 500 million guests. The company has also confirmed that there has been unauthorized access to the Starwood network since 2014.

According to a report from the BBC, for roughly 327 million guests, the attacker was able to access personally identifiable information including a combination of name, address, phone number, email address, passport number, account information, date of birth, and gender. In some cases, the compromised records also included encrypted credit card information. The company is still trying to determine whether or not the encryption keys have also been stolen.

Source de l’article sur DZONE


Introduction

In very simple language, Pattern Recognition is a type of problem while Machine Learning is a type of solution. Pattern recognition is closely related to Artificial Intelligence and Machine Learning. Pattern Recognition is an engineering application of Machine Learning. Machine Learning deals with the construction and study of systems that can learn from data, rather than follow only explicitly programmed instructions whereas Pattern recognition is the recognition of patterns and regularities in data.

  1. Machine Learning

The goal of Machine Learning is never to make "perfect" guesses because Machine Learning deals in domains where there is no such thing. The goal is to make guesses that are good enough to be useful. Machine Learning is a method of data analysis that automates analytical model building. Machine Learning is a field that uses algorithms to learn from data and make predictions. A Machine Learning algorithm then takes these examples and produces a program that does the job. Machine Learning builds heavily on statistics. For example, when we train our machine to learn, we have to give it a statistically significant random sample as training data. If the training set is not random, we run the risk of the Machine Learning patterns that aren’t actually there.

Source de l’article sur DZONE

To understand the current and future state of big data, we spoke to 31 IT executives from 28 organizations. We asked them, "What’s the future of big data ingestion, management, and analysis from your perspective – where do the greatest opportunities lie?" Here’s what they told us:

AI/ML

  • We’ll see the transition from on-prem to the cloud and subsequently see traditional Hadoop make the transition to the cloud. This will lead to higher adoption of AI/ML. 
  • Just drive the digitization agenda of the company. You have sufficient compute power and data – what could you do? Take advantage of the capability. Use AI/ML to filter through the data. Enable more people to get involved. 
  • Leverage big data and ML anomaly detection with more sensors entering the world. Cameras checking on safety helmets, ML models from city sensors early warning indicators. The entire economy becomes information driven. Understand why anomalies might happen. 
  • 1) AI/ML becoming less hype and more of a trend. ML needs big data to work. Any ML requires big data. Big data is not useful by itself. Ability to have an engine automatically see trends and make suggestions on what to look at is valuable. 2) Expect more tools for visualizing and reporting on big data. Salesforce has Einstein. Tableau has a tool. Expect thousands more we haven’t seen yet. AI/ML will become more prevalent. 
  • AI protected systems. Maintain and keep the data safer. Create ethical and moral dilemmas for humans. Protect the data because at some point it will be turned over to machines which is terrifying because you don’t know what the machine may do with it and you cannot recover. 
  • The use of AI and ML technologies, like TensorFlow, providing the greatest possible future opportunities for big data applications. With AI the computer uncovers patterns that a human is unable to see. 
  • We’re going to suffer a talent problem in organizations. Ability to make the value of the data visible to people who are not data scientists is an important factor to deal with. AI/ML will focus on making sense of data to provide answers to people. Context is also important — how can we create context and get it out of people’s heads?
  • Maturing past the Hadoop data lake world. Hadoop is a workload good for some things and not good for everything. Everyone is taking a deep breath. Hadoop is good for these things. THe same is true for the data lake. You have to go through the growing pains to figure it out. Opportunity increases as we get more into the world of AI and the system is finding things in the future, that’s the reality, we’ll get there as an industry. Huge opportunity to do across data and workloads. You have to scope that. Some use cases and workloads.

Streaming

  • Streaming for more real-time, faster ingestion, and analysis. Still in the early days of animated actions with the data.
  • How to connect real-time data to get the most out of big data. Connect data and dots to explore and predict relationships.

Tools

  • Recognition about the variety of data. Rationalize across all the different kinds of data. Aggregation of variables – credit bureau, core banking system, data on Hadoop. There’s an opportunity with the proliferation of tools you can put in the hands of the data analysts or business users rather than relying on data governance or DBAs. Give people access to the data and the tools to manipulate. 
  • More maturity and more tools with the ability to interpret. 1) More data, more types, streaming more quickly. 2) Analytical methods used to process the data. 3) Automation of an insight. 
  • The trend to make the common denominator across systems more SQL-centric through an API. SQL is how devs interact with data across different systems. Move to more open source and lower cost tooling as a visualization step. The difference between Power BI and Tableau is shrinking. Data-as-a-service makes tools for visualization less critical. Increasing role of the data steward bridge between analyst and data consumer to be more self-sufficient. 
  • There is a continued drive for standardization of data ingestion, with many companies looking to Kafka for on-premises or private cloud or Kinesis for AWS cloud. Data management and analytic tools then become sinks and sources for data for those data movement frameworks, which creates a sort of data utility grid for these companies, sort of like the electrical system in a house. If you need electricity, you just need an appliance with a standard plug and you plug in. The same is occurring with data access — and is already in place at some companies — if you need to get use of data or provide data to someone else, you just plug your application into the data grid using their standard “plug” (or interface). This will also allow for more “best of breed” components to be used, like the best BI or analytics tool or best database for a particular workload rather than having to compromise on an inferior all-in-one product since the data integration will be more standard than custom. Localization of data is a great opportunity, too. That is, having data located in the world where it is needed rather than needed to traverse long networks in order to retrieve it, process it, change it, or analyze it. That means more master-master, active-active architectures which can create application challenges for any enterprise, so the right choice of components will be important.  
  • Leading companies are increasingly standardizing on mature open source technologies like Apache Kafka, Apache Ignite, and Apache Spark to ingest, manage and analyze their big data. All of these projects have experienced major adoption growth in the past few years and it appears likely this will continue for the foreseeable future. As these technologies mature and become increasingly easier to install and use, they will create opportunities for those who know how to use and implement distributed computing technologies for an increasingly real-time world.

Other

  • Look at tagging, get the proper metadata models, ensure the context of the information. Tags and metadata draw context. Ensure proper metadata is wrapped around. Have traceability for reliability.
  • Focus on operationalization driven by the continued emergence of streaming always-on technology. Complete understanding of what’s going on. The cloud drives this home where cloud-based application architectures are always on and being updated. The same needs to happen with data architectures with automation. Customers see themselves going down a data operations path.
  • All three parts of big data can lead to a considerably successful project in terms of ROI as well as data governance. I would order them hierarchically. First, we need to be able to collect data in large amounts from many different sources. Once the data becomes available, the proper management, like the proper creation of informative KPIs, might already lead to some unexpected discoveries. Finally, after the data have been so transformed, their analysis produces even further insights that are vital for the company business. So, as you see, you already get information from step 1. But you can get step 2 without having completed step 1 first.
  • This will all become easier. Things that are challenging today will become second nature and automated in the future. See ease of accessing big data just as easy as anything we do on a computer. Handling, moving, connecting will have far less friction.  Using big data identifying the value proposition within the data is where the opportunities lie within each business.
  • Augmented analytics pulling together natural language, data, and analytics to drive answers. How do we get to analyzing based on identifying what you don’t know to query?
  • Data analysts and scientists don’t care where the data is, they just want the data and the tools they need to analyze it. Catalog and know where the data is. Next step just want data where I want it. Build a virtual catalog to access delivery. There’s a logical progression of what we’re doing.
  • Regardless of on-prem or cloud needs companies to ensure the engine keeps working so you can get value. As a service model doesn’t automatically solve the problems. Need to know and manage performance problems. Bring performance transparency. Think through security from end-to-end.
  • The future is big data analytical platforms that provide proven capabilities for ingestion, management, and analysis at the speed and scale that enables businesses to compete and win. The greatest opportunities are for businesses to no longer be constrained by the imagination of the business in getting accurate insight so that they can act on all opportunities – understand exactly which customers are likely to churn and grow your business, establish entirely new business models based on the revenue-generating capabilities of data (think pay-as-you-drive insurance, as an example). Every single industry can differentiate itself based on the valuable insight of the data. Make an investment in a proven data analytical platform that offers no compromises and prepares you for whatever the future holds in terms of deployment models (clouds, on-premises, on Hadoop, etc.) or seamless integration with emerging technologies in the data ecosystem.
  • The greatest opportunities lie in delivering true agile data engineering processes that allow companies to quickly create data pipelines to answer new business questions without requiring business people to depend on IT. This requires the automation of the end-to-end development, operationalization, and ongoing governance of big data environments in an integrated fashion. The key to success is automating away the complexity so organizations can use people with basic SQL and data management skills to fully leverage big data for competitive advantage.
  • There is a very bright future ahead for all of these. One area of great opportunity is in the IoT arena. There are over 9 billion devices deployed and the rate of deployment is speeding up as the cost of devices decreases and the sophistication of devices increases. This device data requires very high-speed ingestion and robust management. It is also ripe for advanced analytics such as machine learning for outlier detection.
  • We see three mission-critical opportunities in the future of data-driven marketing and sales. 1) Cord-Cutters — Our clients’ customers are more mobile and digital than ever. Traditional data elements and IDs such as home phone, home address, business extension, etc. have to be complemented with digital IDs such as mobile phone number, GPS coordinates, cookie ID, device ID, MAIDs, etc. 2) Predictive World — Artificial intelligence is woven throughout our everyday lives and experiences. Our phones predict the next few words in the sentence we are texting. Our thermostats predict what temperature is optimal for personal warmth and cost savings. Our cars brake for us before an accident happens. Consumers now expect marketing and sales experiences will also be predictive, using data and intelligence to improve their brand experiences in real-time. 3) B2B2C Life — There is a blending of our business and consumer selves. Research shows that approximately 43% of consumer work remotely and the number of people that spend > 50% of their time working at home has grown 115% over the past 10 years. Therefore, marketers must be able to connect the data IDs, attributes and behaviors of individuals versus siloed B2B or B2C targeting. 

Here’s who we spoke to:

Source de l’article sur DZONE

Multiple tech news sites have reported that  event-stream , a popular JavaScript library with almost two million weekly downloads, contained malicious code for several weeks. It was first reported in a GitHub post last week, but developers were only able to decipher the purpose of the code yesterday. The code targeted users of the cryptocurrency wallet Copay, and was designed to steal Bitcoin and Bitcoin Cash.

GitHub user right9ctrl injected the malicious code in a package called flatmap-stream after offering to help maintain the library, according to a post on the Snyk security blog. Handing off control to other users happens frequently in open source communities, as the original authors and maintainers move on to other projects. Adding libraries as dependencies is also common in open source, although the  event-stream npm package hadn’t been substantially updated for about two years prior to September 2018, when flatmap-stream was first added. According to Ars Technica, the next phase took place on October 5, when malicious code was implemented to transfer the balances of Copay wallets to a server in Kuala Lumpur. 

Source de l’article sur DZONE