Artificial intelligence is a trending topic that more and more companies want to use. Intelligent and automatic data evaluation, is of particular interest. However, successful use of machine learning requires pervasive data sets because the AI model is trained with these over many iterations to reliable output results ultimately.
But what does the IT architecture behind it have to look like? After all, it must be able to process the sometimes vast amounts of data and scale quickly. This is anything but a trivial matter, so a conventional architecture is no longer sufficient. Instead, innovative data platforms are needed for this new type of digital application. Below we present an overview of the structure of such an architecture, which we developed in a customer project using the Google Cloud Stack.
The first challenge is the scaling of the IT infrastructure concerning the amount of data. In the next three to four years, an increase of about fivefold is expected. The IT infrastructure that is to house an AI solution for data analysis must therefore be designed for growth from the outset. The increase in continuous data streams—up to 25 percent overall—makes stream processing preferable to batch processing. This often entails a technical change.
To keep up with this, companies have to set a new course concerning the IT architecture and the entire organization. To benefit sustainably from data analyzes of business processes, it is not enough to examine the data pools of isolated silos. Instead, the organization must adapt to a “data culture,” connect previous silos and feed data from all company areas to the AI.
A large part of the data that will flow into analysis processes in the future will be unstructured data – for Example, images, video and audio files, or continuous text. It makes sense to store and process this data using non-relational (or NoSQL ) databases such as MongoDB or CouchDB. However, structured data in SQL databases will by no means lose their validity in the medium term. Therefore, the unstructured data must be combined and merged with structured data, representing an additional challenge.
In addition to all these challenges, know-how and human resources in AI/ ML represent a bottleneck. The organization and infrastructure must generate as much output as possible from as few input hours as possible. This works best with a central Enterprise Data Warehouse (EDW), the structure of which is shown in the next section. For the customer project mentioned, an EDW was introduced with this methodology.
To successfully move from a silo to an EDW infrastructure, the following approach has emerged:
Once all five steps have been completed, the user company receives a powerful solution to gain decision-relevant knowledge from all data streams.
A few obstacles usually need to be cleared on the way to the EDW. First, there are legacy systems, which are relatively expensive to modernize and maintain. This limits scalability so that the infrastructure cannot withstand the rapid growth of data. Therefore, the question must be asked: Are the existing systems even able to support AI and ML solutions? Is the effort involved in running and “tuning” them reasonable given the insights they end up generating?
But not only in the infrastructure but also in the process of data collection, some obstacles need to be overcome. Excessively restrictive data protection and security regulations can significantly limit the necessary consolidation of data streams. In addition, the data sources are often not suitable for permanently storing or feeding in the current data. However, AI insights are only as good and extensive as the available database. Therefore, data quality is the fundamental success factor for any AI strategy.
Our practical Example of a data platform that enables AI analysis functions is based on Google Cloud. However, it could also be built on a comparable provider’s cloud stack, such as Amazon Web Services (AWS) or Microsoft Azure.
The platform is orchestrated according to Continuous Integration / Continuous Delivery (CI/CD) principles. This overcomes previous integration problems so that the developers involved can seamlessly integrate their code into the existing one. Automation comes into play in almost all phases of application development. The following diagram shows what this can look like in practice:
Such a CI/CD pipeline creates a continuous stream of data that leads to insights for the relevant decisions. The solution can react to changes in nearly real-time and take into account feedback loops. For Example, this makes it possible to implement “early warning systems” that enable decisive action in the event of rapid changes.
Finally, it should be mentioned that business analytics is not a purely technical task and that AI/ML models do not lead to results “by themselves.” The contextualization of analysis results and their understanding as a basis for decision-making are still with people – more precisely, with management.
Nevertheless, companies that invest in the appropriate infrastructure today will be able to use the insights from AI analysis for themselves sooner. Over time, your competitive advantage over those competitors who do not want to or cannot raise the data treasure in their company will continue to increase.
ALSO READ: Companies Should Pay Attention To When It Comes To Radio Technology
The Google Threat Horizons report is a document that should be consulted by those involved…
Julius computer-based intelligence is an artificial brainpower ideal for investigating information from Succeed. An instrument…
For CA Technologies, agility, DevOps, feedback, and security constitute the strategic pillars of business development.…
The migration from hybrid Cloud to multi-cloud is of interest to the vast majority of…
The Internet has made the world an actual global village. Its advent broke down physical,…
With the blast in the notoriety of virtual entertainment, it is progressively challenging for a…