24
Data Management Challenges in Production Machine Learning An Academic presentation by Dr. Nancy Agnes, Head, Technical Operations, Phdassistance Group www.phdassistance.com Email: [email protected]

Data Management Challenges In Production Machine Learning - Phdassistance

Embed Size (px)

DESCRIPTION

Machine learning’s importance in modern computing cannot be overstated. Machine learning is becoming increasingly popular as a method for extracting knowledge from data and tackling a wide range of computationally difficult tasks, including machine perception, language understanding, health care, genetics, and even the conservation of endangered species [1]. Machine learning is often used to describe the one-time application of a learning algorithm to a given dataset. The user of machine learning in these instances is usually a data scientist or analyst who wants to try it out or utilizes it to extract knowledge from data. Learn More:https://bit.ly/3zk9UkG Contact Us: Website: https://www.phdassistance.com/ UK: +44 7537144372 India No:+91-9176966446 Email: [email protected]

Citation preview

Data Management Challenges in Production Machine LearningAn Academic presentation byDr. Nancy Agnes, Head, Technical Operations, Phdassistance Group www.phdassistance.comEmail: [email protected]

Today's Discussion

Production Machine Learning: Overview and

Assumptions

Data Issues in Production Machine Learning

Enrichment

Future Scope

INTRODUCTION in modern computingMachine learning's importance

cannot be overstated.

Machine learning is becoming increasingly popular as a method for extracting knowledge from data and tackling a wide range of computationally difficult tasks, including machine perception, language understanding, health care, genetics, and even the conservation of endangered species.

Machine learning is often used to describe the one-time application of a learning algorithm to a given dataset.

Contd...

The user of machine learning in these instances is usually a data scientist or analyst who wants to try it out or utilises it to extract knowledge from data.

Our focus here is different, and it considers machine learning implementation in production.

This entails creating a pipeline that reliably ingests training datasets as input and produces a model as output, in most cases constantly and gracefully dealing with various forms of failures.

This scenario usually involves a group of engineers that spend a substantial amount of their time to the less glamorous parts of machine learning, such as maintaining and monitoring machine learning pipelines.

PRODUCTION MACHINE LEARNING: OVERVIEW AND ASSUMPTIONS

A high-level representation of a production machine learning pipeline is shown in Figure 1.

The training datasets that will be provided to the machine learning algorithm are the system's input.

The result is a machine-learned model, which is picked up by serving infrastructure and combined with serving data to provide predictions.

Contd...

Many of the issues we'll discuss below are also applicable in a pure streaming system, as well as for one-time data processing on a single batch.

PhD Assistance experts has experience in handling dissertation and assignment in computer science research with assured 2:1 distinction. Talk to Experts Now

DATA ISSUES IN PRODUCTION MACHINE LEARNINGThe primary issues in handling data for production machine learning pipelines are discussed in this section.

UNDERSTANDING

Engineers who are first setting up a machine learning pipeline spend a large amount of time evaluating their raw data.

Contd...

This procedure entails creating and visualising key aspects of the data, as well as recognising any anomalies or outliers.

It can be difficult to scale this technique to enormous amounts of training data.

Techniques established for online analytical processing [3], data-driven visualisation recommendation [4], and approximation query processing [5] can all be used to create tools that help people comprehend their own data.

Another important step for engineers is to figure out how to encode their data into features that the trainer can understand.

Contd...

For example, if a string feature in the raw data contains nation identifiers; one hot encoding can be used to transform it to an integer feature.

A fascinating and relatively unexplored research field is automatically recommending and producing transformations from raw data to features based on data qualities.

When it comes to comprehending facts, context is equally crucial.

Contd...

In order to design a maintainable machine learning pipeline, it is critical to clearly identify explicit and implicit data dependencies, as described in [2].

Many of the tools developed for data-provenance management may be used to track some of these dependencies, allowing us to better understand how data travels through these complicated pipelines.

VALIDATION

It is difficult to overlook the fact that data validity has a significant impact on the quality of the model developed.

Validity entails ensuring that training data contains the expected characteristics that these features have the expected values that features are associated as expected, and that serving data does not diverge from training data.

Some of the issues can be solved by using well-knowndatabase system technologies.

Contd...

The predicted properties and the characteristics of their values, for example, can be encoded using something close to a training data format.

Hire PhD Assistance experts to develop your algorithm and coding implementation for your Computer Science dissertation Services.

Furthermore, machine learning introduces new restrictions that must be verified, such as bounds on the drift in the statistical distribution of feature values in the training data, or the usage of an embedding for some input feature if and only if other features are normalised in a specified way.

Contd...

Furthermore, unlike a traditional DBMS, any schema over training data must be flexible enough to allow changes in training data features as they reflect real- world occurrences.

In production machine learning pipelines, the difference between serving and training data is a primary source of issues.

The underlying problem is that the data used to build the model differs from the data used to test it, which almost always means that the predictions provided are inaccurate.

Contd...

The final stage is to clean the data in order to correct the problem.

Cleaning can be accomplished by addressing the source of the problem.

Patching the data within the machine learning pipeline as a temporary workaround until the fundamental problem is properly fixed is another option.

This method is based on a large body of research on database repair for specific sorts of constraints [4].

A recent study [6] looked at how similar strategies could be used to a specific class of machine learning algorithms.

ENRICHMENT

Enrichment is the addition of new features to the training and serving data in order to increase the quality of the created model.

toaugmentcurrent acommon form of

Joining in a new datasource features withnew signals is enrichment.Discovering which extra signals or changes can meaningfully enrich the data is a major difficulty in this situation.

Contd...

A catalogue of sources and signals can serve as a starting point for discovery, and recent research has looked into the difficulty of data cataloguing in many contexts [7] as well as the finding of links between sources and signals.

Another significant issue is assisting the team in comprehending the increase in model quality achieved by adding a specific collection of characteristics to the data.

This data will aid the team in determining whether or not to devote resources to applying the enrichment in production.

Contd...

This topic was investigated in a recent study [3] for the situation of joining with new data sources and a certain class of methods, and it would be interesting to consider extensions to additional cases.

Another wrinkle is that data sources may contain sensitive\information and consequently may not be accessible unless the team goes through an access review.

Going through a review and gaining access to sensitive data, on the other hand, can result in operating costs.

As a result, it's worth considering whether the enrichment effect can be approximated in a privacy-preserving manner without access to sensitive data, in order to assist the team in deciding whether to apply for access.

One option is to use techniques from privacy- preserving learning [7], while past research has focused on learning a privacy-preserving model rather than simulating the influence of new characteristics on model quality.

FUTURE SCOPEIT architectures will need to change to accommodate it, but almost every department within a company will undergo adjustments to allow big data to inform and reveal.

Data analysis will change, becoming part of a business process instead of a distinct function performed only by trained specialists.

Big data productivity will come as a result of giving users across the organization the power to work with diverse data sets through self-services tools.

Contd...

Achieving the vast potential of big data demands a thoughtful, holistic approach to data management, analysis and information intelligence.

Across industries, organizations that get ahead of big data will create new operational efficiencies, new revenue streams, differentiated competitive advantage and entirely new business models.

Business leaders should start thinking strategically about how to prepare the organizations for big data.

UNITED KINGDOM+44 7537144372

INDIA+91-9176966446

[email protected]

Contact Us