Model Evaluation and Validation in Bioprocess Engineering
Expert-defined terms from the Professional Certificate in AI Applications in Bioprocess Engineering course at LearnUNI. Free to read, free to share, paired with a globally recognised certification pathway.
Accuracy #
A performance metric for classification models that measures the proportion of correct predictions out of all predictions made. It is calculated as the number of true positives and true negatives divided by the total number of samples.
Area Under the ROC Curve (AUC #
ROC): A performance metric for classification models that measures the model's ability to discriminate between positive and negative classes. It is calculated as the area under the Receiver Operating Characteristic (ROC) curve, which plots the true positive rate against the false positive rate at various threshold settings.
Bias #
A measure of the difference between the expected predictions of a model and the true values. A high bias model is overly simplistic and tends to underfit the data, resulting in poor performance on both training and testing data.
Calibration Curve #
A graphical representation of the relationship between the predicted probabilities of a model and the true probabilities. It is used to evaluate the reliability of a model's predictions and to detect any systematic biases.
Cross #
Validation: A technique for evaluating the performance of a model by dividing the data into multiple folds, training the model on one fold and testing it on the remaining folds, and repeating the process for each fold. This provides a more reliable estimate of the model's performance than using a single training and testing dataset.
F1 Score #
A performance metric for classification models that combines precision and recall into a single metric. It is calculated as the harmonic mean of precision and recall and ranges from 0 to 1, with a higher value indicating better performance.
Holdout Validation #
A technique for evaluating the performance of a model by dividing the data into a training set and a testing set. The model is trained on the training set and tested on the testing set to evaluate its performance.
K #
Fold Cross-Validation: A type of cross-validation where the data is divided into K folds, and the model is trained and tested K times, with each fold serving as the testing set once. The average performance across all K runs is used as the final performance metric.
Log Loss #
A performance metric for classification models that measures the accuracy of the predicted probabilities. It is calculated as the negative logarithm of the likelihood function and penalizes models that assign low probabilities to true positive and true negative samples.
Mean Absolute Error (MAE) #
A performance metric for regression models that measures the average difference between the predicted and actual values. It is calculated as the absolute difference between the predicted and actual values, averaged over all samples.
Mean Squared Error (MSE) #
A performance metric for regression models that measures the average of the squared differences between the predicted and actual values. It is calculated as the squared difference between the predicted and actual values, averaged over all samples.
Model Evaluation #
The process of assessing the performance of a model using various performance metrics and validation techniques. The goal is to determine the model's ability to generalize to new, unseen data and to identify any weaknesses or biases in the model.
Model Validation #
The process of evaluating the performance of a model using a separate testing dataset. The goal is to assess the model's ability to generalize to new, unseen data and to ensure that the model's performance is not over-estimated due to overfitting.
Overfitting #
A situation where a model is too complex and learns the noise in the training data, resulting in poor performance on new, unseen data. Overfitting is characterized by high performance on the training data and low performance on the testing data.
Precision #
A performance metric for classification models that measures the proportion of true positive predictions out of all positive predictions. It is calculated as the number of true positives divided by the sum of true positives and false positives.
Receiver Operating Characteristic (ROC) Curve #
A graphical representation of the performance of a classification model at various threshold settings. It plots the true positive rate against the false positive rate and is used to evaluate the model's ability to discriminate between positive and negative classes.
Recall #
A performance metric for classification models that measures the proportion of true positive predictions out of all actual positive samples. It is calculated as the number of true positives divided by the sum of true positives and false negatives.
Regression #
A type of supervised learning task where the goal is to predict a continuous output variable based on one or more input variables.
Residual Analysis #
The process of analyzing the residuals, or the differences between the predicted and actual values, to evaluate the performance of a regression model. It is used to detect any patterns or systematic errors in the residuals, which may indicate a poor fit or a biased model.
Root Mean Squared Error (RMSE) #
A performance metric for regression models that measures the root of the average of the squared differences between the predicted and actual values. It is calculated as the square root of the mean squared error and is used to evaluate the accuracy of the model's predictions.
Specificity #
A performance metric for classification models that measures the proportion of true negative predictions out of all actual negative samples. It is calculated as the number of true negatives divided by the sum of true negatives and false positives.
Supervised Learning #
A type of machine learning where the goal is to learn a mapping between input variables and output variables based on labeled training data.
Training Set #
A subset of the data used to train a model. The model learns the relationship between the input variables and the output variables based on the training data.
Testing Set #
A subset of the data used to evaluate the performance of a trained model. The model's predictions on the testing data are compared to the actual values to assess the model's ability to generalize to new, unseen data.
Underfitting #
A situation where a model is too simple and fails to capture the underlying pattern in the data, resulting in poor performance on both training and testing data. Underfitting is characterized by high bias and high variance.
Variance #
A measure of the difference between the expected predictions of a model and the predictions on the training data. A high variance model is overly complex and tends to overfit the data, resulting in poor performance on new, unseen data.
Validation Curve #
A graphical representation of the performance of a model as a function of its complexity. It is used to detect any biases or variances in the model and to identify the optimal complexity for the model.
Visual Evaluation #
The process of visually inspecting the data and the model's predictions to evaluate its performance. Visual evaluation is used to detect any patterns or anomalies in the data and to assess the model's ability to capture these patterns.