Chapter 2
Chapter 2
College of Science
Data Science Department
Course Title: Machine Learning
By
C1
• Instances/feature vector C2
41
3.6
63
1.2
1.5
2
4
1
0 3.5
• Entity with certain attribute values C1 109 0.4 6 1 2.4
Examples:
•Spam Detection: Classifies emails as spam or not spam.
•House Price Prediction: Estimates house prices based on features like
size and location.
•Medical Diagnosis: Identifies diseases from symptoms and patient data.
Steps Involved in Supervised Learning
1
5
Single label classification
• Single label classification: learning from a set of data points that are
associated with only one target label from a set of disjoint
labels Y , with |Y| ≥ 2.
• If |Y| = 2, the learning problem is called binary classification
e.g disease diagnosis, spam, malware detection, etc.
0.6 0.6 0 0 1
Advantages
0.4 0.1 0 1 0 • Computationally efficient
Disadvantages
0.5 0.2 1 0 1
•Does not capture the dependence relations among the
class variables
9
Label Powerset Method
• Transform each label combination to a class value and then learn a
multi-class classifier with the new class values
X1 X2 Y1 Y2 Y3 X1 X2
Ycomb
• X-->Ycomb
0.3 0.7 1 1 0 0.3 0.7 1
Advantages
• Learns the full joint of the class variables and each of the new class values maps to a label
combination
Disadvantages
• The number
classifier of choices inchoices
on exponential the newisclass can be exponential (|YLP | = O(2d )) and Learning
1 a
expensive
multi-class 0
Classifier Chains
• Similar to Binary relevance, doing multiple single-label problem but we use the previous prediction as
input feature vector.
X1 X2 Y1 X1 X2 Y1 Y2 X1 X2 Y1 Y2 Y3
Y1 as a class
Y1 as a feature and Y2 as a class Y1,Y2 as feature and Y3 as a
class
Limitation: The result can vary for different order of chains Solution:
1
ensemble 1
Classification-A Two-Step Process
•Model construction (Training phase): describing a set of predetermined classes
• Model construction (Training phase): describing a set of predetermined classes
• Each tuple/sample is assumed to belong to a predefined class, as determined by the
class label attribute
• The set of tuples used for model construction is training set
• The model is represented as classification rules, decision trees, or math formulae
•Model usage (Testing phase): for classifying future or unknown objects
• Estimate accuracy of the model 22
• The known label of test sample is compared with the classified result from the model
• Accuracy rate is the percentage of test set samples that are correctly classified by the model
• Test set is independent of training set (otherwise overfitting) If the accuracy is acceptable,
use the model to classify new data
• Note: If the test set is used to select models, it is called validation (test) set.
Common Algorithms in Supervised Learning
K-Nearest-Neighbor
Decision Tree
Random Forest
Support Vector Machine (SVM)
Anomaly Detection
Linear Regression with Regularization
Logistic Regression with Regularization
K-Nearest-Neighbor
•The K-Nearest Neighbors (KNN) algorithm is a simple
yet powerful supervised machine-learning technique
used for classification tasks.
•It works based on the idea that similar data points are
often near each other in feature space.
•KNN is like the "ask your neighbors" rule in real life.
When you need to make a decision or guess something,
you check what the closest people (or examples) around
you are doing and follow the majority.
•It’s a simple machine learning algorithm that looks at
who’s closest to you to make a prediction. 11
22
Applications of KNN
Training
Records
Choose k of the
nearest records
11
22
K-Nearest-Neighbor
Real-Life Example: Fruit Identification
Imagine you find a fruit and don’t know what it is. You compare
it with nearby fruits based on: Size, Color, Weight
If it looks similar to apples near you, you call it an apple!
For Example: you find a fruit that is: Size: 7.5 cm, Color: Red,
Weight: 155 g
You compare it to the data below. It’s closest to the apples, so
you classify
No Size(cm) it as an
Color Apple.
Weight (g) Label
1 10 Red 190 tomato
2 6 Yellow 120 Banna
3 8 Red 160 Apple
4 12 Yellow 250 Orange
5 9 Red 165 Apple
11
22
Key Concepts of KNN
1. Instance-based learning: Predictions are made based on
the similarity of a new data point to existing instances.
* KNN does not explicitly learn a model but memorizes the
training dataset. *
2. Distance Metric: KNN relies on measuring the distance
between data points. Common distance metrics include:
• Euclidean distance, * Manhattan distance, * Minkowski
distance
3. Number of Neighbors (K):
The parameter K determines how many nearest neighbors
are considered for classification or regression.
Small K may lead to noisy predictions (overfitting), while
large K may oversimplify the model (underfitting). 1
2
1
2
Nearest Neighbor Classification
28
Steps of K-NN Algorithm
k-nearest neighbors algorithm steps.
Step 2. Calculate the distance between the query-instance and all the training
samples
Step 3. Sort the distance and determine nearest neighbors based on the k-th
minimum distance
7 7 Bad
7 4 Bad
3 4 Good
1 4 Good
30
k-NN Example
Step2. Calculate the distance between the query-instance and all the
training examples
31
k-NN Example
Step 3. Sort the distance and determine nearest neighbors based on the
k-th minimum distance
Acid Strengt Square distance to Rank Is it included
Durability h (X2) query instance (3,7) minimum in 3- nearest
(X1 ) distance neighbors
7 7 (7-3)2 + (7-7)2 =16 3 Yes
33
k-NN Example
Step 5. Use simple majority of the category of nearest neighbors as
the prediction value of the query instance.
34
K-Nearest-Neighbor
Advantages of KNN
Simple to understand and implement.
No assumptions about the underlying data distribution.
Effective for small datasets with well-separated classes.
Versatile, applicable to both classification and regression.
Disadvantages of KNN
Computationally expensive during prediction since it requires calculating distances
for all training data points.
Memory-intensive as it requires storing the entire training set.
Sensitive to irrelevant or noisy features.
Requires careful selection of K and the distance metric. 1 1
22
Common Algorithms in Supervised Learning
K-Nearest-Neighbor
Decision Tree
Random Forest
Support Vector Machine (SVM)
Anomaly Detection
Linear Regression with Regularization
Logistic Regression with Regularization
Decision Tree
Decision Tree
• Root node: has no incoming edges and zero or more outgoing edges.
• Internal nodes: each of which has exactly one incoming edge and two or
more outgoing edges.
• Leaf or terminal node, each of which has exactly one incoming and no
outgoing edges.
22
88
Decision Tree Example
Output: A Decision Tree for
“buys_computer”
Training Dataset Decision Tree
age income student credit_rati buys_compute
ng r age?
<=30 high no fair no
<=30 high no excellent no
31…40 high no fair yes <=30 overcast
30..40 >40
>40 medium no fair yes
>40 low yes fair yes
>40 low yes excellent no student? yes credit rating?
31…40 low yes excellent yes
<=30 medium no fair no
<=30 low yes fair yes no yes excellent
>40 medium yes fair yes
<=30 medium yes excellent yes
31…40 medium no excellent yes no yes no yes
31…40 high yes fair yes
>40 medium no excellent no
Decision Tree
Algorithm:
• The algorithm is a greedy algorithm that doesn’t warranty optimality
• The algorithm for decision tree involves three general phases as stated below:
• Phase I: Find Splitting Criteria based on all the sample set at the splitting point (node)
• Phase II: Split all the sample data based on the splitting criteria and form branches and each
successor nodes will have the samples
• Phase III: Do phase one and phase two iteratively until stopping criterion get fulfilled
It computes the linear relationship between the dependent variable and one or more
independent features by fitting a linear equation with observed data.
It predicts the continuous output variables based on the independent input variable.
For Example: if we want to predict house price we consider various factor such as
house age, distance from the main road, location, area and number of room,
linear regression uses all these parameter to predict house price as it consider a
linear relation between all these features and price of house.
Why Linear Regression is Important?
Simplicity & Interpretability – Linear regression provides a straightforward way to
understand r/ship b/n dependent & independent variables, making it easy to interpret.
Predictive Modeling – It helps in forecasting and predicting outcomes based on
historical data, such as predicting sales based on advertising spending.
Identifying Trends & Relationships – It helps in understanding the impact of one or
more independent variables on a dependent variable (e.g., how temperature affects ice
cream sales).
Feature Importance – It helps determine which variables (features) have the most
influence on the target variable.
Baseline Model for Comparison – In machine learning, linear regression serves as a
benchmark model to compare with more complex models.
Efficiency and Low Computational Cost – It is computationally inexpensive and
works well with small to moderately large datasets.
Widespread Applications – Used in finance, healthcare, economics, marketing, and
many other fields for decision-making and analysis.
Types of Linear Regression
Simple Linear Regression (Univariate Linear Regression):
with one independent variable
Multiple Linear Regression (Multivariate Regression)
with multiple independent variables
Simple Linear Regression
Simple linear regression is the simplest form of linear regression and it involves only
one independent variable and one dependent variable.
The relationship is modeled using a straight line.
The equation for simple linear regression is: y=β0+β1X
where:
• Y is the dependent variable
• X is the independent variable
• β0 is the intercept
• β1 is the slope
Assumptions of Simple Linear Regression
Linearity: independent & dependent variables have a linear r/ship with one another.
This implies that changes in the dependent variable follow those in the independent
variable(s) in a linear fashion. This means that there should be a straight line that can
be drawn through the data points.
If the relationship is not linear, then linear regression will not be an accurate model.
Assumptions of Simple Linear Regression
Homoscedasticity: Across all levels of the independent variable(s), the variance of
the errors is constant.
This indicates that the amount of the independent variable(s) has no impact on the
variance of the errors.
If the variance of the residuals is not constant, then linear regression will not be an
accurate model.
Assumptions of Simple Linear Regression
Independence: The observations in the dataset are independent of each other.
This means that the value of the dependent variable for one observation does not
depend on the value of the dependent variable for another observation.
If the observations are not independent, then linear regression will not be an accurate
model.
Normality: The residuals should be normally distributed.
This means that the residuals should follow a bell-shaped curve.
If the residuals are not normally distributed, then linear regression will not be an
accurate model.
Multiple Linear Regression
Multiple linear regression involves more than one independent variable and one
dependent variable.
The equation for multiple linear regression is. y=β0+β1X1+β2X2+
………βnXn
where:
• Y is the dependent variable
• X1, X2, …, Xn are the independent variables
• β0 is the intercept
• β1, β2, …, βn are the slopes
The goal of the algorithm is to find the best Fit Line equation that can predict the
values based on the independent variables.
Assumptions of Multiple Linear Regression
For Multiple Linear Regression, all four of the assumptions from Simple Linear
Regression apply. In addition to this, below are few more:
No multicollinearity: There is no high correlation b/n the independent variables.
This indicates that there is little or no correlation between the independent variables.
Multicollinearity occurs when two or more independent variables are highly
correlated with each other, which can make it difficult to determine the individual
effect of each variable on the dependent variable.
If there is multicollinearity, then multiple linear regression will not be an accurate
model.
Additivity: The model assumes that the effect of changes in a predictor variable on
the response variable is consistent regardless of the values of the other variables.
This assumption implies that there is no interaction between variables in their effects
on the dependent variable.
Assumptions of Multiple Linear Regression
Feature Selection: In multiple linear regression, it is essential to carefully select the
independent variables that will be included in the model.
Including irrelevant or redundant variables may lead to overfitting and complicate
the interpretation of the model.
Overfitting: Overfitting occurs when the model fits the training data too closely,
capturing noise or random fluctuations that do not represent the true underlying
relationship between variables.
This can lead to poor generalization performance on new, unseen data.
Multicollinearity
Multicollinearity is a statistical phenomenon where two or more independent variables
in a multiple regression model are highly correlated, making it difficult to assess the
individual effects of each variable on the dependent variable.
Detecting Multicollinearity includes two techniques:
• Correlation Matrix: High correlations (near 1 or -1) potential multicollinearity.
• VIF (Variance Inflation Factor): A VIF above 10 suggests multicollinearity.
Use Case of Multiple Linear Regression
Real Estate Pricing: In real estate MLR is used to predict property prices based on
multiple factors such as location, size, number of bedrooms, etc. This helps buyers
and sellers understand market trends and set competitive prices.
Financial Forecasting: Financial analysts use MLR to predict stock prices or
economic indicators based on multiple influencing factors such as interest rates,
inflation rates and market trends. This enables better investment strategies and risk
management24.
Agricultural Yield Prediction: Farmers can use MLR to estimate crop yields based
on several variables like rainfall, temperature, soil quality and fertilizer usage. This
information helps in planning agricultural practices for optimal productivity
E-commerce Sales Analysis: An e-commerce company can utilize MLR to assess
how various factors such as product price, marketing promotions and seasonal trends
impact sales.
common Algorithms in Supervised Learning
K-Nearest-Neighbor
Decision Tree
Random Forest
Support Vector Machine (SVM)
Linear Regression with Regularization
Logistic Regression with Regularization
Anomaly Detection
what is Logistic Regression?
Logistic regression is a supervised ML algorithm used for classification tasks where
the goal is to predict the probability that an instance belongs to a given class or not.
Logistic regression is a statistical algorithm which analyze the relationship between
two data factors.
Logistic regression is used for binary classification where we use sigmoid function,
that takes input as independent variables and produces a probability value b/n 0 & 1.
For example, we have two classes Class 0 and Class 1 if the value of the logistic
function for an input is greater than 0.5 (threshold value) then it belongs to Class 1
otherwise it belongs to Class 0. It’s referred to as regression because it is the
extension of linear regression but is mainly used for classification problems.
Key Point of Logistic Regression?
Logistic regression predicts the output of a categorical dependent variable. Therefore,
the outcome must be a categorical or discrete value.
It can be either Yes or No, 0 or 1, true or False, etc. but instead of giving the exact
value as 0 and 1, it gives the probabilistic values which lie between 0 and 1.
In Logistic regression, instead of fitting a regression line, we fit an “S” shaped
logistic function, which predicts two maximum values (0 or 1).
Types of Logistic Regression
On the basis of the categories, Logistic Regression can be classified into three types:
• Binomial: In binomial Logistic regression, there can be only two possible types of
the dependent variables, such as 0 or 1, Pass or Fail, etc.
• Multinomial: In multinomial Logistic regression, there can be 3 or more possible
unordered types of the dependent variable, such as “cat”, “dogs”, or “sheep”
• Ordinal: In ordinal Logistic regression, there can be 3 or more possible ordered types
of dependent variables, such as “low”, “Medium”, or “High”.
Assumptions of Logistic Regression
Independent observations: Each observation is independent of the other. meaning
there is no correlation between any input variables.
Binary dependent variables: It takes the assumption that the dependent variable
must be binary or dichotomous, meaning it can take only two values. For more than
two categories SoftMax functions are used.
Linearity r/ship b/n independent variables and log odds: The relationship between
the independent variables and the log odds of the dependent variable should be linear.
No outliers: There should be no outliers in the dataset.
Large sample size: The sample size is sufficiently large
Differences Between Linear and Logistic
Linear Regression
Regression
Logistic Regression
It is used to predict the continuous dependent variable itis used to predict the categorical dependent
using a given set of independent variables. variable using a given set of independent variables.
Linear regression is used for solving regression
It is used for solving classification problems.
problem.
In this we predict the value of continuous variables In this we predict values of categorical variables
In this we find best fit line. In this we find S-Curve.
Least square estimation method is used for estimation Maximum likelihood estimation method is used for
of accuracy. Estimation of accuracy.
The output must be continuous value, such as price, Output must be categorical value such as 0 or 1,
age, etc. Yes or no, etc.
It required linear r/ship b/n dependent & independent. It not required linear relationship.
There may be collinearity between the independent There should be little to no collinearity between
variables. independent variables.
common Algorithms in Supervised Learning
K-Nearest-Neighbor
Decision Tree
Random Forest
Support Vector Machine (SVM)
Linear Regression with Regularization
Logistic Regression with Regularization
Anomaly Detection
Anomaly Detection
What is Anomaly Detection?
Anomaly Detection
• Anomaly Detection (Outlier Detection) is a technique in data analysis and machine
learning that identifies data points, events, or observations that significantly deviate
from the normal pattern of a dataset.
• Anomaly Detection is critical in lots of fields, which includes finance for detecting
fraud transactions, manufacturing finds defects, healthcare for odd clinical
conditions, and cybersecurity for detecting protection breaches or threats.
• These irregular findings often signal major troubles. Think mistakes, wrongdoing, or
unauthorized access.
Types of Anomalies
1. Individual Point Anomalist: point anomaly occurs when a single data point deviates
significantly from the overall distribution.
• Example: In the case of the credit card transaction analysis, a point anomaly may be the
transaction that has this value significantly bigger than any other average values recorded
for that account and potential fraud.
3. Collective Anomalies: Consolidated anomalies occur when a group of data points seems
normal individually but forms an outlier collectively, often seen in sequential data like
telecom and healthcare monitoring systems.
Why is Anomaly Detection Important?
Early Issue and Threat Detection: Anomaly detection helps identify potential problems
and risks early, preventing significant damage. In cybersecurity, detecting unusual network
traffic can signal a breach, enabling proactive prevention of data theft.
Fraud Prevention: In finance helps identify and prevent fraudulent transactions by
spotting deviations from a user's normal behavior, protecting assets and saving millions.
Quality Control & Maintenance: In manufacturing, anomaly detection ensures product
quality and enables predictive maintenance by identifying defects and abnormal equipment
behavior, reducing downtime and costs.
Healthcare Monitoring: Anomaly detection helps track patient health by identifying
abnormal vital signs, enabling early detection of potential issues or deterioration.
Improving the Customer Experience: Companies employ anomaly detection to track
service performance and user interactions.
Enhanced Security: Anomaly detection boosts safety and security by identifying suspicious
activities or behaviors, complementing cybersecurity and fraud prevention efforts.
Anomaly Detection Use Cases
1. Fraud Detection:
Banking and Finance: Identifies potential fraud by automatically flagging unusual
activities such as large amounts, foreign transactions, or rapid sequences of transactions.
Insurance: Flags suspicious claims, such as when damage seems inconsistent with
reported losses or multiple claims are made for the same incident.
2. Intrusion Detection (Cybersecurity):
• Network Security: Monitors network traffic for unusual events such as DoS attacks,
phishing, or malware spreading, based on deviations from normal traffic patterns.
• System Security: Tracks system operations and alerts on suspicious activities, like
unauthorized access or abnormal access patterns.
3. Industrial Anomaly Detection:
Manufacturing Processes: Monitors production lines to identify and remove defective or
non-compliant products, preventing supply chain issues.
Oil and Gas: Tracks infrastructure and machinery with sensor data to detect failures or
safety risks early.
Model performance evaluation, diagnostics &
predictions
Regularization and Bias/Variance
Evaluating hypothesis
Model selection (Train/ Test / Validation)
Learning Curves
MSE, lift, AUC, Type 1 vs 2
Regularization Techniques in Machine
Regularization is a technique usedLearning
to prevent overfitting by adding a penalty term to
the model's objective function during training.
The objective is to discourage the model from fitting the training data too closely and
promote simpler models that generalize better to unseen data.
which provides a methodical way to avoid overfitting and enhance the capacity of
machine learning models to generalize.
Overfitting occurs when a model captures noise and patterns specific to the training
data, leading to poor generalization on unseen data.
Underfitting arises when a model is too simple to capture the underlying patterns in
the data, resulting in poor performance on both the training and testing datasets.
• Testing set is usually a properly organized dataset having all kinds of data for scenarios
that the model would probably be facing when used in the real world.
• Often the validation and testing set combined is used as a testing set which is not
considered a good practice.
• If the accuracy of the model on training data is greater than that on testing data then the
model is said to have overfitting.
• This data is approximately 20-25% of the total data available for the project.
Validation Set
• The validation set is used to fine-tune the hyperparameters of the model and is considered
a part of the training of the model.
• The model only sees this data for evaluation but does not learn from this data, providing
an objective unbiased evaluation of the model.
• Validation dataset can be utilized for regression as well by interrupting training of model
when loss of validation dataset becomes greater than loss of training dataset .i.e. reducing
bias and variance.
• This data is approximately 10-15% of the total data available for the project but this can
change depending upon the number of hyperparameters .i.e. if model has quite many
hyperparameters then using large validation set will give better results.
• Now, whenever the accuracy of model on validation data is greater than that on training
data then the model is said to have generalized well.
Cross-Validation
• Cross-validation is a statistical method used in machine learning to evaluate how well a
model performs on an independent data set.
• It involves dividing the available data into multiple folds or subsets, using one of these
folds as a validation set and training the model on the remaining folds.
• This process is repeated multiple times each time using a different fold as the validation
set.
• Finally, the results from each validation step are averaged to produce a more robust
estimate of the model’s performance.
• The main purpose of cross validation is to prevent overfitting which occurs when a model
is trained too well on the training data and performs poorly on new, unseen data.
• By evaluating the model on multiple validation sets, cross validation provides a more
realistic estimate of the model’s generalization performance i.e. its ability to perform well
on new, unseen data.
• If you want to make sure your machine learning model is not just memorizing the training
data but is capable of adapting to real-world data cross-validation is a commonly used
technique.
Types of Cross-Validation
• Holdout Validation splits the dataset into two equal parts: 50% for training and 50% for
testing. It’s quick but may miss important data in the unused half, leading to higher bias.
• LOOCV(Leave-One-Out Cross Validation) trains the model on all but one data point and
tests on the omitted point, repeating for each data point. It uses all data, minimizing bias,
but can result in higher variation and longer execution time due to testing each data point.
• Stratified Cross-Validation ensures that each fold of the dataset has the same class
distribution as the entire dataset. It’s useful for imbalanced datasets, ensuring balanced
representation in each fold, which helps in classification tasks.
• K-Fold Cross Validation splits the dataset into k subsets. The model is trained on k-1
subsets, with one reserved for testing. This process is repeated k times, using each subset
for testing once.
Model performance evaluation, diagnostics &
predictions
Regularization and Bias/Variance
Evaluating hypothesis
Model selection (Train/ Test / Validation)
Learning Curves
MSE, lift, AUC, Type 1 vs 2
Learning Curves
• Machine learning models are employed to learn patterns in data.
• The best models can generalize well when faced with instances that were not part of the
initial training data.
• During the research phase, several experiments are conducted to find the solution that
best solves the business's problem, and reduces the error being made by the model.
• An error may be defined as the difference between the prediction of observation and the
true value of the observation.
• There are two major causes for errors in machine learning models:
• Bias describes a model which makes simplified assumptions so the target function is
easier to approximate; a model may learn that every 5'9 male in the world wears a size
medium top - this is clearly biased.
• Variance describes the variability in the model prediction; how much the prediction of the
model changes when we change the data used to train it.
Model performance evaluation, diagnostics &
predictions
Regularization and Bias/Variance
Evaluating hypothesis
Model selection (Train/ Test / Validation)
Learning Curves
MSE, lift, AUC, Type 1 vs 2
MSE, lift, AUC, Type 1 vs 2
MSE (Mean Squared Error)
• Definition: MSE is a metric used to evaluate the accuracy of a regression model. It
measures the average squared difference between the predicted and actual values.
• Purpose: It helps to quantify how well a regression model fits the data.
• Goal: Minimize MSE to achieve a better-fitting model, as lower MSE indicates higher
accuracy in predictions.
Lift
Definition: Lift is a metric used in classification models to measure how much better a
model performs compared to random guessing. It is commonly used in marketing and
sales to evaluate targeting strategies.
Purpose: To measure the effectiveness of a classification model in identifying true positive
outcomes.
Goal: A higher lift means the model is better at predicting the target variable, particularly
in situations with imbalanced datasets.
MSE, lift, AUC, Type 1 vs 2
AUC (Area Under the Curve)
• Definition: AUC is the area under the ROC curve (Receiver Operating Characteristic
curve) and is used to evaluate the performance of a binary classification model. It
measures the ability of the model to distinguish between positive and negative classes.
• Purpose: It provides a single value that reflects the overall performance of a classifier
across all possible classification thresholds.
• Goal: Maximize AUC (closer to 1) to achieve a model with better discriminative ability.
• Interpretation:
• AUC = 0.5: No discrimination (random classifier)
• AUC = 1.0: Perfect classification
• AUC < 0.5: Worse than random guessing.
Type 1 vs. Type 2 Errors
• Type 1 Error (False Positive): Incorrectly rejecting the null hypothesis when it is true. It
occurs when the model incorrectly predicts a positive outcome (e.g., predicting a disease
when the person is healthy).
• Type 2 Error (False Negative): Failing to reject the null hypothesis when it is false. It
occurs when the model incorrectly predicts a negative outcome (e.g., predicting no disease
when the person is actually sick).
• Purpose: To understand the trade-offs between false positives and false negatives in model
predictions, especially in classification tasks.
Goal:
• Minimize Type 1 Errors when the cost of false positives is high (e.g., unnecessary medical
treatments).
• Minimize Type 2 Errors when the cost of false negatives is high (e.g., missing a critical
diagnosis).
• Relationship: Type 1 and Type 2 errors are inversely related. Reducing one often increases
the other, so finding a balance depending on the problem context is important.
Other Related Topics
Enkutatash Tech
THANK TOU
Wisdom is the essence of our
uniqueness!
Adane Kasie Chekole
+251-938427723
adanekasie26@gmail.com
www.enkutatashplc.com