ml
ml
Ans Human learning is a complex and ongoing process through which individuals acquire knowledge,
skills, attitudes, or behaviors. It occurs as a result of experience, observation, practice, study, or
teaching. Learning is not just about memorizing facts—it involves understanding, applying, and
sometimes even creating new ideas. It leads to a relatively permanent change in behavior or
thinking, meaning that what is learned can be applied in real-life situations and retained over time.
Human learning can be formal, such as in schools and colleges, where structured lessons and
assessments are used, or informal, like learning through everyday life experiences, observation, or
problem-solving. It involves cognitive, emotional, and social aspects, and each individual may learn
differently depending on their abilities, motivation, and environment.
-Learning to ride a bicycle: This is a skill-based learning where a person gains balance, coordination,
and control through repeated practice and experience. Once learned, the person can ride a bicycle
for years without forgetting, demonstrating permanent change.
-Learning a new language: This involves acquiring knowledge of vocabulary, grammar, pronunciation,
and cultural nuances through listening, speaking, reading, and writing. Over time, a person becomes
capable of communicating fluently in that language, showing both cognitive and behavioral learning.
Machine Learning is a branch of artificial intelligence (AI) that enables computers and systems to
learn from data, identify patterns, and make decisions or predictions without being explicitly
programmed. Instead of following strict rules coded by a human, a machine learning model improves
its performance over time as it is exposed to more data.
In simpler terms, machine learning allows computers to “learn from experience” much like humans
do, but using mathematical algorithms and statistical techniques.
Machine learning tasks are generally divided into the following main types:
-Supervised Learning
The model is trained on labeled data, meaning the input comes with the correct output.
Examples:
-Unsupervised Learning
The model is trained on unlabeled data and must find patterns, relationships, or structures on its
own.
Goal: Group or organize data meaningfully.
Examples:
-Reinforcement Learning
The model learns by trial and error, receiving rewards or penalties based on its actions.
Examples:
Uses a small amount of labeled data and a large amount of unlabeled data to improve learning
accuracy.
Example: Facial recognition systems trained with a few labeled images but thousands of unlabeled
ones.
✅ In short: Machine learning is about teaching machines to learn patterns from data, and its key tasks
involve prediction, classification, clustering, and decision-making.
1. **Prediction**
2. **Classification**
* **Example:** Email spam detection (spam vs. not spam), tumor type classification (benign vs.
malignant).
3. **Pattern Recognition**
* ML identifies patterns and regularities in data that may not be obvious to humans.
* ML helps make decisions by analyzing data and choosing the best possible action.
6. **Automation / Self-Improvement**
* **Example:** Google’s search algorithm improves over time by learning user behavior.
| ---------------------------------------------- | -------------------------------------------------- |
| Learn from experience and past outcomes | Predict outcomes from historical data |
| Learn through trial and error | Reinforcement learning models improve via feedback |
| Improve skills over time | ML models improve automatically with more data |
✅ **In short:** Machine learning is like teaching a computer to **think, learn, and improve** in
ways similar to humans, but using **mathematical algorithms and data** instead of the human
brain.
4. Define machine learning and explain the different elements with a real example.
Ans Definition:
Machine Learning (ML) is a branch of Artificial Intelligence (AI) that allows computers to learn from
data and experience without explicit programming. It uses algorithms to identify patterns, make
predictions, or take actions based on data.
Example: In a house price prediction model, data includes size, location, number of rooms, age of the
house, etc.
Example: In house price prediction, features are size ([Link].), number of bedrooms, and location.
Learning Algorithm – The method used to train the model and find patterns in the data.
Prediction / Output – The result provided by the trained model for new input data.
Example: For a 1200 [Link]. house in a city, the model predicts the price as $150,000.
Abstraction is the process of simplifying a complex system or problem by focusing only on the
essential details while ignoring irrelevant or less important information. In other words, it helps us
reduce complexity and deal with only the important aspects of a problem.
In computer science and machine learning, abstraction allows us to create models, algorithms, or
systems without getting overwhelmed by unnecessary details.
-Determine essential features – Decide which aspects are important for solving the problem.
-Ignore irrelevant details – Remove information that doesn’t affect the outcome.
-Create a simplified model – Use only the relevant features to represent the problem.
-Solve or analyze the problem – Work with the simplified model to find a solution.
Example of Abstraction
Complex Reality: A real environment has cars, pedestrians, traffic signals, road signs, weather
conditions, road bumps, birds, buildings, etc.
Abstraction: Focus only on the essential information needed for safe driving:
Road lanes (to stay on track)
The self-driving car ignores irrelevant details like the color of buildings, trees, or advertisements,
making the problem simpler and solvable.
Result: The car can make decisions efficiently using only the important information.
6. What is generalization? What role does it play in the process of machine learning?
Ans Generalization in machine learning refers to the ability of a model to perform well on new,
unseen data, not just the data it was trained on. When a machine learning model learns from a
dataset, it tries to identify patterns that represent the underlying relationship in the data. If the
model memorizes every detail of the training data, it may perform perfectly on that data but fail to
make accurate predictions on new data; this problem is called overfitting. On the other hand, if the
model fails to capture the important patterns, it will perform poorly even on the training data, which
is called underfitting. Generalization is achieved when the model captures the essential patterns
without memorizing noise, allowing it to make correct predictions on data it has never seen before.
For example, consider a model trained to recognize cats in images. If it has good generalization, it can
correctly identify cats in new images that were not part of its training dataset. If the model only
memorized the specific images it was trained on, it might fail when it encounters cats in different
poses, backgrounds, or lighting conditions. The ability to generalize is what makes machine learning
models useful in real-world applications, such as predicting house prices, detecting spam emails, or
recognizing speech. Without generalization, a model’s predictions would be unreliable outside the
training environment, limiting its practical value.
Ans In reinforcement learning (RL), an agent learns to make decisions by interacting with its
environment. The agent performs actions, observes the results, and receives feedback that helps it
learn which actions are good or bad. This feedback comes in the form of rewards and penalties.
A reward is a positive signal given to the agent when it takes a correct or desirable action. It
encourages the agent to repeat that action in the future. A penalty (or negative reward) is given
when the agent takes a wrong or undesirable action. It discourages the agent from repeating that
action. Over time, by trying actions, receiving rewards, and avoiding penalties, the agent learns the
best strategy to maximize its total reward.
Example: Imagine a robot learning to navigate a maze. If the robot moves closer to the exit, it
receives a reward (+10 points). If it bumps into a wall, it gets a penalty (-5 points). By repeating this
process, the robot learns the most efficient path to reach the exit while avoiding walls.
In short, rewards guide the agent toward correct actions, and penalties help it avoid mistakes,
allowing it to learn optimal behavior through trial and error.
8. What is regression? Give example of some practical problems solved using regression.
Ans Regression is a type of supervised machine learning used to predict a continuous numerical
value based on input data. Unlike classification, which predicts categories, regression predicts
quantities that can take any value within a range. The model learns patterns from historical data and
uses them to estimate future or unknown values.
Predicting house prices based on features like size, location, number of bedrooms, and age of the
house.
Forecasting stock market prices using historical price trends and economic indicators.
Predicting a person’s weight based on height, age, and other health metrics.
In all these examples, regression helps quantify relationships between variables and make
predictions that guide decisions in real-world scenarios.
9. What is classification? Explain the key differences between classification and regression.
Ans Classification is a type of supervised machine learning where the model learns from labeled data
to assign new inputs into predefined categories or classes. The output is always discrete, meaning it
belongs to one of the set of classes.
Example:
| **Evaluation Metrics** | Accuracy, Precision, Recall | Mean Squared Error (MSE), R², MAE
|
| **Question Answered** | “Which category does it belong to?” | “What is the value?”
|
Ans Clustering is an unsupervised machine learning technique that groups similar data points
together into clusters based on their characteristics, without using any labeled data. The objective is
to ensure that data points within the same cluster are more similar to each other than to those in
different clusters, enabling the discovery of natural groupings and hidden patterns in complex
datasets.
Goal: Discover the natural grouping or structure in unlabeled data without predefined categories.
How: Data points are assigned to clusters based on similarity or distance measures.
Similarity Measures: Can include Euclidean distance, cosine similarity or other metrics depending on
data type and clustering method.
Output: Each group is assigned a cluster ID, representing shared characteristics within the cluster.
Ans Machine Learning (ML) algorithms are widely used to enable systems to learn from data and
make intelligent decisions without explicit programming. Some key applications include:
Finance: Fraud detection, credit scoring, stock price prediction, risk assessment.
Overall, machine learning algorithms improve efficiency, accuracy, and automation across various
industries.
Ans Supervised learning is a type of machine learning in which a model is trained using labeled data.
Each training example consists of an input and a corresponding correct output (label). The algorithm
learns a mapping between inputs and outputs and uses this learned relationship to make predictions
on new, unseen data.
Common supervised learning tasks include classification (e.g., spam email detection, disease
diagnosis) and regression (e.g., house price prediction, sales forecasting). Popular supervised
learning algorithms include Linear Regression, Logistic Regression, Decision Trees, Support Vector
Machines (SVM), and k-Nearest Neighbors (k-NN).
Supervised learning is widely used because of its high accuracy when sufficient labeled data is
available.
Ans Unsupervised learning is a type of machine learning where the model is trained on unlabeled
data, meaning no predefined output or target is provided. The algorithm discovers hidden patterns,
structures, or relationships within the data on its own.
Common unsupervised learning tasks include clustering (e.g., customer segmentation, grouping
similar data points) and association (e.g., market basket analysis). Popular unsupervised learning
algorithms include K-Means clustering, Hierarchical clustering, DBSCAN, and Apriori algorithm.
Unsupervised learning is widely used in data exploration, pattern recognition, anomaly detection,
and feature extraction, helping organizations gain insights from large datasets.
[Link] learning is a type of machine learning in which an agent learns by interacting with
an environment. The agent takes actions and receives rewards or penalties as feedback, with the
goal of maximizing cumulative reward over time.
Unlike supervised learning, reinforcement learning does not use labeled data. Instead, learning is
based on trial and error. Common components include the agent, environment, actions, states, and
reward. Popular reinforcement learning algorithms include Q-learning, SARSA, and Deep Q-Networks
(DQN).
Reinforcement learning is widely used in robotics, game playing (e.g., AlphaGo), autonomous
vehicles, and recommendation systems, where sequential decision-making is required.
Ans Below is a detailed difference between Abstraction and Generalization, explained clearly for
exams and understanding:
Data Visibility Internal data is hidden Common data and methods are shared
Example A Vehicle interface defines start() and Car and Bike are generalized into a
Aspect Abstraction Generalization
Object-oriented design for simplicity and Object-oriented design for hierarchy and
Used In
security reuse
In Short
Common Logistic Regression, Decision Tree, Naïve Linear Regression, Polynomial Regression,
Algorithms Bayes, SVM, k-NN SVR, Decision Tree
Evaluation
Accuracy, Precision, Recall, F1-score Mean Squared Error (MSE), RMSE, R²
Metrics
In short:
18. What are the main activities involved in machine learning? What is meant by data pre-
processing?
Data Pre-processing
Data pre-processing is the process of cleaning, transforming, and organizing raw data into a suitable
format for machine learning algorithms. Real-world data often contains missing values, noise, and
inconsistencies, which can affect model accuracy.
It includes:
Data transformation
In brief:
Data pre-processing improves data quality and ensures better model performance.
19. Explain qualitative and quantitative data in details. Differentiate between the two.
Data can be broadly classified into qualitative and quantitative based on the nature of information
they represent.
Qualitative Data
Characteristics
Non-numeric
Often subjective
Examples
Eye color
Type of vehicle
Marital status
Uses
Quantitative Data
Quantitative data is numerical in nature and represents measurable quantities. It can be analyzed
using mathematical and statistical methods.
Characteristics
Numeric
Measurable
Objective
Examples
Age
Salary
Distance
Marks obtained
Uses
In Short
20. What are the different causes of data issues in machine learning? What are the fallouts?
In machine learning, the quality of data directly affects model performance. Poor or problematic
data leads to inaccurate and unreliable models.
1. Missing Data
Occurs due to data entry errors, system failures, or incomplete records.
2. Noisy Data
Data contains random errors or incorrect values caused by faulty sensors or human mistakes.
3. Outliers
Extreme values that differ significantly from most data points.
4. Inconsistent Data
Conflicting information in different sources (e.g., different formats or units).
5. Duplicate Data
Repeated records due to data integration from multiple sources.
6. Imbalanced Data
One class has significantly more samples than others.
8. Biased Data
Data does not represent real-world diversity.
2. Overfitting or Underfitting
The model learns noise or fails to learn meaningful patterns.
3. Biased Predictions
Unfair or unethical decisions.
4. Poor Generalization
Model performs well on training data but poorly on new data.
5. Increased Training Time
More computational resources required.
In Short
3. Ordinal Data
Categorical data with a meaningful order.
Example: Rating (Low, Medium, High), Education level
4. Binary Data
Data with only two possible values.
Example: Yes/No, 0/1, Pass/Fail
5. Time-Series Data
Data collected over time intervals.
Example: Daily stock prices, monthly sales
22. Differentiate
Meaning Reduces features by creating new ones Selects a subset of existing features
Before building a machine learning model, several key activities ensure the model is effective and
accurate. These activities are part of the data preparation and understanding process:
2. Data Collection
o Gather relevant data from databases, files, APIs, sensors, or online sources.
4. Data Cleaning
5. Data Pre-processing
o Normalize or scale numerical features.
o Check for imbalanced classes and biases that may affect the model.
In short:
Proper preparation before modelling ensures that the data is clean, relevant, and structured,
which is essential for building accurate, reliable, and efficient machine learning models.
unit 2
model?
ans **Model:**
**Training a Model:**
1. **Collect Dataset:** Gather relevant and sufficient data related to the problem.
2. **Preprocess Data:** Clean the data by handling missing values, removing noise, and scaling
features if required.
3. **Split the Data:** Divide the dataset into training and testing (and sometimes validation) sets.
6. **Adjust Parameters:** Optimize model parameters to minimize errors using a loss function.
7. **Evaluate the Model:** Test the trained model on unseen test data to check accuracy and
performance.
8. **Improve the Model:** If performance is low, repeat the process by tuning parameters or
improving data quality.
1. The complete dataset is divided into **K equal-sized parts**, called **folds**.
2. In the first iteration, **one fold is selected as the test set** and the remaining **K−1 folds are
used for training** the model.
3. The model is trained on the training folds and evaluated on the test fold.
4. This process is **repeated K times**, each time using a different fold as the test set.
5. Thus, **every data point is used once for testing and K−1 times for training**.
6. The performance scores (accuracy, error, etc.) from all K iterations are recorded.
7. The **final model performance** is calculated by taking the **average of all K results**.
**Benefits:**
Bootstrap Sampling:
Bootstrap sampling is a resampling technique in which random samples are drawn from the
original dataset with replacement. Because sampling is done with replacement, the same data
point can appear multiple times in a single sample, while some data points may not appear at all.
Each bootstrap sample is usually of the same size as the original dataset.
Q4 What is the main purpose of a descriptive model? State some real-world problems solved using
descriptive models.
**Purpose:**
Descriptive models are used to **analyze and describe patterns, relationships, trends, or
structures** present in a dataset. These models help in **understanding the data** rather than
predicting future outcomes. They summarize large and complex datasets into meaningful
information and provide insights that support decision-making.
**Key Points:**
4. **Clustering Students Based on Performance:** Grouping students into categories such as high,
average, and low performers.
**Purpose:**
Descriptive models are used to **analyze and describe patterns, relationships, trends, or
structures** present in a dataset. These models help in **understanding the data** rather than
predicting future outcomes. They summarize large and complex datasets into meaningful
information and provide insights that support decision-making.
**Key Points:**
4. **Clustering Students Based on Performance:** Grouping students into categories such as high,
average, and low performers.
ans
Occurs when the model is too simple to learn the Occurs when the model is too complex and
underlying patterns in the data. It shows high bias learns noise along with patterns. It shows high
and gives poor performance on both training and variance and performs poorly on new or
testing data. unseen data.
Cross-Validation Bootstrapping
Mainly used for model validation and performance Mainly used for resampling and creating
evaluation to estimate how well a model will multiple datasets, often to improve model
generalize to unseen data. stability and reduce variance.
Q7What is a target function? Express target function in context of a reallife example. How is the
fitness of a target function measured?
A target function in machine learning is a mathematical function that defines the relationship
between input features and the output value that the model is trying to learn. The main goal of a
learning algorithm is to approximate this target function as accurately as possible.
Example:
In a house price prediction problem, the target function can be written as:
House Price = f(size, location, number of rooms)
Here, size, location, and rooms are input features, and house price is the output.
2. Loss function value, which shows how far predictions are from actual values.
3. Validation performance, which checks model accuracy on unseen data to ensure good
generalization.
Q8 What are predictive models? What are descriptive models? Give examples of both types of
models. Explain the difference between these types of models.
Ans Predictive Models:
Predictive models are used to predict future or unknown values based on past data. These models
learn the relationship between input and output variables using labeled data. They are mainly
used for forecasting and decision-making.
Examples:
Weather prediction
Sales forecasting
Descriptive Models:
Descriptive models are used to understand, analyze, and summarize data. They identify patterns,
groups, or relationships in data but do not predict future outcomes. These models usually work on
unlabeled data.
Examples:
Customer clustering
Topic modeling
Difference:
Predictive models focus on prediction of future outcomes, while descriptive models focus on
understanding patterns and structure in existing data.
Q9 What is underfitting in context of machine learning models? What is the major cause of
underfitting?
Underfitting occurs when a machine learning model is too simple to capture the important
patterns, trends, and relationships present in the training data. Such a model is unable to learn
from the data properly and therefore gives poor performance on both training and test datasets.
Underfitting leads to inaccurate predictions and low overall model accuracy because the model
does not represent the data well.
1. Very simple model selection, such as using a linear model for a complex problem.
2. Insufficient training time, where the model is not allowed to learn enough.
4. High bias, meaning the model makes strong assumptions about the data.
Overfitting occurs when a machine learning model learns the training data in too much detail,
including noise, outliers, and random variations, instead of learning the true underlying patterns.
Because of this, the model shows very high accuracy on training data but performs poorly on
unseen or test data. This means the model fails to generalize well to new data.
1. When the model is too complex, such as having too many layers or parameters.
2. When the number of features is very high compared to the dataset size.
4. When the model is trained for too many iterations without proper control.
The bias–variance trade-off explains the balance between two types of errors that affect the
performance of a machine learning model: bias and variance. The goal of model fitting is to
achieve a proper balance between them to get good accuracy on both training and unseen data.
Bias refers to the error caused by overly simple models that make strong assumptions about the
data. High bias models fail to capture important patterns, leading to underfitting and poor
performance on both training and test data.
Variance refers to the error caused by overly complex models that are very sensitive to small
changes in the training data. High variance models learn noise along with patterns, leading to
overfitting and poor performance on new data.
The trade-off means that reducing bias usually increases variance, and reducing variance usually
increases bias. A good model finds the right balance, minimizing total error and providing good
generalization on unseen data.
1. Holdout method
2. 10-fold cross-validation
3. Parameter tuning
This method is fast and easy to implement, making it suitable for large datasets. However, it can be
less reliable for small datasets because the evaluation depends heavily on how the data is split.
The performance may vary if a different split is used, leading to inconsistent results.
10-Fold Cross-Validation is a more robust evaluation technique. In this method, the dataset is
divided into 10 equal parts (folds). The model is trained on 9 folds and tested on the remaining
fold. This process is repeated 10 times, with each fold used once as the test set.
The final performance is calculated as the average of all 10 results, making it more accurate than
the holdout method. It is especially useful for small datasets because it ensures that every data
point is used for both training and testing, reducing bias and variance in model evaluation.
Parameter tuning involves adjusting the hyperparameters of a machine learning model to improve
its performance. Hyperparameters are settings that control the learning process, such as learning
rate, number of trees in a random forest, or the number of neighbors in KNN.
Tuning is typically done using methods like Grid Search or Random Search, which systematically
test different combinations of hyperparameters. Proper parameter tuning can significantly improve
model accuracy, prevent overfitting, and help the model generalize better on unseen data.
Bagging Boosting
Works best when the base model is unstable (e.g., Works best with weak learners to convert them
decision trees). into a strong learner.
Bagging Boosting
Summary: Bagging focuses on parallel model training to reduce variance, while Boosting focuses
on sequential training to reduce bias and improve accuracy.
Learns only at the time of prediction. It stores the Learns a general model during training and
training data and waits until a query is made. uses it to make predictions later.
Training is fast, but prediction can be slow because Training is slower, but prediction is fast
computation happens at query time. because the model is already built.
Often used for instance-based learning. Often used for model-based learning.
Works well with small datasets but can be slow for Scales better for large datasets and can
large datasets. generalize well.
Summary: Lazy learners delay learning until prediction, while eager learners build a model in
advance to make faster predictions.
A feature is an individual measurable property, characteristic, or attribute of the data that is used
as an input for a machine learning model. Features play a key role in helping the model understand
patterns, relationships, and trends in the data so it can make accurate predictions or classifications.
Choosing the right features and representing them properly is essential for building an effective
model.
Example:
In a house price prediction problem, the input features could include:
Size of the house (in sq. ft.) – larger houses usually cost more.
Nominal variables are categorical variables that represent names, labels, or categories without any
intrinsic order. Machine learning models require numerical input, so encoding nominal variables is
the process of converting these categories into numbers.
Benefits:
Example:
Original data:
Color
Red
Blue
Green
Color
1 0 0
0 1 0
0 0 1
This numeric representation can now be used in any machine learning algorithm.
Sometimes, numeric features need to be converted into categorical features to make them easier
to interpret or to improve model performance, especially when the numeric values fall into distinct
ranges or groups. This process is also called discretization or binning.
Process:
Benefits:
Improves interpretability
Example:
8 Child
17 Teen
Age Age Group
25 Adult
65 Senior
Q18 Explain the wrapper approach of feature selection. What are the merits and de-merits of this
approach?
The wrapper approach is a method of selecting the best subset of features by evaluating different
combinations of features using a machine learning model. It treats the feature selection process as
a search problem, where various subsets are tested, and the model’s performance determines
which features are selected.
Process:
Merits:
Often provides high accuracy because selection is based on actual model performance.
Demerits:
This method is best for small to medium datasets where accuracy is more important than
computation time.
Q19 What is feature engineering? Explain, in details, the different aspects of feature engineering?
Feature Engineering is the process of creating, modifying, and improving features in a dataset to
help machine learning models learn better and make more accurate predictions.
Key Aspects:
1. Feature Creation: Generate new features from existing data.
Example: From Date of Birth, create Age.
Summary: Feature engineering transforms raw data into meaningful features, improving model
accuracy, interpretability, and training efficiency.
Q20 What is feature selection? Why is it needed? What are the different approaches of feature
selection?
Feature Selection is the process of choosing the most relevant features from a dataset that
contribute the most to a machine learning model’s accuracy, while removing irrelevant or
redundant features.
Why is it Needed?
1. Filter Approach:
2. Wrapper Approach:
o Evaluates different subsets of features using a specific model and selects the best-
performing subset.
3. Embedded Approach:
o Feature selection occurs during model training, e.g., Lasso Regression or Decision
Trees automatically select important features.
Summary: Feature selection helps build efficient, accurate, and interpretable models by using only
the most useful features.
Q21 Explain the filter and wrapper approaches of feature selection. What are the merits and
demerits of these approaches?
1. Filter Approach:
Merits:
Demerits:
2. Wrapper Approach:
Selects features by evaluating different subsets using a machine learning model and
choosing the subset that gives the best performance.
Merits:
Demerits:
Summary:
1. Data Collection:
o Gather raw data from various sources that include all potential features.
2. Preprocessing:
3. Feature Evaluation:
o Assess the importance of each feature using statistical measures (Filter), model
performance (Wrapper), or embedded methods.
o Choose the best subset of features that maximizes model performance while
reducing redundancy.
o Train the model using the selected features and evaluate performance on
validation/test data.
o Refine the feature subset if needed to achieve better accuracy, reduce overfitting,
or improve interpretability.
Summary:
Feature selection ensures that the machine learning model is efficient, accurate, and easier to
interpret by focusing only on the most meaningful features.
Q23 Explain, with an example, the main underlying concept of feature extraction. What are the
most popular algorithms for feature extraction?
Feature Extraction is the process of transforming the original features of a dataset into a new set of
features that are more informative and easier for a machine learning model to use. Unlike feature
selection, which picks existing features, feature extraction creates new features by combining or
reducing the original ones. This helps in reducing dimensionality and improving model
performance.
Example:
In an image recognition problem, an image may have thousands of pixel values as features.
Feature extraction can transform these into a smaller set of important features like edges, shapes,
or textures that capture the most relevant information. This reduces complexity while retaining
meaningful data for classification.
2. Linear Discriminant Analysis (LDA): Finds features that best separate classes.
Summary:
Feature extraction transforms raw data into informative, reduced, and meaningful features,
making models faster, simpler, and more accurate.