Scaling techniques in Machine Learning
Last Updated :
04 Dec, 2021
Definition: Scaling is a technique of generating an endless sequence of values, upon which the measured objects are placed. Several scaling techniques are employed to review the connection between the objects.
Following are the two categories under scaling techniques:

Comparative scales:
It involves the direct comparison of objects. Comparative scale data must be interpreted in corresponding terms and have either ordinal or rank order properties.
Types of comparative scales are:
1. Paired comparison:
- This technique is a widely used comparative scaling technique.
- In this technique, the respondent is asked to pick one object among the two objects with the help of some criterion.
- The respondent makes a series of judgements between objects.
- The data obtained is ordinal in nature.
- With n brands, [n(n-1)/2] paired comparisons are required.
For example: A survey was conducted to find out consumer's preference for dark chocolate or white chocolate. The outcome was as follows:
Dark chocolate= 30%
White chocolate= 70%
Thus, it is visible that consumers prefer white chocolate over dark chocolate.
2. Rank order:
- In this technique, the respondent judges one item against others.
- Respondent are present with several objects and are asked to rank or order them according to some criterion.
- Rank order scaling is also ordinal in nature.
- Only (n-1) scaling decisions need to be made in this technique.
For example: A respondent is asked to rate the following soft drinks:
Drinks | Rank |
Pepsi | 2 |
Thumbs Up | 1 |
Mountain dew | 3 |
Mirinda | 4 |
3. Constant sum scaling:
- In this technique, the respondent is assigned with the constant sum of units, such as 100 points to attributes of a product to reflect their importance.
- If the attribute is not important, the respondent assigns it 0 or no points.
- If an attribute is twice as important as another attribute, it receives twice as many points.
- The sum of all points is 100, that is, constant. Hence, the name of the scale.
For example: A respondent is asked to spend 500 rupees on product A, product B and product C of foods?
Products | Money |
Product A | 250 |
Product B | 150 |
Product C | 100 |
Total | 500 |
4. Q sort:
- It is a sophisticated form of rank order.
- In this technique, a set of objects is given to an individual to sort into piles to specified rating categories.
For example: A respondent is given 10 brands of shampoos and asked to place them in 2 piles, ranging from "most preferred" to "least preferred".
pile 1
Most preferred
Clinic plus | Head n shoulder | dove | L'Oreal paris | pantene |
pile 2
Least preferred
TRESemme | Biotique | Sunsilk | Himalya | Patanjali |
Note: Generally the most preferred shampoo is placed on the top while the least preferred at the bottom.
Non-comparative scales:
In non-comparative scales, each object of the stimulus set is scaled independently of the others. The resulting data are generally assumed to be ratio scaled.
Types of Non-comparative scales are:
1. Continuous rating scales:
- It is a graphic continuum typically coordinated by two extremes.
- The extreme values are not predefined.
- It can be constructed easily and is simple to use.
- The respondent rates by placing the mark on a continuous line.
For example: A respondent is asked to rate the service of Domino's:
Type 1
Following are the two categories under scaling techniques:

Type 2
2. Itemised rating scales:
- It is a graphic continuum typically coordinated by two extremes.
- It is simple to use and can be constructed easily.
- The respondent is provided with a scale that has a number or brief description associated with each category.
- The categories are ordered in terms of scale position, and therefore the respondents are required to pick the required category that best describes the object being rated.
The different forms of Itemised rating scales are - a. Itemised graphic scale, b. Itemised verbal scale, c. Itemised numeric scale
Types of Itemised rating scaling are:
a. Likert scale:
- This scale requires the respondent to indicate a degree of agreement or disagreement with the statements mentions on the left side of the object.
- The analysis is often conducted on an item-by-item basis, or a total score can be calculated.
- When arriving at a total score, the categories assigned to the negative statements by the respondent is scored by reversing the scale.
For example: A well-known shampoo brand carried out Likert scaling technique to find the agreement or disagreement for ayurvedic shampoo.
Statement
| Strongly disagree
| Disagree
| Neither agree nor disagree
| Agree
| Strongly agree
|
Ayurvedic shampoo helps in maintaining hair
| 1
| 2
| 3
| 4
| 5
|
Ayurvedic shampoo damage hair
| 5
| 4
| 3
| 2
| 1
|
Ayurvedic shampoo cleans your hair
| 5
| 4
| 3
| 2
| 1
|
b. Semantic differential scale:
- The semantic differential is a 7 point rating scale with endpoints related to bipolar labels.
- The negative words or phrase sometimes appears on the left side or sometimes right side.
- This controls the tendency of the respondents, particularly those with very positive and very negative attitudes, to mark the right or left sides without reading the labels.
- Individual items on a semantic differential scale could also be scored on either a -3 to +3 or 1 to 7 scale.
For example: A well-known shoe brand carried out semantic differential scaling technique to find out customer's opinion towards their product.
c. Staple scale:
- It is a unipolar rating scale with 10 categories scaled from -5 to +5.
- It does not have a neutral point, that is, zero.
- It is represented vertically.
For example: A well-known shoe brand carried out a staple scaling technique to find out costumer's opinion towards their product.
Similar Reads
Regularization Techniques in Machine Learning
Overfitting is a major concern in the field of machine learning, as models aim to extract complex patterns from data. When a model learns to commit the training data to memory instead of making good generalizations to new data, this is known as overfitting. The model may perform poorly as a result w
10 min read
Quantiles in Machine Learning
Quantiles offers valuable insights into data distribution and helping in various aspects of analysis. This article describes quantiles, looks at how to calculate them, and talks about how important they are for machine learning applications. We also discuss the problems with quantiles and how box pl
6 min read
Large scale Machine Learning
Large-scale machine learning (LML) aims to efficiently learn patterns from big data with comparable performance to traditional machine learning approaches. This article explores the core aspects of LML, including its definition, importance, challenges, and strategies to address these challenges.What
5 min read
Statistics For Machine Learning
Machine Learning Statistics: In the field of machine learning (ML), statistics plays a pivotal role in extracting meaningful insights from data to make informed decisions. Statistics provides the foundation upon which various ML algorithms are built, enabling the analysis, interpretation, and predic
7 min read
Maths for Machine Learning
Mathematics is the foundation of machine learning. Math concepts plays a crucial role in understanding how models learn from data and optimizing their performance. Before diving into machine learning algorithms, it's important to familiarize yourself with foundational topics, like Statistics, Probab
5 min read
Newton's method in Machine Learning
Optimization algorithms are essential tools across various fields, ranging from engineering and computer science to economics and physics. Among these algorithms, Newton's method holds a significant place due to its efficiency and effectiveness in finding the roots of equations and optimizing functi
15 min read
What is Machine Learning Pipeline?
In artificial intelligence, developing a successful machine learning model involves more than selecting the best algorithm; it requires effective data management, training, and deployment in an organized manner. A machine learning pipeline becomes crucial in this situation. A machine learning pipeli
7 min read
Regularization in Machine Learning
In the previous session, we learned how to implement linear regression. Now, weâll move on to regularization, which helps prevent overfitting and makes our models work better with new data. While developing machine learning models we may encounter a situation where model is overfitted. To avoid such
8 min read
Feature Transformation Techniques in Machine Learning
Most machine learning algorithms are statistics dependent, meaning that all of the algorithms are indirectly using a statistical approach to solve the complex problems in the data. In statistics, the normal distribution of the data is one that a statistician desires to be. A normal distribution of t
6 min read
Last Minute Notes (LMNs) - Machine Learning
Machine Learning (ML) is a branch of artificial intelligence where computers learn from data to make decisions or predictions without being explicitly programmed.There are two main types of learning in ML:Supervised Learning: The model is trained on labeled data, meaning it learns from input-output
11 min read