Sampling is the process of selecting a small group from a larger population to gather information and make conclusions about the whole group. It's a practical and cost-effective way to collect data, especially when studying when it is not possible.
Example:- A man wants to buy mangoes but doesn't inspect each one individually. Instead, he picks 2 to 3 mangoes at random from the pile, checks their quality, and decides that the entire batch is good based on his quick inspection.
There are two types of Sampling Methods: Probability sampling methods and Non-probability sampling methods.

Probability Sampling Methods
Every element has a known, non-zero chance of selection. Here we will discuss in detail three probability sampling methods, such as

Random Sampling Method
Random Sampling is a method where every item or individual in a group has an equal chance of being selected. It's like drawing names from a hat—each name has the same probability of being chosen.
Example: You have a class of 30 students, and you want to randomly select 5 students for a group project. You write each student’s name on a separate piece of paper, put them all in a hat, and draw 5 names. Each student has an equal chance of being selected.
Lottery Method: In the Lottery Method, the investigator prepares paper slips for each of the items of the universe and shuffles these slips in a box. After that, some slips are impartially drawn from the box to obtain a sample.
Table of Random Numbers: Use pre-generated tables of random numbers to select items.
Random Sampling and Haphazard Sampling: Random sampling follows a systematic approach based on rules of sampling, whereas haphazard sampling does not follow any systematic rules or methodology. Additionally, in random sampling, each item has an equal chance of being selected. In contrast, haphazard sampling does not ensure equal chances for each item to be selected.
Stratified or Mixed Sampling
Stratified or Mixed Sampling is a method used when a population has different groups with unique characteristics. In this method, the population is divided into smaller groups, called strata, based on these differences. Then, some items are chosen from each group to represent the whole population.
Example:- Surveying 300 customers from an e-commerce site: Split customers into strata (e.g., age groups: 18–30, 31–50, 51+). Randomly sample 100 from each group.
Stratified Sampling Method is also known as Mixed Sampling because it combines both Purposive and Random Sampling methods. The population is divided into different strata purposefully, but the items are selected randomly from each stratum.
Systematic Sampling
Systematic Sampling is a method where the population is arranged in a specific order, such as by number or alphabet. Every nth item or person is then selected for the sample. This method is usually easier than randomly picking each item and helps reduce bias.
Example: Quality testing in a factory: Take every 20th smartphone off the assembly line. Test it for defects.
Cluster Sampling
Cluster sampling is a research method where you split a large population into natural groups (like neighborhoods or schools), randomly pick a few of these groups, and study everyone in the chosen groups.
Example: To study student lunch habits in a city: Divide all students into clusters by school. Randomly select 5 schools out of 50. Survey every student in those 5 schools.
Non-Probability Sampling Methods
Selection not random; generalizability limited. Here we are explaining three types in Non-probability Sampling:

Quota Sampling
Quota Sampling is a method where the population is divided into groups based on certain characteristics, like age, gender, or income. The researcher then picks a fixed number of items from each group to form a sample. This way, the sample represents different parts of the population.
Example:- Study the opinions of 100 people about a new product, divide them into groups based on gender and age, such as 50 men and 50 women. Then, collect responses from 20 people in each age group.
Convenience Sampling
As the name suggests, Convenience Sampling is a method of collecting data where the investigator selects items from the population based on convenience.
Example:- An investigator who wants to collect data on the average number of females using induction cooktops in the kitchen goes to a shopping mall and collects information from the females visiting that mall. By doing so, the investigator neglects females who were not present in the mall that day or who did not visit the mall. This reduces the reliability of the result, as some females may have induction cooktops in the kitchen but were not present in the mall at that time.
Purposive or Deliberate Sampling
Purposive Sampling, also known as Judgmental or Deliberate Sampling, is a non-random sampling method where the researcher intentionally selects individuals or items that are most relevant to the research objectives. This approach is used when specific characteristics or expertise are needed to address particular research questions.
Example:- if an investigation is about FMCG companies, then the inclusion of companies like Nestle, Hindustan Unilever Ltd., etc., is essential in the sample. However, the chances of personal biases in this method of sampling are higher, which reduces its credibility.
Snowball Sampling
Snowball sampling is a research method where existing participants recruit new participants from their own network of contacts. This sampling starts with few people and relies on referrals. Sampling size grows like chain reaction. This is used in studying hard-to-reach or hidden population.
Example:- A researcher studying homelessness starts by interviewing one person living on the streets. That person refers the researcher to two friends who are also homeless. Those two friends each refer two more people. The sample "snowballs" from 1 person → 3 people → 7 people, growing through participant referrals.
Summary of Sampling Methods
A table summarizing the sampling methods, along with their definitions and examples.
Method of Sampling | Definition | Example |
---|
Random Sampling | A sampling method where every item of the population has an equal chance of being selected. It is impartial and does not involve investigator control. | Drawing names from a lottery to select participants. |
Stratified Sampling | The population is divided into subgroups (strata) based on distinct characteristics, and samples are selected proportionally from each subgroup. | Dividing students into Arts, Commerce, and Science groups to study academic performance. |
Systematic Sampling | The population is arranged in order, and every nth item is selected to form the sample. | Selecting every 10th person from a list of 200 for a survey. |
Cluster Sampling | Split large population into small groups and pick few groups and study samples from chosen group. | Selecting some schools from clusters of schools. |
Quota Sampling | The population is divided into groups based on certain characteristics, and fixed numbers are selected from each group to ensure diversity in the sample. | Surveying a set percentage of people from different age groups. |
Convenience Sampling | The investigator selects items based on convenience, often using readily accessible individuals or items. | Interviewing people at a local mall for a survey about cooking habits. |
Purposive Sampling | The investigator deliberately selects a sample based on their judgment, focusing on items that are deemed most relevant to the study. | Selecting top FMCG companies like Nestle and Hindustan Unilever for a market study. |
Snowball Sampling | A research method where sample is increased when existing subjects recruit by referrals. | Study about Homelessness, where a person refer to two homeless and those people then refer to more homeless people. |
Solved Problem Based on Data Sampling
Question 1. What is the purpose of data sampling?
Answer :- The main purposes of data sampling are Efficient Data Collection, Cost-Effective, and Statistical Inference.
Question 2. In a factory with 600 employees, you want to select 100 employees for a survey. You choose every 6th person from a list, starting with the 4th person. The selection process is systematic, ensuring that every 6th person is chosen. What type of sampling is being used?
Answer :- In Systematic Sampling, you select every 6th person from the list, starting at a randomly chosen point (the 4th person). This method ensures a consistent, systematic approach to selecting participants
Question 3. In which sampling method is group formation important?
Answer :- Group formation is important in stratified sampling because the population is divided into distinct subgroups (strata), and a sample is taken from each group.
Question 4. Which sampling method can generate bias?
Answer :- Convenience Sampling can generate bias because it relies on selecting easily accessible participants, which may not accurately represent the entire population.
Practice Problem based on Data Sampling
Question 1. Which sampling method is used when specific individuals are chosen based on their knowledge or expertise?
Question 2. Which sampling method involves selecting a fixed number of participants from each subgroup in proportion to their occurrence in the population?
Question 3. Which sampling method selects participants based on predetermined characteristics or criteria?
Question 4.What type of sampling ensures that each member of the population has an equal chance of being selected?
Answer :-
1. Purposive Sampling 2. Quota Sampling 3. Purposive Sampling 4. Random Sampling
Related Articles
Similar Reads
Data Science Tutorial Data Science is a field that combines statistics, machine learning and data visualization to extract meaningful insights from vast amounts of raw data and make informed decisions, helping businesses and industries to optimize their operations and predict future trends.This Data Science tutorial offe
3 min read
Introduction to Machine Learning
What is Data Science?Data science is the study of data that helps us derive useful insight for business decision making. Data Science is all about using tools, techniques, and creativity to uncover insights hidden within data. It combines math, computer science, and domain expertise to tackle real-world challenges in a
8 min read
Top 25 Python Libraries for Data Science in 2025Data Science continues to evolve with new challenges and innovations. In 2025, the role of Python has only grown stronger as it powers data science workflows. It will remain the dominant programming language in the field of data science. Its extensive ecosystem of libraries makes data manipulation,
10 min read
Difference between Structured, Semi-structured and Unstructured dataBig Data includes huge volume, high velocity, and extensible variety of data. There are 3 types: Structured data, Semi-structured data, and Unstructured data. Structured data - Structured data is data whose elements are addressable for effective analysis. It has been organized into a formatted repos
2 min read
Types of Machine LearningMachine learning is the branch of Artificial Intelligence that focuses on developing models and algorithms that let computers learn from data and improve from previous experience without being explicitly programmed for every task.In simple words, ML teaches the systems to think and understand like h
13 min read
What's Data Science Pipeline?Data Science is a field that focuses on extracting knowledge from data sets that are huge in amount. It includes preparing data, doing analysis and presenting findings to make informed decisions in an organization. A pipeline in data science is a set of actions which changes the raw data from variou
3 min read
Applications of Data ScienceData Science is the deep study of a large quantity of data, which involves extracting some meaning from the raw, structured, and unstructured data. Extracting meaningful data from large amounts usesalgorithms processing of data and this processing can be done using statistical techniques and algorit
6 min read
Python for Machine Learning
Learn Data Science Tutorial With PythonData Science has become one of the fastest-growing fields in recent years, helping organizations to make informed decisions, solve problems and understand human behavior. As the volume of data grows so does the demand for skilled data scientists. The most common languages used for data science are P
3 min read
Pandas TutorialPandas (stands for Python Data Analysis) is an open-source software library designed for data manipulation and analysis. Revolves around two primary Data structures: Series (1D) and DataFrame (2D)Built on top of NumPy, efficiently manages large datasets, offering tools for data cleaning, transformat
6 min read
NumPy Tutorial - Python LibraryNumPy is a core Python library for numerical computing, built for handling large arrays and matrices efficiently.ndarray object â Stores homogeneous data in n-dimensional arrays for fast processing.Vectorized operations â Perform element-wise calculations without explicit loops.Broadcasting â Apply
3 min read
Scikit Learn TutorialScikit-learn (also known as sklearn) is a widely-used open-source Python library for machine learning. It builds on other scientific libraries like NumPy, SciPy and Matplotlib to provide efficient tools for predictive data analysis and data mining.It offers a consistent and simple interface for a ra
3 min read
ML | Data Preprocessing in PythonData preprocessing is a important step in the data science transforming raw data into a clean structured format for analysis. It involves tasks like handling missing values, normalizing data and encoding variables. Mastering preprocessing in Python ensures reliable insights for accurate predictions
6 min read
EDA - Exploratory Data Analysis in PythonExploratory Data Analysis (EDA) is a important step in data analysis which focuses on understanding patterns, trends and relationships through statistical tools and visualizations. Python offers various libraries like pandas, numPy, matplotlib, seaborn and plotly which enables effective exploration
6 min read
Introduction to Statistics
Statistics For Data ScienceStatistics is like a toolkit we use to understand and make sense of information. It helps us collect, organize, analyze and interpret data to find patterns, trends and relationships in the world around us.From analyzing scientific experiments to making informed business decisions, statistics plays a
12 min read
Descriptive StatisticStatistics is the foundation of data science. Descriptive statistics are simple tools that help us understand and summarize data. They show the basic features of a dataset, like the average, highest and lowest values and how spread out the numbers are. It's the first step in making sense of informat
5 min read
What is Inferential Statistics?Inferential statistics is an important tool that allows us to make predictions and conclusions about a population based on sample data. Unlike descriptive statistics, which only summarize data, inferential statistics let us test hypotheses, make estimates, and measure the uncertainty about our predi
7 min read
Bayes' TheoremBayes' Theorem is a mathematical formula used to determine the conditional probability of an event based on prior knowledge and new evidence. It adjusts probabilities when new information comes in and helps make better decisions in uncertain situations.Bayes' Theorem helps us update probabilities ba
13 min read
Probability Data Distributions in Data ScienceUnderstanding how data behaves is one of the first steps in data science. Before we dive into building models or running analysis, we need to understand how the values in our dataset are spread out and thatâs where probability distributions come in.Let us start with a simple example: If you roll a f
8 min read
Parametric Methods in StatisticsParametric statistical methods are those that make assumptions regarding the distribution of the population. These methods presume that the data have a known distribution (e.g., normal, binomial, Poisson) and rely on parameters (e.g., mean and variance) to define the data.Key AssumptionsParametric t
6 min read
Non-Parametric TestsNon-parametric tests are applied in hypothesis testing when the data does not satisfy the assumptions necessary for parametric tests, such as normality or equal variances. These tests are especially helpful for analyzing ordinal data, small sample sizes, or data with outliers.Common Non-Parametric T
5 min read
Hypothesis TestingHypothesis testing compares two opposite ideas about a group of people or things and uses data from a small part of that group (a sample) to decide which idea is more likely true. We collect and study the sample data to check if the claim is correct.Hypothesis TestingFor example, if a company says i
9 min read
ANOVA for Data Science and Data AnalyticsANOVA is useful when we need to compare more than two groups and determine whether their means are significantly different. Suppose you're trying to understand which ingredients in a recipe affect its taste. Some ingredients, like spices might have a strong influence while others like a pinch of sal
9 min read
Bayesian Statistics & ProbabilityBayesian statistics sees unknown values as things that can change and updates what we believe about them whenever we get new information. It uses Bayesâ Theorem to combine what we already know with new data to get better estimates. In simple words, it means changing our initial guesses based on the
6 min read
Feature Engineering
Model Evaluation and Tuning
Data Science Practice