6.study Material MR
6.study Material MR
6.study Material MR
PROCESS
Learning Outcomes:
Able to appreciate the role of marketing research in marketing decision making.
Define marketing research problems and develop an approach to problems.
Classify and compare different Research Methods and identify appropriate research
method depending on the Research problem / objective
INFORMATION NEEDS
Information needs are facts that a person wants, needs or expects in an information source.
This can be driven by curiosity or the need to make a decision or complete a task. Information
needs are related to, but distinct from information requirements. They are studied for:
The explanation of observed phenomena of information use or expressed need;
The prediction of instances of information uses;
The control and thereby improvement of the utilization of information manipulation of
essential conditions
The following are common examples of an information need.
Consumer: Researching a purchase of a product or service.
Business: The data and information required for the strategy, planning and operations
of a business.
Research: The data, concepts and guidelines required to conduct research.
Technology: Technology information such as how to configure a web server.
Transport: Information about transportation systems such as a train schedule.
CONSUMER INSIGHT
Understanding your consumers’ needs and wants has become essential to ensure your company
is future-proof. Listening empowers you to learn from and talk to your customers in order to
create ever-improving experiences. Businesses with quality data are more likely to collect
actionable customer insights, which can help them grow their bottoms line in the long run.
1
A consumer insight is an interpretation used by businesses to gain a deeper understanding of
how their audience thinks and feels. Analysing human behaviours allows companies to really
understand what their consumers want and need, and most importantly, why they feel this way.
When consumer insight research is conducted properly, it should improve the effectiveness of
how a company communicates to its customers, which is likely to change consumer behaviour,
and therefore increase sales.
But collecting good consumer insights can be challenging, so here is what you need in order to
collect and use consumer insights properly:
Good data quality: Data quality is vital to the collection of consumer insights. Without
high-quality data, your conclusions or results might suffer.
A dedicated analytics team: The role of your data analytics team is essential in order
to understand how your customers think and behave. If you don’t have the right
analytics team, it’s hard to understand what the data is telling you.
Consumer research: It’s important to understand and acknowledge consumer
behaviour and consumer insights should help you engage with customers emotionally.
In order to do that, it’s essential not to ignore the results of your consumer research,
whether you agree with them or not.
Database and segment marketing: Database marketing is a form of marketing using
databases of customers to generate personalised communications. These databases can
lead to personas, different sets of audiences or segments. As consumer insights remain
theoretical, database marketing is another essential element to your research when it
comes to testing and learning. Indeed, test actions are necessary if you want to turn
insights into facts.
Market research can be defined as an effort to gather information about customers or markets.
It provides information about market needs, market sizes, competitors and customers. It is the
“what” of customers and markets. Market research delivers statistics and knowledge.
Consumer insights deliver the same, but tend to come with recommended actions that will help
you amplify the company’s growth. This means the team in charge of consumer insights will
deliver both data and narrative, allowing you to make use of the data.
2
Long story short, research tells us what is happening, whereas consumer insights tell us why
it’s happening, and will allow us to make changes to our business in order to improve customer
satisfaction, customer retention, and increase our bottom line.
Here are some more differences between research and insight – all food for thought:
There is plenty of research out there. There is much less actionable insight.
Research delivers facts, knowledge and statistics. Insight delivers this as well, but also
delivers an additional layer of recommended actions that can help improve the bottom
line.
Research departments deliver to the marketing function. Insight departments work in
partnership with the marketing function.
Research departments deliver data and statistics. Insight departments deliver data plus
narrative.
Insights are more sought-after and prolific in Anglo and American markets, less so
in Europe. Translation challenges of narratives appear to be the main obstacles
(although here at Feedback Ferret, we are able to manage this very successfully).
MARKETING RESEARCH
Marketing research is the function that links the consumer, customer and public to the marketer
through information. This information is used to
• identify and define marketing opportunities and problems
• generate, refine and evaluate marketing actions
• monitor marketing performance
• improve understanding of marketing as a process.
The American Marketing Association defines marketing research as "the systematic gathering,
recording and analysing of data about problems related to the marketing of goods and services".
Crisp has defined marketing research as "...the systematic, objective and exhaustive search for
and study of the facts relevant to any problem in the field of marketing".
Marketing research specifies the information required to address these issues, designs the
method for collecting information, manages and implements the data collection process,
analyses the results and communicates the findings and their implications.
4
Problem Solving Research
Problem solving research is relevant when the essence of a decision problem is known but
decision makers are unable to solve the problem with their current state of knowledge. In
contrast to the problem identification research, this type of inquiry is characterized by precise
formulation of research objectives and usually calls for more formalized, quantitative methods
including surveys, experiments and observations. As a rule, the problem solving research is
more expensive and time consuming in comparison to exploratory studies.
5
Motivation Research: It studies consumer buying motives. It studies those factors that
motivate consumers to buy a product and mainly finds out the reasons for consumers
buy the product. It also helps in finding out the causes of consumer behavior in the
market.
Ethics in marketing research refers to moral principles or values that researcher or research
organization should consider. Marketing research has experienced a resurgence with the
widespread use of the internet and the popularity of social networking. It is easier than ever
before for companies to connect directly with customers and collect individual information that
goes into a computer database to be matched with other pieces of data collected during
unrelated transactions. The way a company conducts its market research these days can have
serious ethical repercussions, impacting the lives of consumers in ways that have yet to be fully
understood. Further, companies can be faced with a public backlash if their market research
practices are perceived as unethical.
The Use of Deceptive Practices
The ease with which a company can access and gather data about its customers can lead
to deceptive practices and dishonesty in the company's research methods. This type of
ethical problem can run the gamut — from not telling customers that information is
being collected when they visit a website to misrepresenting research results by using
faulty data. At no time should a researcher ever coerce or pressure a respondent to give
a particular answer. Any action that uses lies and deception to find out or establish
information about consumer opinions falls under this category.
Invasion of Privacy
One of the most serious ethical considerations involved in market research is invasion
of privacy. Companies have an unprecedented ability to collect, store and match
information relating to customers that can infringe on a person's right to privacy. In
many instances, the customer does not know or understand the extent of the company's
infiltration into his life. The company uses this information to reach the customer with
targeted advertising, but the process of targeting can have a chilling effect on personal
freedom.
6
Breaches of Confidentiality
Another significant ethical consideration involved in market research involves breaches
of confidentiality. Companies regularly share information about customers with
partners and affiliates, requiring the customer to opt-out of the sharing if he doesn't
want to be involved. Some companies sell information they have gathered on customers
to outside companies. Ethically, any unauthorized disclosure of customer information
is problematic.
Undertaking Objective Market Research
Marketing and advertising have a significant impact on public perceptions. Market
researchers have an ethical obligation to conduct research objectively, so that available
data allows for the development of a balanced or reality-based picture. Researchers who
allow their own prejudices to skew their work tend to contribute to the perpetuation of
stereotypes in advertising, the development of destructive social constructs and the
enabling of unjust profiting from poverty.
RESEARCH PROCESS
Research is a systematic process of collecting and analysing information to increase our
understanding of the phenomenon under study. The sequential steps involved in the completion
of a research under consideration are referred to as research process. The researcher must
complete each step meticulously to accomplish the research under consideration most
effectively. Following are steps involved in the research process:
Formulating the research problem
Extensive literature survey
Development of working hypothesis
Preparing the research design
Determining sample design
Collecting the data
Execution of the project
Analysis of data
Hypothesis testing
Generalizations & interpretations
Preparation of the report or the thesis
7
1. Selecting the research problem (or an opportunity)
Defining the goals and objectives of a research project is one of the most important
steps in the research process. Clearly stated goals keep a research project focused. The
process of goal definition usually begins by writing down the broad and general goals
of the study. As the process continues, the goals become more clearly defined and the
research issues are narrowed. In this stage the researcher identifies a few problems areas
after developing and evaluating several alternatives.
3. Review of Literature
One of the most important early steps in a research is the conducting of the literature
review. A literature review is an evaluative report of information found in the literature
related to the selected area of study. An elaborate literature review helps researcher to
know whether any research is done in this direction. If some research has already gone
8
into the problem under consideration, the researcher can know how his research is
different from the research which is already conducted. Literature review gives an idea
about the tools and techniques used by the earlier researcher. Information in literature
review should be organized and related directly to the research platform. It assists the
researcher in knowing the ways and means to deal with the research problem.
Researcher should undertake extensive literature survey connected with the problem.
For this purpose, the abstracting and indexing journals and published or unpublished
bibliographies are the first place to go to. Academic journals, conference proceedings,
government reports, books etc., must be tapped depending on the nature of the problem.
In this process, it should be remembered that one source will lead to another. The earlier
studies, if any, which are similar to the study in hand should be carefully studied. A
good library will be a great help to the researcher at this stage.
4. Development of hypotheses
A hypothesis is an educated guess. A hypothesis is provisional statement that proposes
a possible explanation to some phenomenon or event. In other words a hypothesis is a
precise testable statement a prediction of what the researcher expects to find or prove.
Research hypothesis should be specified prior to the collection of any data.
9
5. Research Design
Research design is a plan for collecting and utilizing data so that desired information
can be obtained with sufficient precision or so that a hypothesis can be tested properly.
Research design can be thought of as the structure of research. It is a plan which covers
all the aspects of the selected research. It explains: Why the study is conducted? What
is the time involved in completing the work? Objectives of the study, Types of data
needed for the study under consideration, Sources of data, sampling technique selected
for the proposed research, tools used for data collection, techniques used for data
analysis, the pattern of report to be prepared etc.
10
7. Data Collection
The search for answers to research questions calls for collection of data. Data are facts,
figures and other relevant materials, past and present, serving as bases for study and
analysis. The research outcome is based on the nature and the quality of data collected
by the researcher. The required information may be collected from people directly
thorough interviews, questionnaires or by observations. The information can also be
collected from the published sources like reports, journals, newspapers, internet etc.
9. Testing of Hypothesis
Hypothesis testing is the use of statistics to determine the probability that a given
hypothesis is true. The usual process of hypothesis testing consists of four steps.
a. Formulate the null hypothesis and the alternative hypothesis
b. Identify a test statistic that can be used to assess the truth of the null hypothesis.
Various tests such as t-test, z-test, chi-square test etc are used for the purpose
c. The next step is to select a significance level. Typically the 0.05 or the 0.01 level is
used.
d. The next step is to calculate a statistic analogous to the parameter specified by the
null hypothesis
e. Compare the probability value to the significance level and make the statistical
decision.
11
10. Report writing
Report writing is the final stage of a research project. The purpose of the report is to
communicate the research findings to the interested persons. It explains, the problem
investigated, method adopted to collect the data, method adopted to analyse the data
and the inference.
Following guidelines will help the researcher to write the report more effectively
1. Style of the report: It is important to communicate in a way that is both acceptable
and clear to the readers.
2. Selectivity: Careful choice of words can enable the researcher to convey the exact
tone of the information in the right spirit
3. Accuracy: accurate information is essential for effective communication and
decision making. Therefore, the researcher should ensure that everything he writes
is factually accurate
4. Objectivity A report should not be an essay reflecting personal emotions and
opinions. The researcher should have an open mind when writing report. The
emphasis, therefore, should be on the factual material presented and the conclusions
drawn, rather than on any personal beliefs, biases or prejudices.
5. Conciseness: The report should be concise. A concise report is short but still
contains all the essential details. Therefore, the researcher should ask the following
question while including some information in the report- Is it necessary for this
information to be included?’
6. Simplicity: The information should be presented as simple as it can be. The
researcher should keep asking the question whether or not the reader will be able to
follow the logic of presentation.
7. Clarity and Consistency: The best way to achieve clarity in the writing is to allow
some time to elapse between the first draft and its revision. The time gap allows the
researcher to think and refine the presentation of the report.
8. Avoid Pointless Words: Sometimes we use words and phrases like basically,
actually, undoubtedly, each and every one and during the course of our investigation
etc. These words add nothing to the message and often can be eliminated without
changing the meaning or the tone of the statement.
12
TYPES OF DATA
As already discussed, a researcher needs to collect different types of data for different purposes.
There are mainly two types of data, as shown in Figure-1:
Types of Data
Primary Data: Refers to the data that is collected by the researcher for a researcher for a
particular research. Primary data is the data that does not have any prior existence and collected
directly from the respondents. It is considered very reliable in comparison to all other forms of
data. However, its reliability may come under scrutiny for various reasons.
For example, the researcher may be biased while collecting data, the respondents may not feel
comfortable to answer the questions, and the researcher may influence the respondents. In all
these scenarios, primary data would not be very dependable.
Therefore, primary data collection should be done with utmost caution and prudence. Primary
data helps the researchers in understanding the real situation of a problem. It presents the
current scenario in front of the researchers; therefore, it is more effective in taking the business
decisions.
Secondary Data: Refers to the data that is collected in the past, but can be utilized in the
present scenario/research work. The collection of secondary data requires less time in
comparison to primary data.
Secondary data is the data that have been already collected by and readily available from other
sources. Such data are cheaper and more quickly obtainable than the primary data and also may
be available when primary data cannot be obtained at all.
13
Advantages of Secondary data
It is economical. It saves efforts and expenses.
It is time saving.
It helps to make primary data collection more specific since with the help of secondary
data, we are able to make out what are the gaps and deficiencies and what additional
information needs to be collected.
It helps to improve the understanding of the problem.
It provides a basis for comparison for the data that is collected by the researcher.
14
3. Accuracy: In order to find how accurate the data is, the following points must be
considered:
a. Specification and methodology used;
b. Margin of error should be examined;
c. The dependability of the source must be seen.
4. Sufficiency: Adequate data should be available.
RESEARCH DESIGN
A research design is the arrangement of conditions for collection and analysis of data in a
manner that aims to combine relevance to the research purpose with economy in procedure. A
traditional research design is a blueprint or detailed plan for how a research study is to be
completed – operationalizing variables so they can be measured, selecting a sample of interest
to study, collecting data to be used as a basis for testing hypothesis and analyzing the results.
Decisions regarding what, where, when, how much, by what means concerning an inquiry or a
research study constitute a research design. The research design is the conceptual structure
within which research is conducted. It constitutes the blueprint for the collection, measurement
and analysis of data.
Research must ensure to:
Convert the research question and the stated hypothesis into operational variables that
can be measured.
Specify the process that would be followed to complete the above task, as efficiently
and economically as possible.
Specify the control mechanisms that would be used to ensure that the effect of other
variables that could impact the outcome of the study have been controlled.
15
Need for Research Design
It facilitates the smooth sailing of the various research operations.
Maximal information with minimal expenditure of effort, time and money.
It helps the researcher to organize his ideas in a form whereby it will be possible for
him to look for flaws and inadequacies.
It stands for advance planning of the methods to be adopted for collecting the relevant
data and the techniques to be used in their analysis, keeping in view the objective of the
research and the availability of staff, time and money.
16
Research hypothesis: When a prediction or a hypothesized relationship is to be tested
by scientific methods, it is termed as research hypothesis.
The research hypothesis is a predictive statement that relates an independent
variable to a dependent variable.
Experimental and non-experimental hypothesis-testing research:
When the purpose of research is to test a research hypothesis, it is termed as
hypothesis-testing research.
It can be of the experimental design or of the non-experimental design.
For instance, a researcher wants to study whether intelligence affects reading
ability for a group of students.
Treatments: It is the process of using an agent, a procedure on the subject.
For example, if we want to determine through an experiment the comparative
impact of three varieties of fertilizers on the yield of wheat, in that case the three
varieties of fertilizers will be treated as three treatments.
Experiment: Examining the truth of a statistical hypothesis, relating to some research
problem, is known as an experiment.
For example, an experiment to examine the usefulness of a certain newly
developed drug
Experiments can be of two types viz., absolute experiment and comparative
experiment.
If we want to determine the impact of a fertilizer on the yield of a crop,
it is a case of absolute experiment.
but if we want to determine the impact of one fertilizer as compared to
the impact of some other fertilizer, our experiment then will be termed
as a comparative experiment.
Experimental unit: The pre-determined plots or the blocks, where different treatments
are used, are known as experimental units.
17
Types of Research Design
18
• Well-grounded picture of the situation being developed.
• Generation of new ideas and assumptions.
• Development of tentative theories or hypotheses.
• Determination about whether a study is feasible in the future.
• Issues get refined for more systematic investigation and formulation of new research
questions.
• Direction for future research and techniques get developed
20
explanations that reflect tests of hypotheses. Causal effect (nomothetic perspective) occurs
when variation in one phenomenon, an independent variable, leads to or results, on average, in
variation in another phenomenon, the dependent variable.
Conditions necessary for determining causality:
• Empirical association -- a valid conclusion is based on finding an association between
the independent variable and the dependent variable.
• Appropriate time order -- to conclude that causation was involved, one must see that
cases were exposed to variation in the independent variable before variation in the
dependent variable.
• Non-spuriousness -- a relationship between two variables that is not due to variation
in a third variable.
21
CROSS-SECTIONAL STUDIES
It involves a slice of the population and studies the nature of the relevant variables being
investigated.
Two essential characteristics:
It is carried out at a single moment in time and thus the applicability is most relevant
for a specific period. (Example: study on the attitude of Americans towards Asian-
Americans, pre and post 9/11, was vastly different and a study done in 2011 would
reveal a different attitude and behaviour which might not be absolutely in line with that
found earlier).
Study is carried out on a section of respondents from the population units under study.
Cross-sectional research designs have three distinctive features: no time dimension; a
reliance on existing differences rather than change following intervention; and, groups
are selected based on existing differences rather than random allocation. The cross-
sectional design can only measure differences between or from among a variety of
people, subjects, or phenomena rather than a process of change. As such, researchers
using this design can only employ a relatively passive approach to making causal
inferences based on findings.
22
6. Can estimate prevalence of an outcome of interest because the sample is usually taken
from the whole population.
7. Because cross-sectional designs generally use survey techniques to gather data, they
are relatively inexpensive and take up little time to conduct.
23
3. The design permits the measurement of differences or change in a variable from
one period to another [i.e., the description of patterns of change over time].
4. Longitudinal studies facilitate the prediction of future outcomes based upon earlier
factors.
24
The total error is the variation between the true mean value in the population of the variable
of interest and the observed mean value obtained in the marketing research project.
Random sampling error is the variation between the true mean value for the population and
the true mean value for the original sample.
Non-sampling errors can be attributed to sources other than sampling, and they may be
random or nonrandom: including errors in problem definition, approach, scales, questionnaire
design, interviewing methods and data preparation and analysis. Non-sampling errors consist
of non-response errors and response errors.
Non-response error arises when some of the respondents included in the sample do not
respond.
Response error arises when respondents give inaccurate answers or their answers are mis-
recorded or mis-analyzed.
QUALITITATIVE RESEARCH
Qualitative Research is primarily exploratory research. It is used to gain an understanding of
underlying reasons, opinions, and motivations. It provides insights into the problem or helps to
develop ideas or hypotheses for potential quantitative research. Qualitative Research is also
used to uncover trends in thought and opinions, and dive deeper into the problem.
Qualitative data collection methods vary using unstructured or semi-structured techniques.
Some common methods include focus groups (group discussions), individual interviews, and
participation/observations. The sample size is typically small, and respondents are selected to
fulfil a given quota.
QUANTITATIVE RESEARCH
Quantitative Research is used to quantify the problem by way of generating numerical data or
data that can be transformed into usable statistics. It is used to quantify attitudes, opinions,
behaviors, and other defined variables – and generalize results from a larger sample population.
Quantitative data is designed to collect cold, hard facts. It provides support when you need to
draw general conclusions from your research.
Quantitative Research uses measurable data to formulate facts and uncover patterns in research.
Quantitative data collection methods are much more structured than Qualitative data collection
methods. Quantitative data collection methods include various forms of surveys – online
surveys, paper surveys, mobile surveys and kiosk surveys, face-to-face interviews, telephone
25
interviews, longitudinal studies, website interceptors, online polls, and systematic
observations.
This approach can be further sub-classified into
Inferential,
Experimental and
Simulation approaches to research.
The purpose of inferential approach to research is to form a data base from which to infer
characteristics or relationships of population. This usually means survey research where a
sample of population is studied (questioned or observed) to determine its characteristics, and it
is then inferred that the population has the same characteristics.
Experimental approach is characterised by much greater control over the research
environment and in this case some variables are manipulated to observe their effect on other
variables.
Simulation approach involves the construction of an artificial environment within which
relevant information and data can be generated. This permits an observation of the dynamic
behaviour of a system (or its sub-system) under controlled conditions. The term ‘simulation’
in the context of business and social sciences applications refers to “the operation of a
numerical model that represents the structure of a dynamic process. Given the values of initial
conditions, parameters and exogenous variables, a simulation is run to represent the behaviour
of the process over time.” Simulation approach can also be useful in building models for
understanding future conditions.
26
or just someone’s perception? The hard facts obtained will enable you to make
decisions based on objective observations.
Finding general answers: Quantitative research usually has more respondents than
qualitative research because it is easier to conduct a multiple-choice survey than a series
of interviews or focus groups. Therefore it can help you definitely answer broad
questions like: Do people prefer you to your competitors? Which of your company’s
services are most important? What ad is most appealing?
Incorporating the human element: Qualitative research can also help in the final
stages of your project. The quotes you obtained from open-ended questions can put a
human voice to the objective numbers and trends in your results. Many times it helps
to hear your customers describe your company in their own words to uncover your blind
spots. Qualitative data will get you that.
27
Focus Group
The term, focus group, was coined by Ernest Ditcher, an eminent psychologist and marketing
expert. Focus group is a form of qualitative research, which is based on communication
between research participants to generate data. The focus group method of data generation
seems to be similar to group interviews. However, focus group lays more emphasis on informal
group interaction; whereas; in interview, formal interaction takes place between the interviewer
and the respondent. In focus group, the researcher does not ask questions explicitly, instead
people in the group are encouraged to talk to each other, ask questions from each other, generate
different views about the topic, and comment and add upon other’s views and experiences.
There are various types of focus groups. Some of the important types of focus groups are
mentioned as follows:
Two-way focus group: Refers to the type of focus group in which one focus
group observes and discusses the interaction of the other focus group.
Dual moderator focus group: Refers to the focus group, which includes
two moderators. One moderator ensures smooth discussion while the other
ensures that all topics are covered.
Dueling moderator focus group: Refers to the focus group, which is
similar to dual moderator focus group, except the differences that two
moderators intentionally take opposite sides on the topics are discussion.
Respondent moderator focus group: Refers to the group in which one of
the respondents acts as a moderator temporarily.
Client participant focus group: Refers to the focus group in which client
representatives take part in the discussion.
Mini focus group: Refers to a small group composed of 3 to 4 members
instead of 6 to 12.
Teleconference focus group: Refers to the group in which members are not
physically present together. The members interact with each other through
a telephone network.
Online focus group: Refers to the group in which members are connected
via the internet.
Benefits of focus groups can be mentioned as follows:
Helps in generating new ideas, which are difficult to obtain by other
means
28
Refers to an efficient way of discussion any topic or acquiring
feedback or spreading any idea
Helps in marketing research where customer feedback is required on
any new product
Helps the government in acquiring feedback of people on its various
plans and policies
The disadvantages of focus group are as follows:
Reduces the control of the researcher over the group
Finds it difficult to analyze the conclusion drawn in a focus group as
the communication takes place in an informal way
Lacks anonymity that may stop participants to freely express their
views
Depth Interview
It is an unstructured personal interview conducted one on one.
Projective Technique
Projective Techniques are indirect and unstructured methods of investigation which have
been developed by the psychologists and use projection of respondents for inferring about
underline motives, urges or intentions which cannot be secure through direct questioning as the
respondent either resists to reveal them or is unable to figure out himself. These techniques are
useful in giving respondents opportunities to express their attitudes without personal
embarrassment.
These techniques helps the respondents to project his own attitude and feelings unconsciously
on the subject under study. Thus Projective Techniques play a important role in motivational
researches or in attitude surveys.
29
1. Word Association Test: An individual is given a clue or hint and asked to respond to
the first thing that comes to mind. The association can take the shape of a picture or a
word. There can be many interpretations of the same thing. A list of words is given and
you don’t know in which word they are most interested. The interviewer records the
responses which reveal the inner feeling of the respondents. The frequency with which
any word is given a response and the amount of time that elapses before the response
is given are important for the researcher. For eg: Out of 50 respondents 20 people
associate the word “ Fair” with “Complexion”.
2. Completion Test: In this the respondents are asked to complete an incomplete sentence
or story. The completion will reflect their attitude and state of mind.
3. Construction Test: This is more or less like completion test. They can give you a
picture and you are asked to write a story about it. The initial structure is limited and
not detailed like the completion test. For eg: 2 cartoons are given and a dialogue is to
written.
4. Expression Techniques: In this the people are asked to express the feeling or attitude
of other people.
30
MODULE 2: MEASUREMENT AND SCALING AND QUESTIONNAIRE
DESIGN
Learning Outcomes:
Identify Primary Scales of Measurement
Distinguish between Comparative and non-comparative scaling techniques.
Examine purpose and process of designing a questionnaire.
Measurement is the process of observing and recording the observations that are collected as
part of a research effort. It is a process of mapping aspects of a domain onto other aspects of a
range according to some rule of correspondence.
Researchers always face the problem of how to measure or quantify abstract concepts and how
to relate one individual’s response to that with another response. Hence, the problem is two-
fold. First is, how to quantify a response in absolute terms, and the second is, how to relate it
with other responses. This problem is resolved by scaling process i.e. assigning the numbers or
values to responses.
Attitude measurement tends to focus on beliefs and emotional feelings. It is important for
marketers to try and understand these attitudes and influence them to gain an advantage in the
market. Measuring attitudes is a highly difficult process and unlike measurement scales in the
physical sciences like measuring height, weight, etc., measurement scales for attitude are less
precise. There are different attitude scales that can be used to measure attitudes as precisely as
possible.
MEASUREMENT TECHNIQUES
There are four techniques available:
Questionnaires
Depth interviews
Attitude scales
Observation
Questionnaires: This is a set of questions, used as an instrument for seeking relevant
information directly from respondents. Generally, a question or set of questions represents a
variable used in research. These are especially designed for a particular research and then
suitably validated before its use.
31
Depth Interviews: These are interviews in which individuals are made to express their feelings
freely and without fear of dispute or disapproval. The details are recorded in specifically
designed sheets. They are useful when one wants detailed information about a person’s
thoughts and behaviours or want to explore new issues in depth.
Attitude Scales: These scales elicit self-reports of beliefs and feelings towards an object. There
are different types of attitude scales:
Rating scales that require the respondent to place the object at some point on a
continuum that is numerically ordered.
Composite scales that require the respondent to express a degree of belief with regard
to several attributes of an object
Multi-dimensional scales and scales developed using conjoint analysis used for
inferring specific aspects of an individual’s attitude towards an object at against direct
evaluation of the respondents.
Observation: This is a direct examination of behaviour or results of behaviour. These
procedures require a person to dedicate his or her attention to the behaviours of an individual
or group in a natural setting for a certain period of time.
DIFFICULTIES IN MEASUREMENT
Irrelevant Date
Inaccurate Response
Researchers and Analysts get no respect
Training in measurement is rare
Delegating measurement strategy
SCALES OF MEASUREMENT
Scaling is the procedure of assigning numbers to various degrees of opinions, attitude and
other concepts. Scale is continuum, consisting of highest and lowest points. Scaling is applied
to the procedures for attempting to determine quantitative measures of subjective abstract
concepts. It can be done in two ways:
Making a judgment about some characteristic of an individual & then placing
him directly on the scale that has been defined in terms of that characteristic.
Constructing questionnaires in such a way that the score of individual’s
responses assign him a place on a scale
32
There are four types of measurement scales: Nominal, Ordinal, Interval and Ratio Scales.
Nominal Scale: Nominal variables (also called categorical variables) can be placed into
categories. They don’t have a numeric value and so cannot be added, subtracted, divided or
multiplied. They also have no order; if they appear to have an order then you probably have
ordinal variables instead. The nominal scale, sometimes called the qualitative type, places non-
numerical data into categories or classifications. For example:
Placing cats into breed type. Example: a Persian is a breed of cat.
Putting cities into states. Example: Jacksonville is a city in Florida.
Surveying people to find out if men or women have higher self-esteem.
Finding out if introverts or extroverts are more likely to be philanthropic.
These pieces of information aren’t numerical. They are assigned a category (breeds of cat, cities
in Florida, men and women, introvert and extrovert). Qualitative variables are measured on the
nominal scale.
33
Ordinal Scale: The ordinal scale contains things that you can place in order. For example,
hottest to coldest, lightest to heaviest, richest to poorest. Basically, if you can rank data by 1st,
2nd, 3rd place (and so on), then you have data that’s on an ordinal scale. Median is the
appropriate measure of central tendency. Percentile / quartile is the appropriate measure of
dispersion. No correlation can be obtained for ordinal data. Only non-parametric methods can
be used in these cases.
Ordinal data is made up of ordinal variables. In other words, if you have a list that can be placed
in “first, second, third…” order, you have ordinal data. It sounds simple, but there are a couple
of elements that can be confusing:
1. You don’t have to have the exact words “first, second, third…etc.” Instead, you can have
different rating scales, like “Hot, hotter, hottest” or “Agree, strongly agree, disagree.”
2. You don’t know if the intervals between the values are equal. We know that a list of
cardinal numbers like 1, 5, 10 have a set value between them (in this case, 5) but with
ordinal data you just don’t know. For example, in a marathon you might have first, second
and third place. But if you don’t know the exact finishing times, you don’t know what the
interval between first and second, or second and third is.
The Likert Scale gives another example of how you can’t be sure about intervals with ordinal
data. What is the interval between “strongly agrees” and “agrees”? It’s practically impossible
to put any kind of number to that interval. Even if you could put a number to the interval, the
gap between “strongly agree” and “agree” is likely to be much smaller than the gap between
34
“agree” and “no opinion.” Think of someone being asked to rate a question like “Chocolate is
irresistible.” Someone who likes chocolate a lot might have their pencil hover between
answering “strongly agree” or “agree”, but their pencil never hovers over “no opinion.”
Interval Scale: An interval scale has ordered numbers with meaningful divisions. Temperature
is on the interval scale: a difference of 10 degrees between 90 and 100 means the same as 10
degrees between 150 and 160. Compare that to high school ranking (which is ordinal), where
the difference between 1st and 2nd might be .01 and between 10th and 11th .5. If you have
meaningful divisions, you have something on the interval scale.
Dates are also measured on an interval scale. For example, there’s 100 years between the 20th
and 21st, and also the 21st and 22 centuries. Dates illustrate a major problem with interval
scales: the zero is arbitrary. Year zero doesn’t exist in the A.D. system (which starts at year 1)
but the Buddhist and Hindu calendars include it. Arbitrary zeros are one reason why you can’t
say that “the 10th century is twice as long as the fifth century.” This leads to another issue with
zeros in the interval scale: Zero doesn’t mean that something doesn’t exist. For example, the
year 0 doesn’t imply that time didn’t exist. And similarly, a temperature of zero doesn’t mean
that temperature doesn’t exist at that point. Arbitrary zeros (and the inability to calculate ratios
because of it) are one reason why the ratio scale — which does have meaningful zeros — is
sometimes preferred.
Ratio Scale: The ratio scale is exactly the same as the interval scale with one major difference:
zero is meaningful. For example, a height of zero is meaningful (it means you don’t exist).
Compare that to a temperature of zero, which while it exists, it doesn’t mean anything in
particular (although admittedly, in the Celsius scale it’s the freezing point for water).
A ratio scale has all the properties of an interval scale. Ratio data on the ratio scale has
measurable intervals. For example, the difference between a height of six feet and five feet is
the same as the interval between two feet and three feet. Where the ratio scale differs from the
interval scale is that it also has a meaningful zero. The zero in a ratio scale means that
35
something doesn’t exist. For example, the zero in the Kelvin temperature scale means that heat
does not exist at zero. Other examples of the ratio scale:
Age. The clock starts ticking when you are born, but an age of “0” technically means
you don’t exist.
Weight. At 0 pounds, you would weight nothing and therefore wouldn’t exist.
Height. If you were 0″, you would have no height.
Sales figures. Sales of zero means that you sold nothing and so sales didn’t exist.
Quantity purchased. If you bought 0 items, then there were no quantities purchased.
Time measured from the “Big Bang.”
In Marketing Research, several scaling techniques are employed to study the relationship
between the objects. The most commonly used techniques can be classified as Comparative
and Non-Comparative Scales.
36
ink pen and gel pen. The comparative data can only be interpreted in relative terms and hence
possess the ordinal or rank-order properties. This is the reason why the comparative scaling is
also called as nonmetric scaling. The Comparative Scaling includes the following techniques:
Paired Comparison Scaling
Rank Order Scaling
Constant Sum Scaling
Q-Sort Scaling
Paired Comparison Scaling: In paired comparison scaling, respondents are asked to choose
one among two alternatives on a selected criterion. For example, a respondent may be asked to
choose between two well-known toothpaste brands on the criterion of quality. The data
obtained from paired comparison scaling is ordinal in nature. When there are more than two
stimuli involved paired comparison scaling can still be useful technique to compare various
stimuli. Let’s say, a researcher is interested in knowing consumers preference among three
different toothpaste brands, A, B and C. Using the paired comparison scaling researcher will
create three questions for respondents namely:
1. Preference between toothpaste brand A and brand B
2. Preference between toothpaste brand B and brand C
3. Preference between toothpaste brand A and brand C.
If for the first question respondent choose brand A over brand B and in the second question
chose brand B over brand C, using simple logic researcher can derive that brand A will be more
preferred in comparison to brand C. In simple terms, using paired comparison scaling
researcher can generate a rank order among stimuli. Paired comparison scaling is used in
pricing decisions frequently. It is quite helpful when the number of stimuli is limited. In such
circumstances, paired comparison can reveal direct comparisons and overt choice. However,
when large number of stimuli is involved, paired comparison scaling becomes a tedious
technique. Paired comparison scaling is highly used in product testing. Many food companies
and other Fast Moving Consumer Goods (FMCG) companies use this technique to compare
their existing product with an upcoming variant or with their competitor’s products.
The Rank Order Scaling is a yet another comparative scaling technique wherein the
respondents are presented with numerous objects simultaneously and are required to order or
37
rank these according to some specified criterion. The Rank order scaling is often used to
measure the preference for the brand and attributes.
The Constant-Sum Scaling is a technique wherein the respondents are asked to allocate a
constant sum of units, such as points, dollars, chips or chits among the stimulus objects
according to some specified criterion. For example, suppose a respondent is asked to allocate
100 points to the attributes of a body wash on the basis of the importance he attaches to each
attribute. In case he feels any attribute being unimportant can allocate zero points and in case
some attribute is twice as important as any other attribute can assign it twice the points. The
sum of all the points allocated to each attribute should be equal to 100. Once the points are
allocated, the attributes are scaled by counting the points as assigned by the respondents to each
attribute and then dividing it by a number of respondents under analysis. Such type of
information cannot be obtained from rank order data unless it is transformed into interval data.
The constant sum scaling is considered as an ordinal scale because of its comparative nature
and lack of generalization.
The Q-Sort Scaling is a Rank order scaling technique wherein the respondents are asked to
sort the presented objects into piles based on similarity according to a specified criterion such
as preference, attitude, perception, etc. For example, suppose the respondents are given 100
motivational statements on individual cards and are asked to place these in 11 piles, ranging
from the “most agreed with” to the “least agreed with”. Generally, the most agreed statement
is placed on the top while the least agreed statement at the bottom.
38
The Continuous Rating Scale is a technique wherein the respondents are asked to rate the
stimulus objects by placing a point/mark appropriately on a line running from one extreme of
the criterion to the other variable criterion. The continuous rating scale is also called as a
Graphic Rating Scale. Here the respondent can place a mark anywhere on the line based on his
opinion and is not restricted to select from the values as previously set by the researcher. The
continuous scale can observe many forms, i.e. it can either be vertical or horizontal; scale
points, in the form of numbers or brief descriptions, may be provided, and if these are provided,
then the scale points might be few or many.
The Itemized Rating Scale is an Ordinal Scale that has a brief description or numbers
associated with each category, ordered in terms of scale positions. The respondents are asked
to select the category that best describes the stimulus object being rated.
The following are the most commonly used itemized rating scales:
Likert Scale: A Likert Scale is a scale with five response categories that ranges from
“strongly disagree” to “strongly agree”, wherein the respondent is asked to indicate the
degree of agreement or disagreement with each of the statements related to the stimulus
object under analysis.
39
Semantic Differential Scale: The semantic differential scale is a seven-point rating
scale with the extreme points having the semantic meaning. The scale is used to
measure the meaning or semantics of words, especially the bi-polar adjectives (such as
“evil” or “good”, “warm” or “cold”) to derive the respondent’s attitude towards the
stimulus object.
40
Stapel Scale: Stapel scale is a single adjective rating scale with 10 categories ranging
from -5 to +5 with no zero points. This scale is usually a vertical scale in which the
single adjective is placed in the middle of the even-numbered range (-5 to +5). The
respondent is asked to identify how correctly or incorrectly each term describes the
stimulus object by choosing an appropriate response category.
The itemized rating scale is widely used in marketing research and serve as a basic component
of more complex scales, such as Multi-Item Scales.
Thus, the researcher can apply any of the scaling techniques to determine the characteristic of
an individual and then locating him on the scale that best fits the defined characteristics.
41
METHODS OF DATA COLLECTION
BASIS FOR
PRIMARY DATA SECONDARY DATA
COMPARISON
Meaning Primary data refers to the first Secondary data means data collected by
hand data gathered by the someone else earlier.
researcher himself.
42
Disadvantages of Secondary Data
Inappropriateness
Lack of control over data quality
May not be specific to the need
Biasness
Lack of Availability
Time lag issues
This method of data collection is quite popular, particularly in case of big enquiries. In this
method a questionnaire is sent (usually by post) to the persons concerned with a request to
answer the questions and return the questionnaire. A questionnaire consists of a number of
questions printed or typed in a definite order on a form or set of forms. The questionnaire is
mailed to respondents who are expected to read and understand the questions and write down
the reply in the space meant for the purpose in the questionnaire itself. The respondents have
to answer the questions on their own.
The method of collecting data by administering the questionnaires to respondents is most
extensively employed in various economic and business surveys.
43
There is inbuilt inflexibility because of the difficulty of amending the approach once
questionnaires have been despatched.
There is also the possibility of ambiguous replies or omission of replies altogether to
certain questions; interpretation of omissions is difficult.
It is difficult to know whether willing respondents are truly representative.
This method is likely to be the slowest of all.
Essentials of a questionnaire
Minimum size
Logical sequencing (easy to difficult)
Place personal/intimate questions at the end
Technical terms & Vague expression should be avoided
Control questions to indicate the reliability of the respondents
Avoid affecting sentiments
Adequate space to be provided
Provision for uncertainty
Direction to fill up
Physical Appearance of the questionnaire
TYPES OF QUESTIONNAIRE
Open question questionnaires. Open questions differ from other types of questions used in
questionnaires in a way that open questions may produce unexpected results, which can make
the research more original and valuable. However, it is difficult to analyze the results of the
findings when the data is obtained through the questionnaire with open questions.
Multiple choice questions. Respondents are offered a set of answers they have to choose from.
The downsize of questionnaire with multiple choice questions is that, if there are too many
answers to choose from, it makes the questionnaire, confusing and boring, and discourages the
respondent to answer the questionnaire.
Dichotomous Questions. This type of questions gives two options to respondents – yes or no,
to choose from. It is the easiest form of questionnaire for the respondent in terms of responding
it.
44
Scaling Questions. Also referred to as ranking questions, they present an option for
respondents to rank the available answers to the questions on the scale of given range of values
(for example from 1 to 10).
DESIGNING A QUESTIONNAIRE
Specify the Information Needed: The researcher must specify the information needed
from the respondents such that the objective of the survey is fulfilled. The researcher
must completely review the components of the problem, particularly the hypothesis,
research questions, and the information needed.
Define the Target Respondent: At the very outset, the researcher must identify the
target respondent from whom the information is to be collected. The questions must be
designed keeping in mind the type of respondents under study. The less diversified
respondent group shall be selected because the more diversified the group is, the more
difficult it will be to design a single questionnaire that is appropriate for the entire
group.
Specify the type of Interviewing Method: The next step is to identify the way in which
the respondents are reached. In personal interviews, the respondent is presented with a
questionnaire and interacts face-to-face with the interviewer. In telephone interviews,
the respondent is required to give answers to the questions over the telephone. For a
questionnaire that is sent through mail or post, it should be self-explanatory and contain
all the important information such that the respondent is able to understand every
question and gives a complete response. The electronic questionnaires are sent directly
to the mail ids of the respondents and are required to give answers online.
Determine the Content of Individual Questions: The researcher must decide on what
should be included in the question such that it contribute to the information needed or
serve some specific purpose. The researcher must try to avoid the use of double-
barrelled questions. A question that talks about two issues simultaneously, such as Is
the Real juice tasty and a refreshing health drink?
Overcome Respondent’s Inability and Unwillingness to Answer: The questions
must be designed in a simple and easy language such that it is easily understood by each
respondent. In situations, where the respondent is not at all informed about the topic of
interest, then the researcher may ask the filter questions, an initial question asked in the
questionnaire to identify the prospective respondents to ensure that they fulfil the
45
requirements of the sample. Despite being able to answer the question, the respondent
is unwilling to devote time in providing information. The researcher must attempt to
understand the reason behind such unwillingness and design the questionnaire in such
a way that it helps in retaining the respondent’s attention.
Decide on the Question Structure: The researcher must decide on the structure of
questions to be included in the questionnaire. The question can be structured or
unstructured.
Determine the Question Wording: The desired question content and structure must
be translated into words which are easily understood by the respondents. At this step,
the researcher must translate the questions in easy words such that the information
received from the respondents is similar to what was intended. In case, the respondent
is reluctant to give answers, then “nonresponse error” arises which increases the
complexity of data analysis. On the other hand, if the wrong information is given, then
“response error” arises due to which the result is biased.
Determine the Order of Questions: At this step, the researcher must decide the
sequence in which the questions are to be asked. The opening questions are crucial in
establishing respondent’s involvement and rapport, and therefore, these questions must
be interesting, non-threatening and easy. Usually, the open-ended questions which ask
respondents for their opinions are considered as good opening questions, because
people like to express their opinions.
Identify the Form and Layout: The format, positioning and spacing of questions has
a significant effect on the results. The layout of a questionnaire is specifically important
for the self-administered questionnaires. The questionnaires must be divided into
several parts, and each part shall be numbered accurately to clearly define the branches
of a question.
Reproduction of Questionnaire: Here, we talk about the appearance of the
questionnaire, i.e. the quality of paper on which the questionnaire is either written or
printed. In case, the questionnaire is reproduced on a poor-quality paper; then the
respondent might feel the research is unimportant due to which the quality of response
gets adversely affected. In case, the questionnaire has several pages, then it should be
presented in the form of a booklet rather than the sheets clipped or stapled together.
Pretesting: Pretesting means testing the questionnaires on a few selected respondents
or a small sample of actual respondents with a purpose of improving the questionnaire
46
by identifying and eliminating the potential problems. All the aspects of the
questionnaire must be tested such as question content, structure, wording, sequence,
form and layout, instructions, and question difficulty. The researcher must ensure that
the respondents in the pre-test should be similar to those who are to be finally surveyed.
RELIABILITY TEST
Reliability is a measure of the stability or consistency of test scores. You can also think of it as
the ability for a test or research findings to be repeatable. For example, a medical thermometer
is a reliable tool that would measure the correct temperature each time it is used. In the same
way, a reliable math test will accurately measure mathematical knowledge for every student
who takes it and reliable research findings can be replicated over and over. Of course, it’s not
quite as simple as saying one think a test is reliable. There are many statistical tools you can
use to measure reliability. For example:
Kuder-Richardson 20: a measure of internal reliability for a binary test (i.e. one with
right or wrong answers).
Cronbach’s alpha: measures internal reliability for tests with multiple possible answers.
Internal reliability, or internal consistency, is a measure of how well your test is actually
measuring what you want it to measure.
External reliability means that your test or measure can be generalized beyond what you’re
using it for. For example, a claim that individual tutoring improves test scores should apply to
more than one subject. A test for depression should be able to detect depression in different
age groups, for people in different socio-economic statuses, or introverts.
47
Pearson’s correlation can be used to estimate the theoretical reliability coefficient
between parallel tests.
The Spearman Brown formula is a measure of reliability for split-half tests.
Cohen’s Kappa measures interrater reliability.
The range of the reliability coefficient is from 0 to 1.
VALIDITY TEST
Validity simply means that a test or instrument is accurately measuring what it’s supposed to.
Reliability implies consistency: if you take the ACT five times, you should get roughly the
same results every time. A test is valid if it measures what it’s supposed to. When a measure
has good test-retest reliability and internal consistency, researchers should be more confident
that the scores represent what they are supposed to. There has to be more to it, however, because
a measure can be extremely reliable but have no validity whatsoever.
As an absurd example, imagine someone who believes that people’s index finger length reflects
their self-esteem and therefore tries to measure self-esteem by holding a ruler up to people’s
index fingers. Although this measure would have extremely good test-retest reliability, it would
have absolutely no validity. The fact that one person’s index finger is a centimetre longer than
another’s would indicate nothing about which one had higher self-esteem.
Discussions of validity usually divide it into several distinct “types.” But a good way to
interpret these types is that they are other kinds of evidence—in addition to reliability—that
should be taken into account when judging the validity of a measure. Here we consider three
basic kinds: face validity, content validity, and criterion validity.
Face validity is the extent to which a measurement method appears “on its face” to
measure the construct of interest. Most people would expect a self-esteem questionnaire
to include items about whether they see themselves as a person of worth and whether
they think they have good qualities. So a questionnaire that included these kinds of
items would have good face validity. The finger-length method of measuring self-
esteem, on the other hand, seems to have nothing to do with self-esteem and therefore
has poor face validity. Face validity is at best a very weak kind of evidence that a
measurement method is measuring what it is supposed to. One reason is that it is based
on people’s intuitions about human behavior, which are frequently wrong.
Content validity is the extent to which a measure “covers” the construct of interest.
For example, if a researcher conceptually defines test anxiety as involving both
48
sympathetic nervous system activation (leading to nervous feelings) and negative
thoughts, then his measure of test anxiety should include items about both nervous
feelings and negative thoughts. Or consider that attitudes are usually defined as
involving thoughts, feelings, and actions toward something. By this conceptual
definition, a person has a positive attitude toward exercise to the extent that he or she
thinks positive thoughts about exercising, feels good about exercising, and actually
exercises. So to have good content validity, a measure of people’s attitudes toward
exercise would have to reflect all three of these aspects. Like face validity, content
validity is not usually assessed quantitatively. Instead, it is assessed by carefully
checking the measurement method against the conceptual definition of the construct.
Criterion validity is the extent to which people’s scores on a measure are correlated
with other variables (known as criteria) that one would expect them to be correlated
with. For example, people’s scores on a new measure of test anxiety should be
negatively correlated with their performance on an important school exam. If it were
found that people’s scores were in fact negatively correlated with their exam
performance, then this would be a piece of evidence that these scores really represent
people’s test anxiety. But if it were found that people scored equally well on the exam
regardless of their test anxiety scores, then this would cast doubt on the validity of the
measure.
A criterion can be any variable that one has reason to think should be correlated with
the construct being measured, and there will usually be many of them. When the
criterion is measured at the same time as the construct, criterion validity is referred to
as concurrent validity; however, when the criterion is measured at some point in the
future (after the construct has been measured), it is referred to as predictive validity
(because scores on the measure have “predicted” a future outcome).
49
MODULE 3: SAMPLING METHODS AND FIELD PROCEDURE
Learning Outcomes:
Develop and use sampling design process
Take statistical approach to determining sample size and the construction of
confidence intervals.
SAMPLE DESIGN
The researcher must decide the way of selecting a sample, which is popularly known as sample
design. Sample design is a definite plan determined before any data are actually collected for
obtaining a sample from a given population.
SAMPLING
Sampling is a process of selecting representative units from an entire population of a study.
The process of selecting a number of individuals for a study in such a way that the individuals
represent the larger group from which they were selected. In other words, sampling is the
selection of some part of an aggregate or a whole on the basis of which judgments or inferences
about the aggregate or mass is made. It is a process of obtaining information regarding a
phenomenon about entire population by examining a part of it.
A good sample is one which satisfies all or few of the following conditions:
Representativeness: When sampling method is adopted by the researcher, the basic
assumption is that the samples so selected out of the population are the best
representative of the population under study. Thus good samples are those who
accurately represent the population. Probability sampling technique yield
representative samples. On measurement terms, the sample must be valid. The validity
of a sample depends upon its accuracy.
Accuracy: Accuracy is defined as the degree to which bias is absent from the sample.
An accurate (unbiased) sample is one which exactly represents the population. It is free
from any influence that causes any differences between sample value and population
value.
50
Size: A good sample must be adequate in size and reliable. The sample size should be
such that the inferences drawn from the sample are accurate to a given level of
confidence to represent the entire population under study.
Purpose of Sampling
Economical
Improved Quality of Data
Quick Study Results
Precision and Accuracy of Data
The whole purpose of sampling is to gather data about the population in order to make an
inference that can be generalized to the population.
51
Sampling Process
An operational sampling process can be divided into seven steps as given below:
Defining the target population.
Specifying the sampling frame.
Specifying the sampling unit.
Selection of the sampling method.
Determination of sample size.
Specifying the sampling plan.
Selecting the sample.
52
contains all the households that have a monthly income above Rs.20,000. However, in
practice it is difficult to get an exhaustive sampling frame that exactly fits the
requirements of a particular research. In general, researchers use easily available
sampling frames like telephone directories and lists of credit card and mobile phone
users. Various private players provide databases developed along various demographic
and economic variables. Sometimes, maps and aerial pictures are also used as sampling
frames. Whatever may be the case, an ideal sampling frame is one that entire population
and lists the names of its elements only once.
A sampling frame error pops up when the sampling frame does not accurately represent
the total population or when some elements of the population are missing another
drawback in the sampling frame is over –representation. A telephone directory can be
over represented by names/household that have two or more connections.
53
primary importance and are worth mentioning are whether the technique deals with
fixed or sequential sampling and whether its logic is based on traditional or Bayesian
methods. In non-probability sampling procedures, the allocation of budget, thumb rules
and number of sub groups to be analyzed, importance of the decision, number of
variables, nature of analysis, incidence rates, and completion rates play a major role in
sample size determination. In the case of probability sampling, however, formulas are
used to calculate the sample size after the levels of acceptable error and level of
confidence are specified.
54
Sampling Methods
Probability Sampling: A method of sampling that enables the researcher to specify for each
case in the population the probability of its inclusion in the sample. A probability sampling
scheme is one in which every unit in the population has a chance (greater than zero) of being
selected in the sample, and this probability can be accurately determined. The combination of
these traits makes it possible to produce unbiased estimates of population totals, by weighting
sampled units according to their probability of selection.
Non-Probability Sampling:
Nonprobability sampling is any sampling method where some elements of the population have
no chance of selection (these are sometimes referred to as ‘out of coverage’/’undercovered’),
or where the probability of selection can’t be accurately determined. It involves the selection
of elements based on assumptions regarding the population of interest, which forms the criteria
for selection. Hence, because the selection of elements is non-random, nonprobability sampling
does not allow the estimation of sampling errors. These conditions place limits on how much
information a sample can provide about the population. Information about the relationship
between sample and population is limited, making it difficult to extrapolate from the sample to
the population.
Systematic Sampling:
Systematic sampling involves the selection of every kth element from a sampling frame.
Here ‘k’ represents the skip interval and is calculated using the following formulae.
Skip interval (k) = population size/Sample size
Often used as a substitute to simple random sample, it involves the selection of units
from a list using a skip interval (k) so that every k’th element on the list, following a
random start between 1 and k, is included in the sample. For ex, if k were to equal 6,
and the random start were2, then the sample would consists of 2nd, 8th, 14th, 20th
…….elements of the sampling frame.
It is to be noted here that if the skip interval is not a whole number then it is rounded
off to the nearest whole number. This sampling method can be used industrial
operations where the equipment and machinery in the production line are checked for
proper functioning as per the specifications. The manufacturer can select every k’th
item to ensure consistent quality or for detection of defects. Therefore, he requires the
first item to be selected at the random as the starting point and subsequently he can
choose every k’th item for evaluation against specifications. It also finds its
applicability while questioning people in a sample survey where the interviewer may
catch hold of every 10th person entering a particular shop. However, in every case, the
researcher has to determine the skip interval and proceed thereafter. In both the cases,
it is necessary to select the first item in the population in a random manner and thereafter
follow the skip interval. This method is more economical and less time consuming than
simple random sampling.
56
independently. This often improves the representativeness of the sample by reducing
the sampling error.
The number of units drawn for sampling from each stratum depends on the
homogeneity of the elements. A smaller sample can be drawn from the known to have
the elements with the same value whereas sample can be drawn in much higher
proportion from another stratum where the values are known to differ. This is because
in the former case the information from the smaller number of respondents can be
enumerated to the whole sample stratum. However in the latter case with much
variability among the elements the higher elements value will keep the sampling to
minimum errors to minimum value. The smaller errors may be due to groups are
appreciably represented when strata are combined.
Proportionate stratified sample – The size of the sample selected from each subgroup
is proportional to the size of that subgroup in the entire population.
Disproportionate stratified sample – The size of the sample selected from each
subgroup is disproportional to the size of that subgroup in the population.
Cluster sampling:
Clustering involves grouping the population into various clusters and selecting few
clusters for study. Cluster sampling is suitable for conducting research studies that
cover large geographic area. Once the cluster is formed the researcher can either go for
one stage, two stages, or multi stage cluster sampling. In single stage, all the elements
from each selected are studied, whereas in two stages, the researchers use random to
select few elements from clusters.
Multistage sampling involves selecting a sample in two or more successive stages.
Here the cluster selected in the first stage can be divided into cluster units.
For example consider the case where a company decides to interview 400 households
about the likeability of its new detergent in a metropolitan city. To minimize the
resources and time researchers divide the city into separate blocks say 40, each block
consist of heterogeneous units. The researcher may opt for the two stage cluster
sampling if he finds that individual clusters have little heterogeneity to other clusters.
Similarly a multi stage cluster sampling involves three or more sampling steps, it differs
from stratified sampling that is done in cluster in contrast to elements within strata as
is the case in the stratified sampling. Elements are randomly selected from each stratum
57
in each stratum in case of stratified sampling whereas only selected clusters are studied
in cluster sampling.
BASIS FOR
STRATIFIED SAMPLING CLUSTER SAMPLING
COMPARISON
Sample Randomly selected individuals All the individuals are taken from
are taken from all the strata. randomly selected clusters.
58
Non-Probability Sampling Methods
Sample Size:
The sample size is an important feature of any empirical study in which the goal is to make
inferences about a population from a sample. Sample size determination is the act of choosing
the number of observations or replicates to include in a statistical sample. The thumb rule has
been proposed for determining sample size where sample size larger than 30 and less than 500
are appropriate for most research, and the minimum size of sample should be 30% of the
population.
60
The size of the sample depends on a number of factors and the researchers have to give the
statistically information before they can get an answer. For example, confidence level, standard
deviation, margin of error and population size are considered to determine the sample size. The
following points are taken into account for determining the sample size in this approach.
The variability of the population: Higher the variability as measured by the population
standard deviation, larger will be the size of the sample.
The confidence attached to the estimate: Higher the confidence the researcher wants for
the estimate, larger will be sample size.
The allowable error or margin of error: Greater the precision the research seeks, larger
would be the size of the sample.
Sample size for estimating population mean - The formula for determining sample size is given
as:
Where
n = Sample size
σ = Population standard deviation
e = Margin of error
Z = The value for the given confidence interval
Field Research
Field research is defined as a qualitative method of data collection that aims to observe, interact
and understand people while they are in a natural environment. For example, nature
conservationists observe behavior of animals in their natural surroundings and the way they
react to certain scenarios. In the same way, social scientists conducting field research may
conduct interviews or observe people from a distance to understand how they behave in a social
environment and how they react to situations around them.
61
Steps in conducting Field Research:
Build the Right Team
Recruiting People for Study
Data Collection Methodology
Site Visit
Data Analysis
Communicating the Results
62
MODULE 4: STATISTICAL TESTING
Learning Outcomes:
Turn statistical findings into marketing information that gets attention.
Concepts of correlation and linear functions.
Analysis of data is a process of inspecting, cleaning, transforming, and modeling data with the
goal of discovering useful information, suggesting conclusions, and supporting decision-
making. In statistical applications, some people divide data analysis into exploratory data
analysis (EDA), descriptive statistics, and inferential Statistics.
Exploratory data analysis (EDA) is an approach to analyzing data sets to summarize their
main characteristics, often with visual methods. A statistical model can be used or not, but
primarily EDA is for seeing what the data can tell us beyond the formal modeling or hypothesis
testing task. Exploratory data analysis was promoted by John Tukey to encourage statisticians
to explore the data, and possibly formulate hypotheses that could lead to new data collection
and experiments.
There are a number of tools that are useful for EDA, but EDA is characterized more by the
attitude taken than by particular techniques.
Typical graphical techniques used in EDA are:
Box plot
Histogram
Multi-vari chart
Pareto chart
Scatter plot
Stem-and-leaf plot
Descriptive statistics is the term given to the analysis of data that helps describe, show or
summarize data in a meaningful way such that patterns might emerge from the data. Descriptive
statistics do not, however, allow us to make conclusions beyond the data we have analysed or
reach conclusions. They are simply a way to describe our data. Descriptive statistics therefore
enables us to present the data in a more meaningful way, which allows simpler interpretation
of the data. For example, if we had the results of 100 pieces of students' coursework, we may
63
be interested in the overall performance of those students. We would also be interested in the
distribution or spread of the marks. Descriptive statistics allow us to do this.
Typically, there are two general types of statistic that are used to describe data:
Measures of central tendency: These are ways of describing the central position of a
frequency distribution for a group of data. In this case, the frequency distribution is simply the
distribution and pattern of marks scored by the 100 students from the lowest to the highest. We
can describe this central position using a number of statistics, including the mode, median, and
mean.
Measures of spread: These are ways of summarizing a group of data by describing how spread
out the scores are. For example, the mean score of our 100 students may be 65 out of 100.
However, not all students will have scored 65 marks. Rather, their scores will be spread out.
Some will be lower and others higher. Measures of spread help us to summarize how spread
out these scores are. To describe this spread, a number of statistics are available to us, including
the range, quartiles, absolute deviation, variance and standard deviation.
Inferential statistics are techniques that allow us to use these samples to make generalizations
about the populations from which the samples were drawn. It is, therefore, important that the
sample accurately represents the population. Inferential statistics arise out of the fact that
sampling naturally incurs sampling error and thus a sample is not expected to perfectly
represent the population. The methods of inferential statistics are (1) the estimation of
parameter(s) and (2) testing of statistical hypotheses.
HYPOTHESIS
A hypothesis may be defined as a proposition or a set of proposition set forth as an explanation
for the occurrence of some specified group of phenomena either asserted merely as a
provisional conjecture to guide some investigation or accepted as highly probable in the light
of established facts. Quite often a research hypothesis is a predictive statement, capable of
being tested by scientific methods, that relates an independent variable to some dependent
variable.
64
These are hypotheses capable of being objectively verified and tested. Thus, we may conclude
that a hypothesis states what we are looking for and it is a proposition which can be put to a
test to determine its validity.
CHARACTERISTICS OF HYPOTHESIS
65
NULL AND ALTERNATE HYPOTHESIS
66
DATA ANALYSIS OUTCOME
Level of Significance:
The significance level, also denoted as alpha or α, is the probability of rejecting the null
hypothesis when it is true (type 1 error). For example, a significance level of 0.05 indicates a
5% risk of concluding that a difference exists when there is no actual difference.
Critical Value
In hypothesis testing, a critical value is a point on the test distribution that is compared to the
test statistic to determine whether to reject the null hypothesis. If the absolute value of your
test statistic is greater than the critical value, you can declare statistical significance and reject
the null hypothesis.
67
one group consisting of values which appear to be consistent with the null
hypothesis and
the other having values which are unlikely to occur if Ho is true.
The first group is called the acceptance region and the second set of values is known as the
rejection region for a test.
The rejection region is also called the critical region. The value(s) that separates the critical
region from the acceptance region is called the critical value(s). The critical value which can
be in the same units as the parameter or in the standardized units, is to be decided by the
experimenter keeping in view the degree of confidence they are willing to have in the null
hypothesis.
As has been stated above that hypothesis testing determines the validity of the assumption
(technically described as null hypothesis) with a view to choose between two conflicting
hypotheses about the value of a population parameter. Hypothesis testing helps to decide on
the basis of a sample data, whether a hypothesis about the population is likely to be true or
false. Statisticians have developed several tests of hypotheses (also known as the tests of
significance) for the purpose of testing of hypotheses which can be classified as: (a) Parametric
tests or standard tests of hypotheses; and (b) Non-parametric tests or distribution-free test of
hypotheses.
Parametric tests usually assume certain properties of the parent population from which we
draw samples. Assumptions like observations come from a normal population, sample size is
large, assumptions about the population parameters like mean, variance, etc., must hold good
before parametric tests can be used. But there are situations when the researcher cannot or does
not want to make such assumptions. In such situations we use statistical methods for testing
hypotheses which are called non-parametric tests because such tests do not depend on any
assumption about the parameters of the parent population. Besides, most non-parametric tests
68
assume only nominal or ordinal data, whereas parametric tests require measurement equivalent
to at least an interval scale. As a result, non-parametric tests need more observations than
parametric tests to achieve the same size of Type I and Type II errors.
t-test is based on t-distribution and is considered an appropriate test for judging the significance
of a sample mean or for judging the significance of difference between the means of two
samples in case of small sample(s) when population variance is not known (in which case we
use variance of the sample as an estimate of the population variance). In case two samples are
69
related, we use paired t-test (or what is known as difference test) for judging the significance
of the mean of difference between the two related samples. It can also be used for judging the
significance of the coefficients of simple and partial correlations. The relevant test statistic, t,
is calculated from the sample data and then compared with its probable value based on t-
distribution (to be read from the table that gives probable values of t for different levels of
significance for different degrees of freedom) at a specified level of significance for concerning
degrees of freedom for accepting or rejecting the null hypothesis. It may be noted that t-test
applies only in case of small sample(s) when population variance is unknown.
χ2 -test is based on chi-square distribution and as a parametric test is used for comparing a
sample variance to a theoretical population variance.
Example:
70
IMPORTANT CHARACTERISTICS OF χ2 TEST
This test (as a non-parametric test) is based on frequencies and not on the parameters
like mean and standard deviation.
The test is used for testing the hypothesis and is not useful for estimation.
This test possesses the additive property as has already been explained.
This test can also be applied to a complex contingency table with several classes and as
such is a very useful test in research work.
This test is an important non-parametric test as no rigid assumptions are necessary in
regard to the type of population, no need of parameter values and relatively less
mathematical details are involved.
F-test is based on F-distribution and is used to compare the variance of the two-independent
samples. This test is also used in the context of analysis of variance (ANOVA) for judging the
significance of more than two sample means at one and the same time. It is also used for judging
the significance of multiple correlation coefficients. Test statistic, F, is calculated and
compared with its probable value (to be seen in the F-ratio tables for different degrees of
freedom for greater and smaller variances at specified level of significance) for accepting or
rejecting the null hypothesis.
71
LIMITATIONS OF THE TESTS OF HYPOTHESES
We have described above some important test often used for testing hypotheses on the basis of
which important decisions may be based. But there are several limitations of the said tests
which should always be borne in mind by a researcher. Important limitations are as follows:
The tests should not be used in a mechanical fashion. It should be kept in view that
testing is not decision-making itself; the tests are only useful aids for decision-making.
Hence “proper interpretation of statistical evidence is important to intelligent
decisions.”6
Test do not explain the reasons as to why does the difference exist, say between the
means of the two samples. They simply indicate whether the difference is due to
fluctuations of sampling or because of other reasons but the tests do not tell us as to
which is/are the other reason(s) causing the difference.
Results of significance tests are based on probabilities and as such cannot be expressed
with full certainty. When a test shows that a difference is statistically significant, then
it simply suggests that the difference is probably not due to chance.
Statistical inferences based on the significance tests cannot be said to be entirely correct
evidences concerning the truth of the hypotheses. This is specially so in case of small
samples where the probability of drawing erring inferences happens to be generally
higher. For greater reliability, the size of samples be sufficiently enlarged.
All these limitations suggest that in problems of statistical significance, the inference
techniques (or the tests) must be combined with adequate knowledge of the subject-matter
along with the ability of good judgement.
72