BBA RM Notes
BBA RM Notes
BBA RM Notes
INTRODUCTION
(a) the sampling design that deals with the method of selecting items to be
observed for the selected study;
(b) the observational design that relates to the conditions under which the
observations are to be made;
(c) the statistical design that concerns with the question of how many items are
to be observed, and how the information and data gathered are to be
analysed; and
(d) the operational design that deals with the techniques by which the
procedures specified in the sampling, statistical and observational designs
can be carried out.
(ii) it constitutes a strategy that specifies the methods of data collection and
analysis which would be adopted; and
(iii) it also specifies the time period of research and monetary budget involved
in conducting the study, which comprise the two major constraints of
undertaking any research.
1.4.2 Concepts Relating to Research Design:
Some of the important concepts relating to Research Design are
discussed below:
1. Dependent and Independent Variables:
A magnitude that varies is known as a variable. The concept may
assume different quantitative values like height, weight, income etc. Qualitative
variables are not quantifiable in the strictest sense of the term. However, the
qualitative phenomena may also be quantified in terms of the presence or
absence of the attribute(s) considered. The phenomena that assume different
values quantitatively even in decimal points are known as ‘continuous
variables’. But all variables need not be continuous. Values that can be
expressed only in integer values are called ‘non-continuous variables’. In
statistical terms, they are also known as ‘discrete variables’. For example, age
is a continuous variable, whereas the number of children is a non-continuous
variable. When changes in one variable depend upon the changes in other
variable or variables, it is known as a dependent or endogenous variable, and the
variables that cause the changes in the dependent variable are known as the
independent or explanatory or exogenous variables. For example, if demand
depends upon price, then demand is a dependent variable, while price is the
independent variable. And, if more variables determine demand, like income
and price of the substitute commodity, then demand also depends upon them in
addition to the price of original commodity. In other words, demand is a
dependent variable which is determined by the independent variables like price
of the original commodity, income and price of substitutes.
2 Extraneous Variable:
The independent variables which are not directly related to the purpose
of the study but affect the dependent variable are known as extraneous variables.
For instance, assume that a researcher wants to test the hypothesis that there is a
relationship between children’s school performance and their self-concepts, in
which case the latter is an independent variable and the former, a dependent
variable. In this context, intelligence may also influence the school
performance. However, since it is not directly related to the purpose of the study
undertaken by the researcher, it would be known as an extraneous variable. The
influence caused by the extraneous variable(s) on the dependent variable is
technically called the ‘experimental error’. Therefore, a research study should
always be framed in such a manner that the influence of extraneous variables on
the dependent variable/s is completely controlled, and the influence of
independent variable/s is clearly evident.
3. Control:
One of the most important features of a good research design is to
minimize the effect of extraneous variable(s). Technically, the term ‘control’ is
used when a researcher designs the study in such a manner that it minimizes the
effects of extraneous variables. The term ‘control’ is used in experimental
research to reflect the restrain in experimental conditions.
4. Confounded Relationship:
The relationship between the dependent and independent variables is
said to be confounded by an extraneous variable, when the dependent variable is
not free from its effects.
5. Research Hypothesis:
When a prediction or a hypothesized relationship is tested by adopting
scientific methods, it is known as research hypothesis. The research hypothesis
is a predictive statement which relates to a dependent variable and an
independent variable. Generally, a research hypothesis must consist of at least
one dependent variable and one independent variable. Whereas, the
relationships that are assumed but not to be tested are predictive statements that
are not to be objectively verified, thus are not classified as research hypotheses.
8. Treatments:
Treatments refer to the different conditions to which the experimental
and control groups are subject to. In the example considered, the two treatments
are the parents with regular earnings and those with no regular earnings.
Likewise, if a research study attempts to examine through an experiment the
comparative effect of three different types of fertilizers on the yield of rice crop,
then the three types of fertilizers would be treated as the three treatments.
9. Experiment:
Experiment refers to the process of verifying the truth of a statistical
hypothesis relating to a given research problem. For instance, an experiment
may be conducted to examine the yield of a certain new variety of rice crop
developed. Further, Experiments may be categorized into two types, namely,
‘absolute experiment’ and ‘comparative experiment’. If a researcher wishes to
determine the impact of a chemical fertilizer on the yield of a particular variety
of rice crop, then it is known as absolute experiment. Meanwhile, if the
researcher wishes to determine the impact of chemical fertilizer as compared to
the impact of bio-fertilizer, then the experiment is known as a comparative
experiment.
10. Experimental Unit(s):
Experimental Units refer to the pre-determined plots, characteristics or
the blocks, to which different treatments are applied. It is worth mentioning
here that such experimental units must be selected with great caution.
(iii) The crucial role of the family-group in transmitting the culture should be
recognized. This means, as an individual is the member of a family, the
role of the family in shaping his/her behaviour should never be ignored.
(ii) A hypothesis must be capable of being put to test. Quite often, the
research programmes fail owing to its incapability of being subject to
testing for validity. Therefore, some prior study may be conducted by the
researcher in order to make a hypothesis testable. A hypothesis “is tested
if other deductions can be made from it, which in turn can be confirmed or
disproved by observation” (Kothari, 1988).
(iii) A hypothesis must state relationship between two variables, in the case of
relational hypotheses.
(iv) A hypothesis must be specific and limited in scope. This is because a
simpler hypothesis generally would be easier to test for the researcher.
And therefore, he/she must formulate such hypotheses.
(v) As far as possible, a hypothesis must be stated in the simplest language, so
as to make it understood by all concerned. However, it should be noted that
simplicity of a hypothesis is not related to its significance.
(vi) A hypothesis must be consistent and derived from the most known facts.
In other words, it should be consistent with a substantial body of
established facts. That is, it must be in the form of a statement which
Judges accept as being the most likely to occur.
(vii) A hypothesis must be amenable to testing within a stipulated or reasonable
period of time. No matter how excellent a hypothesis, a researcher should
not use it if it cannot be tested within a given period of time, as no one can
afford to spend a life-time on collecting data to test it.
(viii) A hypothesis should state the facts that give rise to the necessity of looking
for an explanation. This is to say that by using the hypothesis, and other
known and accepted generalizations, a researcher must be able to derive
the original problem condition. Therefore, a hypothesis should explain
what it actually wants to explain, and for this it should also have an
empirical reference.
1.6.2 Concepts Relating to Testing of Hypotheses:
Testing of hypotheses requires a researcher to be familiar with various
concepts concerned with it such as:
Before the sample is drawn, the researcher has to state the null
hypothesis and the alternative hypothesis. While formulating the null
hypothesis, the following aspects need to be considered:
(a) Alternative hypothesis is usually the one which a researcher wishes to prove,
whereas the null hypothesis is the one which he/she wishes to disprove. Thus, a
null hypothesis is usually the one which a researcher tries to reject, while an
alternative hypothesis is the one that represents all other possibilities.
(b) The rejection of a hypothesis when it is actually true involves great risk, as it
indicates that it is a null hypothesis because then the probability of rejecting it
when it is true is α (i.e., the level of significance) which is chosen very small.
(c) Null hypothesis should always be specific hypothesis i.e., it should not state
about or approximately a certain value.
(2) The Level of Significance:
In the context of hypothesis testing, the level of significance is a very
important concept. It is a certain percentage that should be chosen with great
care, reason and thought. If for instance, the significance level is taken at 5 per
cent, then it means that H0 would be rejected when the sampling result has a less
than 0.05 probability of occurrence when H0 is true. In other words, the five per
cent level of significance implies that the researcher is willing to take a risk of
five per cent of rejecting the null hypothesis, when (H0) is actually true. In sum,
the significance level reflects the maximum value of the probability of rejecting
H0 when it is actually true, and which is usually determined prior to testing the
hypothesis.
(3) Test of Hypothesis or Decision Rule:
Suppose the given hypothesis is H0 and the alternative hypothesis Ha,
then the researcher has to make a rule known as the decision rule. According to
the decision rule, the researcher accepts or rejects H 0. For example, if the H0 is
that certain students are good against the Ha that all the students are good, then
the researcher should decide the number of items to be tested and the criteria on
the basis of which to accept or reject the hypothesis.
(4) Type I and Type II Errors:
As regards the testing of hypotheses, a researcher can make basically two
types of errors. He/she may reject H0 when it is true, or accept H0 when it is
not true. The former is called as Type I error and the latter is known as Type II
error. In other words, Type I error implies the rejection of a hypothesis when it
must have been accepted, while Type II error implies the acceptance of a
hypothesis which must have been rejected. Type I error is denoted by α (alpha)
and is known as α error, while Type II error is usually denoted by β (beta) and is
known as β error.
(5) One-tailed and two-tailed Tests:
These two types of tests are very important in the context of hypothesis
testing. A two-tailed test rejects the null hypothesis, when the sample mean is
significantly greater or lower than the hypothesized value of the mean of the
population. Such a test is suitable when the null hypothesis is some specified
value, the alternative hypothesis is a value that is not equal to the specified value
of the null hypothesis.
1.6.3 Procedure of Hypothesis Testing:
Testing a hypothesis refers to verifying whether the hypothesis is valid
or not. Hypothesis testing attempts to check whether to accept or not to accept
the null hypothesis. The procedure of hypothesis testing includes all the steps
that a researcher undertakes for making a choice between the two alternative
actions of rejecting or accepting a null hypothesis. The various steps involved in
hypothesis testing are as follows:
(i) Making a Formal Statement:
This step involves making a formal statement of the null hypothesis (H0)
and the alternative hypothesis (Ha). This implies that the hypotheses should be
clearly stated within the purview of the research problem. For example, suppose
a school teacher wants to test the understanding capacity of the students which
must be rated more than 90 per cent in terms of marks, the hypotheses may be
stated as follows:
Null Hypothesis H0 : = 100
Alternative Hypothesis Ha : > 100
Sampling error can be computed for a given sample size and design. The
measurement of sampling error is known as ‘precision of the sampling plan’.
When the sample size is increased, the precision can be improved. However,
increasing the sample size has its own limitations. The large sized sample not
only increases the cost of data collection, but also increases the systematic bias.
Thus, an effective way of increasing the precision is generally to choose a better
sampling design, which has smaller sampling error for a given sample size at a
specified cost. In practice, however, researchers generally prefer a less precise
design owing to the ease in adopting the same, in addition to the fact that
systematic bias can be controlled better way in such designs.
In sum, while selecting the sample, a researcher should ensure that the
procedure adopted involves a relatively smaller sampling error and helps to
control systematic bias.
1.7.3 Characteristics of a Good Sample Design:
The following are the characteristic features of a good sample design:
(a) the sample design should yield a truly representative sample;
(b) the sample design should be such that it results in small sampling error;
(c) the sample design should be viable in the context of budgetary
constraints of the research study;
(d) the sample design should be such that the systematic bias can be
controlled; and
(e) the sample must be such that the results of the sample study would be
applicable, in general, to the universe at a reasonable level of confidence.
I. Non-Probability Sampling:
Non-probability sampling is the sampling procedure that does not afford any
basis for estimating the probability that each item in the population would have
an equal chance of being included in the sample. Non-probability sampling is
also known as deliberate sampling, judgment sampling and purposive sampling.
Under this type of sampling, the items for the sample are deliberately chosen by
the researcher; and his/her choice concerning the choice of items remains
supreme. In other words, under non-probability sampling the researchers select
a particular unit of the universe for forming a sample on the basis that the small
number that is thus selected out of a huge one would be typical or representative
of the whole population. For example, to study the economic conditions of
people living in a state, a few towns or village may be purposively selected for
an intensive study based on the principle that they are representative of the
entire state. In such a case, the judgment of the researcher of the study assumes
prime importance in this sampling design.
Quota Sampling:
Quota sampling is also an example of non-probability sampling. Under
this sampling, the researchers simply assume quotas to be filled from different
strata, with certain restrictions imposed on how they should be selected. This
type of sampling is very convenient and is relatively less expensive. However,
the samples selected using this method certainly do not satisfy the characteristics
of random samples. They are essentially judgment samples and inferences
drawn based on that would not be amenable to statistical treatment in a formal
way.
II. Probability Sampling:
Probability sampling is also known as ‘choice sampling’ or ‘random sampling’.
Under this sampling design, every item of the universe has an equal chance of
being included in the sample. In a way, it is a lottery method under which
individual units are selected from the whole group, not deliberately, but by using
some mechanical process. Therefore, only chance would determine whether an
item or the other would be included in the sample or not. The results obtained
from probability or random sampling would be assured in terms of probability.
That is, the researcher can measure the errors of estimation or the significance of
results obtained from the random sample. This is the superiority of random
sampling design over the deliberate sampling design. Random sampling
satisfies the law of Statistical Regularity, according to which if on an average the
sample chosen is random, then it would have the same composition and
characteristics of the universe. This is the reason why the random sampling
method is considered the best technique of choosing a representative sample.
Introduction:
It is important for a researcher to know the sources of data which he
requires for different purposes. Data are nothing but the information. There are
two sources of information or data - Primary data and Secondary data. Primary
data refers to the data collected for the first time, whereas secondary data refers
to the data that have already been collected and used earlier by somebody or
some agency. For example, the statistics collected by the Government of India
relating to the population is primary data for the Government of India since it
has been collected for the first time. Later when the same data are used by a
researcher for his study of a particular problem, then the same data become the
secondary data for the researcher. Both the sources of information have their
merits and demerits. The selection of a particular source depends upon the (a)
purpose and scope of enquiry, (b) availability of time, (c) availability of finance,
(d) accuracy required, (e) statistical tools to be used, (f) sources of information
(data), and (g) method of data collection.
(a) Purpose and Scope of Enquiry: The purpose and scope of data
collection or survey should be clearly set out at the very beginning. It requires
the clear statement of the problem indicating the type of information which is
needed and the use for which it is needed. If for example, the researcher is
interested in knowing the nature of price change over a period of time, it would
be necessary to collect data of commodity prices. It must be decided whether it
would be helpful to study wholesale or retail prices and the possible uses to
which such information could be put. The objective of an enquiry may be either
to collect specific information relating to a problem or adequate data to test a
hypothesis. Failure to set out clearly the purpose of enquiry is bound to lead to
confusion and waste of resources.
After the purpose of enquiry has been clearly defined, the next step is to
decide about the scope of the enquiry. Scope of the enquiry means the coverage
with regard to the type of information, the subject-matter and geographical area.
For instance, an enquiry may relate to India as a whole or a state or an industrial
town wherein a particular problem related to a particular industry can be studied.
(f) Sources of Information (data): After deciding about the unit, a researcher
has to decide about the source from which the information can be obtained or
collected. For any statistical inquiry, the investigator may collect the data first
hand or he may use the data from other published sources, such as publications
of the government/semi-government organizations or journals and magazines
etc.
(g) Method of Data Collection: - There is no problem if secondary data are
used for research. However, if primary data are to be collected, a decision has to
be taken whether (i) census method or (ii) sample technique is to be used for
data collection. In census method, we go for total enumeration i.e., all the units
of a universe have to be investigated. But in sample technique, we inspect or
study only a selected representative and adequate fraction of the population and
after analyzing the results of the sample data we draw conclusions about the
characteristics of the population. Selection of a particular technique becomes
difficult because where population or census method is more scientific and
100% accuracy can be attained through this method, choosing this becomes
difficult because it is time taking, it requires more labor and it is very expensive.
Therefore, for a single researcher or for a small institution it proves to be
unsuitable. On the other hand, sample method is less time taking, less laborious
and less expensive but a 100% accuracy cannot be attained through this method
because of sampling and non-sampling errors attached to this method. Hence, a
researcher has to be very cautious and careful while choosing a particular
method.
Methods of Collecting Primary Data:
Primary data may be obtained by applying any of the following methods:
1. Direct Personal Interviews.
2. Indirect oral interviews.
3. Information from correspondents.
4. Mailed questionnaire methods.
5. Schedule sent through enumerators.
1. Direct personal interviews: A face to face contact is made with the
informants (persons from whom the information is to be obtained) under this
method of collecting data. The interviewer asks them questions pertaining to the
survey and collects the desired information. Thus, if a person wants to collect
data about the working conditions of the workers of the Tata Iron and Steel
Company, Jamshedpur, he would go to the factory, contact the workers and
obtain the desired information. The information collected in this manner is first
hand and also original in character. There are many merits and demerits of this
method, which are discussed as under:
Merits:
1. Most often respondents are happy to pass on the information required
from them when contacted personally and thus response is encouraging.
2. The information collected through this method is normally more accurate
because interviewer can clear doubts of the informants about certain
questions and thus obtain correct information. In case the interviewer
apprehends that the informant is not giving accurate information, he may
cross-examine him and thereby try to obtain the information.
3. This method also provides the scope for getting supplementary
information from the informant, because while interviewing it is possible
to ask some supplementary questions which may be of greater use later.
4. There might be some questions which the interviewer would find
difficult to ask directly, but with some tactfulness, he can mingle such
questions with others and get the desired information. He can twist the
questions keeping in mind the informant’s reaction. Precisely, a delicate
situation can usually he handled more effectively by a personal interview
than by other survey techniques.
5. The interviewer can adjust the language according to the status and
educational level of the person interviewed, and thereby can avoid
inconvenience and misinterpretation on the part of the informant.
Demerits:
2. There is a greater chance of personal bias and prejudice under this method as
compared to other methods.
Conclusion:
Though there are some demerits in this method of data collection still we cannot
say that it is not useful. The matter of fact is that this method is suitable for
intensive rather than extensive field surveys. Hence, it should be used only in
those cases where intensive study of a limited field is desired.
In the present time of extreme advancement in the communication system,
the investigator instead of going personally and conducting a face to face
interview may also obtain information over telephone. A good number of
surveys are being conducted every day by newspapers and television channels
by sending the reply either by e-mail or SMS. This method has become very
popular nowadays as it is less expensive and the response is extremely quick.
But this method suffers from some serious defects, such as (a) very few people
own a phone or a television and hence a limited number of people can be
approached by this method, (b) only few questions can be asked over phone or
through television, (c) the respondents may give a vague and reckless answers
because answers on phone or through SMS would have to be very short.
2. Indirect Oral Interviews: Under this method of data collection, the
investigator contacts third parties generally called ‘witnesses’ who are capable
of supplying necessary information. This method is generally adopted when the
information to be obtained is of a complex nature and informants are not
inclined to respond if approached directly. For example, when the researcher is
trying to obtain data on drug addiction or the habit of taking liquor, there is high
probability that the addicted person will not provide the desired data and hence
will disturb the whole research process. In this situation taking the help of such
persons or agencies or the neighbours who know them well becomes necessary.
Since these people know the person well, they can provide the desired data.
Enquiry Committees and Commissions appointed by the Government generally
adopt this method to get people’s views and all possible details of the facts
related to the enquiry.
Though this method is very popular, its correctness depends upon a number of
factors which are discussed below:
1. The person or persons or agency whose help is solicited must be of proven
integrity; otherwise any bias or prejudice on their part will not bring the correct
information and the whole process of research will become useless.
2. The ability of the interviewers to draw information from witnesses by means
of appropriate questions and cross-examination.
3. It might happen that because of bribery, nepotism or certain other reasons
those who are collecting the information give it such a twist that correct
conclusions are not arrived at.
Therefore, for the success of this method it is necessary that the evidence of
one person alone is not relied upon. Views from other persons and related
agencies should also be ascertained to find the real position .Utmost care must
be exercised in the selection of these persons because it is on their views that the
final conclusions are reached.
3. Information from Correspondents: The investigator appoints local agents
or correspondents in different places to collect information under this method.
These correspondents collect and transmit the information to the central office
where data are processed. This method is generally adopted by news paper
agencies. Correspondents who are posted at different places supply information
relating to such events as accidents, riots, strikes, etc., to the head office. The
correspondents are generally paid staff or sometimes they may be honorary
correspondents also. This method is also adopted generally by the government
departments in such cases where regular information is to be collected from a
wide area. For example, in the construction of a wholesale price index numbers
regular information is obtained from correspondents appointed in different areas.
The biggest advantage of this method is that it is cheap and appropriate for
extensive investigation. But a word of caution is that it may not always ensure
accurate results because of the personal prejudice and bias of the correspondents.
As stated earlier, this method is suitable and adopted in those cases where the
information is to be obtained at regular intervals from a wide area.
4. Mailed Questionnaire Method: Under this method, a list of questions
pertaining to the survey which is known as ‘Questionnaire’ is prepared and
sent to the various informants by post. Sometimes the researcher himself too
contacts the respondents and gets the responses related to various
questions in the questionnaire. The questionnaire contains questions and
provides space for answers. A request is made to the informants through a
covering letter to fill up the questionnaire and send it back within a specified
time. The questionnaire studies can be classified on the basis of:
i. The degree to which the questionnaire is formalized or structured.
ii. The disguise or lack of disguise of the questionnaire and
iii. The communication method used.
There are certain merits and demerits or limitations of this method of data
collection which are discussed below:
Merits:
1. Questionnaire method of data collection can be easily adopted where the
field of investigation is very vast and the informants are spread over a
wide geographical area.
2. This method is relatively cheap and expeditious provided the informants
respond in time.
3. This method has proved to be superior when compared to other methods like
personal interviews or telephone method. This is because when questions
pertaining to personal nature or the ones requiring reaction by the family are
put forth to the informants, there is a chance for them to be embarrassed in
answering them.
Demerits:
1. This method can be adopted only where the informants are literate
people so that they can understand written questions and lend the
answers in writing.
2. It involves some uncertainty about the response. Co-operation on the part of
informants may be difficult to presume.
3. The information provided by the informants may not be correct and it may
be difficult to verify the accuracy.
However, by following the guidelines given below, this method can be made
more effective:
The questionnaires should be made in such a manner that they do not
become an undue burden on the respondents; otherwise the respondents may not
return them back.
i. Prepaid postage stamp should be affixed
ii. The sample should be large
iii. It should be adopted in such enquiries where it is expected that the
respondents would return the questionnaire because of their own interest
in the enquiry.
iv. It should be preferred in such enquiries where there could be a legal
compulsion to provide the information.
Merits:
The main merits or advantages of this method are listed below:
i. It can be adopted in those cases where informants are illiterate.
ii. There is very little scope of non-response as the enumerators go personally
to obtain the information.
iii. The information received is more reliable as the accuracy of statements can
be checked by supplementary questions wherever necessary.
This method too like others is not free from defects or limitations. The
main limitations are listed below:
Demerits:
i. In comparison to other methods of collecting primary data, this method is
quite costly as enumerators are generally paid persons.
ii. The success of the method depends largely upon the training imparted to
the enumerators.
iii. Interviewing is a very skilled work and it requires experience and training.
Many statisticians have the tendency to neglect this extremely important
part of the data collecting process and this result in bad interviews.
Without good interviewing most of the information collected is of doubtful
value.
iv. Interviewing is not only a skilled work but it also requires a great degree of
politeness and thus the way the enumerators conduct the interview would
affect the data collected. When questions are asked by a number of
different interviewers, it is possible that variations in the personalities of
the interviewers will cause variation in the answers obtained. This
variation will not be obvious. Hence, every effort must be made to remove
as much of variation as possible due to different interviewers.
Secondary Data: As stated earlier, secondary data are those data which have
already been collected and analyzed by some earlier agency for its own use, and
later the same data are used by a different agency. According to
W.A.Neiswanger, “A primary source is a publication in which the data are
published by the same authority which gathered and analyzed them. A secondary
source is a publication, reporting the data which was gathered by other
authorities and for which others are responsible.”
Sources of secondary data:-The various sources of secondary data can be
divided into two broad categories:
1. Published sources, and
2. Unpublished sources.
Precautions in the Use of Secondary Data: Since secondary data have already
been obtained, it is highly desirable that a proper scrutiny of such data is made
before they are used by the investigator. In fact the user has to be extra-cautious
while using secondary data. In this context Prof. Bowley rightly points out that
“Secondary data should not be accepted at their face value.” The reason being
that data may be erroneous in many respects due to bias, inadequate size of the
sample, substitution, errors of definition, arithmetical errors etc. Even if there is
no error such data may not be suitable and adequate for the purpose of the
enquiry. Prof. Simon Kuznet’s view in this regard is also of great importance.
According to him, “The degree of reliability of secondary source is to be
assessed from the source, the compiler and his capacity to produce correct
statistics and the users also, for the most part, tend to accept a series particularly
one issued by a government agency at its face value without enquiring its
reliability”.
Therefore, before using the secondary data the investigators should
consider the following factors:
4. The suitability of data: The investigator must satisfy himself that the data
available are suitable for the purpose of enquiry. It can be judged by the
nature and scope of the present enquiry with the original enquiry. For
example, if the object of the present enquiry is to study the trend in retail
prices, and if the data provide only wholesale prices, such data are
unsuitable.
(a) Adequacy of data: If the data are suitable for the purpose of investigation
then we must consider whether the data are useful or adequate for the
present analysis. It can be studied by the geographical area covered by the
original enquiry. The time for which data are available is very important
element. In the above example, if our object is to study the retail price
trend of India, and if the available data cover only the retail price trend in
the State of Bihar, then it would not serve the purpose.
(b) Reliability of data: The reliability of data is must. Without which there is
no meaning in research. The reliability of data can be tested by finding out
the agency that collected such data. If the agency has used proper methods
in collection of data, statistics may be relied upon.
It is not enough to have baskets of data in hand. In fact, data in a raw form are
nothing but a handful of raw material waiting for proper processing so that they
can become useful. Once data have been obtained from primary or secondary
source, the next step in a statistical investigation is to edit the data i.e. to
scrutinize the same. The chief objective of editing is to detect possible errors and
irregularities. The task of editing is a highly specialized one and requires great
care and attention. Negligence in this respect may render useless the findings of
an otherwise valuable study. Editing data collected from internal records and
published sources is relatively simple but the data collected from a survey need
excessive editing.
While editing primary data, the following considerations should be borne in
mind:
1. The data should be complete in every respect
2. The data should be accurate
3. The data should be consistent, and
4. The data should be homogeneous.
Choice between Primary and Secondary Data: As we have already seen, there
are a lot of differences in the methods of collecting Primary and Secondary data.
Primary data which is to be collected originally involves an entire scheme of
plan starting with the definitions of various terms used, units to be employed,
type of enquiry to be conducted, extent of accuracy aimed at etc. For the
collection of secondary data, a mere compilation of the existing data would be
sufficient. A proper choice between the type of data needed for any particular
statistical investigation is to be made after taking into consideration the nature,
objective and scope of the enquiry; the time and the finances at the disposal of
the agency; the degree of precision aimed at and the status of the agency
(whether government- state or central-or private institution of an individual).
In using the secondary data, it is best to obtain the data from the primary source
as far as possible. By doing so, we would at least save ourselves from the errors
of transcription which might have inadvertently crept in the secondary source.
Moreover, the primary source will also provide us with detailed discussion about
the terminology used, statistical units employed, size of the sample and the
technique of sampling (if sampling method was used), methods of data
collection and analysis of results and we can ascertain ourselves if these would
suit our purpose.
Now-a-days in a large number of statistical enquiries, secondary data are
generally used because fairly reliable published data on a large number of
diverse fields are now available in the publications of governments, private
organizations and research institutions, agencies, periodicals and magazines etc.
In fact, primary data are collected only if there do not exist any secondary data
suited to the investigation under study. In some of the investigations both
primary as well as secondary data may be used.
SUMMARY:
There are two types of data, primary and secondary. Data which are collected
first hand are called Primary data and data which have already been collected
and used by somebody are called Secondary data. There are two methods of
collecting data: (a) Survey method or total enumeration method and (b) Sample
method. When a researcher goes for investigating all the units of the subject, it is
called as survey method. On the other hand if he/she resorts to investigating only
a few units of the subject and gives the result on the basis of that, it is known as
sample survey method. There are different sources of collecting Primary and
Secondary data. Some of the important sources of Primary data are—Direct
Personal Interviews, Indirect Oral Interviews, Information from correspondents,
Mailed questionnaire method, Schedules sent through enumerators and so on.
Though all these sources or methods of Primary data have their relative merits
and demerits, a researcher should use a particular method with lot of care. There
are basically two sources of collecting secondary data- (a) Published sources and
(b) Unpublished sources. Published sources are like publications of different
government and semi-government departments, research institutions and
agencies etc. whereas unpublished sources are like records maintained by
different government departments and unpublished theses of different
universities etc. Editing of secondary data is necessary for different purposes as
– editing for completeness, editing for consistency, editing for accuracy and
editing for homogeneity.
Introduction:
Nowadays questionnaire is widely used for data collection in social research. It
is a reasonably fair tool for gathering data from large, diverse, varied and
scattered social groups. The questionnaire is the media of communication
between the investigator and the respondents. According to Bogardus, a
questionnaire is a list of questions sent to a number of persons for their answers
and which obtains standardized results that can be tabulated and treated
statistically. The Dictionary of Statistical Terms defines it as a “group of or
sequence of questions designed to elicit information upon a subject or sequence
of subjects from information.” A questionnaire should be designed or drafted
with utmost care and caution so that all the relevant and essential information
for the enquiry may be collected without any difficulty, ambiguity and
vagueness. Drafting of a good questionnaire is a highly specialized job and
requires great care skill, wisdom, efficiency and experience. No hard and fast
rule can be laid down for designing or framing a questionnaire. However, in this
connection, the following general points may be borne in mind:
7. Types of Questions: Under this head, the questions in the questionnaire may
be classified as follows:
(a) Shut Questions: Shut questions are those where possible answers are
suggested by the framers of the questionnaire and the respondent is required to
tick one of them. Shut questions can further be subdivided into the following
forms:
(i) Simple Alternate Questions: In this type of questions the respondent has to
choose from the two clear cut alternatives like ‘Yes’ or ‘No’, ‘Right or Wrong’
etc. Such questions are also called as dichotomous questions. This technique can
be applied with elegance to situations where two clear cut alternatives exist.
(ii) Multiple Choice Questions: Many a times it becomes difficult to define a
clear cut alternative and accordingly in such a situation additional answers
between Yes and No, like Do not know, No opinion, Occasionally, Casually,
Seldom etc. are added. For example, in order to find if a person smokes or
drinks, the following multiple choice answers may be used:
Do you smoke?
(a)Yes regularly [ ] (b) No never [ ]
(c) Occasionally [] (d) Seldom []
Multiple choice questions are very easy and convenient for the respondents to
answer. Such questions save time and also facilitate tabulation. This method
should be used if only a selected few alternative answers exist to a particular
question.
SAMPLING
Though sampling is not new, the sampling theory has been developed
recently. People knew or not but they have been using the sampling technique in
their day to day life. For example a house wife tests a small quantity of rice to
see whether it has been well-cooked and gives the generalized result about the
whole rice boiling in the vessel. The result arrived at is most of the times 100%
correct. In another example, when a doctor wants to examine the blood for any
deficiency, takes only a few drops of blood of the patient and examines. The
result arrived at is most of the times correct and represent the whole amount of
blood available in the body of the patient. In all these cases, by inspecting a few,
they simply believe that the samples give a correct idea about the population.
Most of our decision are based on the examination of a few items only i.e.
sample studies. In the words of Croxton and Cowdon, “It may be too expensive
or too time consuming to attempt either a complete or a nearly complete
coverage in a statistical study. Further to arrive at valid conclusions, it may not
be necessary to enumerate all or nearly all of a population. We may study a
sample drawn from the large population and if that sample is adequately
representative of the population, we should be able to arrive at valid
conclusions.”
According to Rosander, “The sample has many advantages over a census
or complete enumeration. If carefully designed, the sample is not only
considerably cheaper but may give results which are just accurate and
sometimes more accurate than those of a census. Hence a carefully designed
sample may actually be better than a poorly planned and executed census.”
Merits:
1. It saves time: Sampling method of data collection saves time because
fewer items are collected and processed. When the results are urgently required,
this method is very helpful.
2. It reduces cost: Since only a few and selected items are studied in
sampling, there is reduction in cost of money and reduction in terms of man
hours.
3. More reliable results can be obtained: Through sampling, more
reliable results can be obtained because (a) there are fewer chances of sampling
statistical errors. If there is sampling error, it is possible to estimate and control
the results.(b) Highly experienced and trained persons can be employed for
scientific processing and analyzing of relatively limited data and they can use
their high technical knowledge and get more accurate and reliable results.
4. It provides more detailed information: As it saves time, money and
labor, more detail information can be collected in a sample survey.
5. Sometimes only Sampling method to depend upon: Some times it so
happens that one has to depend upon sampling method alone because if the
population under study is finite, sampling method is the only method to be used.
For example, if someone’s blood has to be examined, it will become fatal to take
all the blood out from the body and study depending upon the total enumeration
method.
6. Administrative convenience: The organization and administration of
sample survey are easy for the reasons which have been discussed earlier.
7. More scientific: Since the methods used to collect data are based on
scientific theory and results obtained can be tested, sampling is a more scientific
method of collecting data.
It is not that sampling is free from demerits or shortcomings. There are certain
shortcomings of this method which are discussed below:
1. Illusory conclusion: If a sample enquiry is not carefully planned and
executed, the conclusions may be inaccurate and misleading.
2. Sample not representative: To make the sample representative is a
difficult task. If a representative sample is taken from the universe, the result is
applicable to the whole population. If the sample is not representative of the
universe the result may be false and misleading.
3. Lack of experts: As there are lack of experts to plan and conduct a
sample survey, its execution and analysis, and its results would be
unsatisfactory and not trustworthy.
4. Sometimes more difficult than census method: Sometimes the
sampling plan may be complicated and requires more money, labor and time
than a census method.
5. Personal bias: There may be personal biases and prejudices with regard
to the choice of technique and drawing of sampling units.
6. Choice of sample size: If the size of the sample is not appropriate then it
may lead to untrue characteristics of the population.
7. Conditions of complete coverage: If the information is required for
each and every item of the universe, then a complete enumeration survey is
better.
Essentials of sampling: In order to reach a clear conclusion, the sampling
should possess the following essentials:
1. It must be representative: The sample selected should possess the
similar characteristics of the original universe from which it has been drawn.
2. Homogeneity: Selected samples from the universe should have similar
nature and should mot have any difference when compared with the universe.
3. Adequate samples: In order to have a more reliable and representative
result, a good number of items are to be included in the sample.
4. Optimization: All efforts should be made to get maximum results both
in terms of cost as well as efficiency. If the size of the sample is larger, there is
better efficiency and at the same time the cost is more. A proper size of sample
is maintained in order to have optimized results in terms of cost and efficiency.
Testing of Hypothesis:
As a part of investigation, samples are drawn from the population and
results are derived to help in taking the decisions. But such decisions involve an
element of uncertainty causing wrong decisions. Hypothesis is an assumption
which may or may not be true about a population parameter. For example, if we toss a
coin 200 times, we may get 110 heads and 90 tails. At this instance, we are interested in
testing whether the coin is unbiased or not.
Therefore, we may conduct a test to judge the significance of the difference of
sampling or otherwise. To carry out a test of significance, the following
procedure has to be followed:
1. Framing the Hypothesis: To verify the assumption, which is based on
sample study, we collect data and find out the difference between the sample
value and the population value. If there is no difference found or the difference
is very small then the hypothetical value is correct. Generally two hypotheses
are constructed, and if one is found correct, the other is rejected.
(a) Null Hypothesis: The random selection of the samples from the given
population makes the tests of significance valid for us. For applying any test of
significance we first set up a hypothesis- a definite statement about the
population parameter/s. Such a statistical hypothesis, which is under test, is
usually a hypothesis of no difference and hence is called Null hypothesis. It is
usually denoted by Ho. In the words of Prof. R.A.Fisher “Null hypothesis is the
hypothesis which is tested for possible rejection under the assumption that
it is true.”
(b) Alternative Hypothesis. Any hypothesis which is complementary to the
null hypothesis is called an alternative hypothesis. It is usually denoted by H 1. It
is very important to explicitly state the alternative hypothesis in respect of any
null hypothesis H0 because the acceptance or rejection of Ho is meaningful only
if it is being tested against a rival hypothesis. For example, if we want to test the
null hypothesis that the population has a specified mean µ0(say), i.e., H0:µ=µ
then the alternative hypothesis could be:
(i) H1:µ≠µ0 (i.e., µ>µ0 or µ<µ0)
(ii) H1: µ>µ0 (iii) H1: µ<µ0
The alternative hypothesis (i) is known as a two-tailed alternative and the
alternatives in (ii) and (iii) are known as right-tailed and left-tailed alternatives.
Accordingly, the corresponding tests of significance are called two-tailed, right-
tailed and left-tailed tests respectively.
The null hypothesis consists of only a single parameter value and is
usually simple while alternative hypothesis is usually composite.
Reject Ho Accept Ho
True State
Ho True Wrong Correct
Type I Error
Ho False Correct Wrong
(H1True) Type II Error.
Thus, in testing of hypothesis we are likely to commit two types of
errors. The error of rejecting Ho when Ho is true is known as Type I error and
the error of accepting Ho when Ho is false is known as Type II Error.
For example, in the Industrial Quality Control, while inspecting the quality of a
manufactured lot, the Inspector commits Type I Error when he rejects a good lot
and he commits Type II Error when he accepts a bad lot.
SUMMARY
Nowadays questionnaire method of data collection has become very popular. It
is a very powerful tool to collect required data in shortest period of time and
with little expense. It is scientific too. But drafting of questionnaire is a very
skilled and careful work. Therefore, there are certain requirements and essentials
which should be followed at the time of framing the questionnaire. They include
the following viz., (i) the size of the questionnaire should be small, (ii) questions
should be very clear in understanding, (iii) questions should be put in a logical
order, (iv) questions should have simple meaning etc. Apart from this, multiple
choice questions should be asked. Questionnaire should be pre tested before
going for final data collection. Information supplied should be cross checked for
any false or insufficient information. After all these formalities have been
completed, a covering note should accompany the questionnaire explaining
various purposes, designs, units and incentives.
There are two ways of survey- Census survey and Sample survey through
which data can be collected. Census survey means total enumeration i.e.,
collecting data from each and every unit of the universe, whereas sample survey
concentrates on collecting data from a few units of the universe selected
scientifically for the purpose. Since census method is more time taking,
expensive and labor intensive, it becomes impractical to depend on it. Therefore,
sample survey is preferred which is scientific, less expensive, less time taking
and less labor intensive too.
But there are merits and demerits of this method which are detailed below:
Merits - It reduces cost, saves time and is more reliable. It provides
more detailed information and is sometimes the only method to depend upon for
administrative convenience and scientifically.
Demerits - Sometimes samples may not be representative and may give
illusory conclusions. There are lack of experts and sometimes it is more difficult
than the census method, since there might arise personal bias and the
determination of the size of the sample might be very difficult.
Apart from these, there are some essentials of sampling which must be
followed. They are: Samples must be representative, samples must be
homogeneous and the number of samples must be adequate. When a researcher
resorts to sampling, he intends to collect some data which would help him to
draw results and finally take a decision. When he takes a decision it’s on the
basis of hypothesis which is precisely assumption and is prone to two types of
errors-Type I Error and Type II Error. When a researcher rejects a correct
hypothesis, he commits type I error and when he accepts a wrong hypothesis he
commits type II error. The researcher should try to avoid both types of errors but
committing type II error is more harmful than type I error.
Research Design
Introduction
The meaning of experiment lies in the process of examining the truth of
a statistical hypothesis related to some research problem. For example, a
researcher can conduct an experiment to examine the newly developed
medicine. Experiment is of two types: absolute experiment and comparative
experiment. When a researcher wants to determine the impact of a fertilizer on
the yield of a crop it is a case of absolute experiment. On the other hand, if he
wants to determine the impact of one fertilizer as compared to the impact of
some other fertilizer, the experiment will then be called as a comparative
experiment. Normally, a researcher conducts a comparative experiment when he
talks of designs of experiments.
The main difficulty of such a design is that with the passage of time
considerable extraneous variations may be there in its treatment effect.
Control area:
The basic assumption in this type of design is that the two areas are identical
with respect to their behavior towards the phenomenon considered. If this
assumption is not true, there is the possibility of extraneous variation entering
into the treatment effect.
This design is superior to the previous two designs because it avoids extraneous
variation resulting both from the passage of time and from non-comparability of
the rest and control areas. But at times, due to lack of historical data time or a
comparable control area, we should prefer to select one of the first two informal
designs stated above.
FERTILITY LEVEL
I II III IV V
X1 A B C D E
X2 B C D E A
X3 C D E A B
X4 D E A B C
X5 E A B C D
From the above diagram, it is clear that in L.S. design the field is divided into as
many blocks as there are varieties of fertilizers. Then, each block is again
divided into as many parts as there are varieties of fertilizers in such a way that
each of the fertilizer variety is used in each of the block only once. The analysis
of L.S. design is very similar to the two-way ANOVA technique.
3. Factorial design:
Factorial designs are used in experiments where the effects of varying
more than one factor are to be determined. These designs are used more in
economic and social matters where usually a large number of factors affect a
particular problem. Factorial designs are usually of two types:
Experimental Variable
Treatment A Treatment B
Control Control Control Control
Variable 2 Variable 2 Variable 2 Variable 2
Level I Level II Level I Level II
Level I Cell 1 Cell 3 Cell 5 Cell 7
Control Cell 2 Cell 4 Cell 6 Cell 8
Variable 2
Level II
Experimental Variable
Treatment Treatment
A B
Control Variable II
Level II
Level I
Control Variable I
Level I
Level II
The dotted line cell in this diagram corresponds to cell I of the above stated
2x2x2 design and is for treatment A, level I of the control variable 1, and level I
of the control variable 2. From this design, it is possible to determine the main
effects for three variables i.e., one experimental and true control variables. The
researcher can also determine the interaction between each possible pair of
variables (such interactions are called ‘first order interactions’) and interaction
between variable taken in triplets (such interactions are called second order
interactions). In case of a 2x2x2 design, the further given first order interactions
are possible:
Experimental variable with control variable 1 (or EV x CV 1);
Experimental variable with control variable 2 (or EV x CV 2);
Control variable 1 with control variable 2 (or CV 1 x CV 2);
There will be one second order interaction as well in the given design (it is
between all the three variables i.e., EV x CV 1 x CV 2).
Similarly, the researcher can determine other first order interactions. The
analysis of the first order interaction in the manner described above is essentially
a simple factorial analysis as only two variables are considered at a time and the
remaining ones are ignored. But the analysis of the second order interaction
would not ignore one of the three independent variables in case of a 2x2x2
design. The analysis would be termed as a complex factorial analysis.
It may, however, be remembered that the complex factorial design need not
necessarily be of 2x2x2 type design, but can be generalized to any number and
combinations of experimental and control independent variables. Of course, the
greater the number of independent variables included in a complex factorial
design, the higher the order of the interaction analysis possible. But the overall
task goes on becoming more and more complicated with the inclusion of more
and more independent variables in our design.
Conclusion
There are several research designs and the researcher must decide in advance of
collection and analysis of data as to which design would prove to be more
appropriate for his research project. He must give due weight to various points
such as type of universe and it’s nature, the objective of the study, the source list
or the sampling frame, desired standard accuracy and the like when taking a
decision in respect of the design for his research project.
Observation
Introduction
Observation is a method that employs vision as its main means of data
collection. It implies the use of eyes rather than of ears and the voice. It is
accurate watching and noting of phenomena as they occur with regard to the
cause and effect or mutual relations. It is watching other persons’ behavior as it
actually happens without controlling it. For example, watching bonded
labourer’s life, or treatment of widows and their drudgery at home, provide
graphic description of their social life and sufferings. Observation is also defined
as “a planned methodical watching that involves constraints to improve
accuracy”.
CHARACTERISTICS OF OBSERVATION
Scientific observation differs from other methods of data collection
specifically in four ways: (i) observation is always direct while other methods
could be direct or indirect; (ii) field observation takes place in a natural setting;
(iii) observation tends to be less structured; and (iv) it makes only the qualitative
(and not the quantitative) study which aims at discovering subjects’ experiences
and how subjects make sense of them (phenomenology) or how subjects
understand their life (interpretivism).
Lofland (1955:101-113) has said that this method is more appropriate for
studying lifestyles or sub-cultures, practices, episodes, encounters, relationships,
groups, organizations, settlements and roles etc. Black and Champion
(1976:330) have given the following characteristics of observation:
Behavior is observed in natural surroundings.
It enables understanding significant events affecting social relations of the
participants.
It determines reality from the perspective of observed person himself.
It identifies regularities and recurrences in social life by comparing data in
our study with that of those in other studies.
***
Normal Distribution and its properties
68.27%
95.45%
99.73%
µ - 36 µ - 26 µ - 16 µ=0 µ + 16 µ + 26 µ + 36
- 3 - 2 -1 Z=0 +1 +2 +3
The following table gives the area under the normal probability curve for
some important value of Z.
Distance from the mean ordinate in Area under the curve
Terms of ± σ
Z = ± 0.6745 0.50
Z = ± 1.0 0.6826
Z = ± 1.96 0.95
Z = ± 2.00 0.9544
Z = ± 2.58 0.99
Z = ± 3.0 0.9973
Illustration: Find the probability that the standard normal value lies between 0
and 1.5
0.4332 (43.32%)
Z=0 Z = 1.5
As the mean, Z = 0.
To find the area between Z = 0 and Z = 1.5, look the area between 0 and 1.5,
from the table. It is 0.4332 (shaded area)
Illustration: The results of a particular examination are given below in a
summary form:
Result Percentage of candidates
Passed with distinction 10
Passed 60
Failed 30
It is known that a candidate gets plucked if he obtains less than 40
marks, out of 100 while he must obtain at least 75 marks in order to pass with
distinction. Determine the mean and standard deviation of the distribution of
marks assuming this to be normal.
Solution:
30% students get marks less than 40.
40 – X
Z =-----------= -0.52 (from the table)
σ
30%20%40%10%
40 – X = -0.52σ-------------------------------------(i)
10% students get more than 75
40% area = 75 – X = 1.28 ------------ (ii)
= 75 – X = 1.28σ
Subtract (ii) from (i)
40 – X = -0.52 σ
75 – X = 1.28 σ
-35 = -1.8 σ
35 = 1.8 σ
1.80 σ = 35
35
σ =--------= 19.4
1.80
X = 1000; σ = 200
X–X
Z=
σ
0.1586
3. TESTING OF HYPOTHESIS
Solution:
Let us take the hypothesis that there is no significant difference between the
sample mean and the hypothetical population mean.
σ 20 20
S.E. X = ---- = -------- = -------- = 2
√n √100 10
Since, the difference is more than 2.58 S.E.(1% level) it could not have arisen
due to fluctuations of sampling. Hence the mean weight of students in the
population could not be 125 lbs.
3.2 Test of Significance for Small Samples
If the sample size is less than 30, then those samples may be regarded as
small samples. As a rule, the methods and the theory of large samples are not
applicable to the small samples. The small samples are used in testing a given
hypothesis, to find out the observed values, which could have arisen by sampling
fluctuations from some values given in advance. In a small sample, the
investigator’s estimate will vary widely from sample to sample. An inference
drawn from a smaller sample result is less precise than the inference drawn from
a large sample result.
t-distribution will be employed, when the sample size is 30 or less and
the population standard deviation is unknown.
The formula is
( X - µ)
t =--------x √n
σ
Solution:
Let us take the hypothesis that µ = 510 gms.
( X - µ)
t = ------- x √n
σ
X = 500; µ = 510; σ = 10; n = 20.
500 – 510
t= x √20
10
The computed value is less than the table value. Hence, our null hypothesis is
accepted.
4. CHI-SQUARE TEST
F, t and Z tests were based on the assumption that the samples were drawn from
normally distributed populations. The testing procedure requires assumption
about the type of population or parameters, and these tests are known as
‘parametric tests’.
There are many situations in which it is not possible to make any rigid
assumption about the distribution of the population from which samples are
being drawn. This limitation has led to the development of a group of
alternative techniques known as non-parametric tests. Chi-square test of
independence and goodness of fit is a prominent example of the use of non-
parametric tests.
Though non-parametric theory developed as early as the middle of the
nineteenth century, it was only after 1945 that non-parametric tests came to be
used widely in sociological and psychological research. The main reasons for
the increasing use of non-parametric tests in business research are:-
(i) These statistical tests are distribution-free
(ii) They are usually computationally easier to handle and understand than
parametric tests; and
(iii) They can be used with type of measurements that prohibit the use of
parametric tests.
The χ2 test is one of the simplest and most widely used non-parametric
tests in statistical work. It is defined as:
∑(O – E)2
χ2 = ------------
E
Where O = the observed frequencies, and E = the expected frequencies.
Steps: The steps required to determine the value of χ2are:
(i) Calculate the expected frequencies. In general the expected frequency
for any cell can be calculated from the following equation:
RXC
E =
N
Where E = Expected frequency, R = row’s total of the respective cell, C =
column’s total of the respective cell and N = the total number of observations.
(ii) Take the difference between observed and expected frequencies and
obtain the squares of these differences. Symbolically, it can be represented as
(O – E)2
The computed value of χ2 is compared with the table value of χ 2 for given
degrees of freedom at a certain specified level of significance. If at the stated
level, the calculated value of χ2 is less than the table value, the difference
between theory and observation is not considered as significant.
The following observation may be made with regard to the χ2
distribution:-
(i) The sum of the observed and expected frequencies is always zero.
Symbolically, ∑(O – E) = ∑O - ∑E = N – N = 0
(ii) The χ2 test depends only on the set of observed and expected frequencies
and on degrees of freedom v. It is a non-parametric test.
(iii) χ2 distribution is a limiting approximation of the multinomial
distribution.
(iv) Even though χ2 distribution is essentially a continuous distribution it can
be applied to discrete random variables whose frequencies can be counted and
tabulated with or without grouping.
The Chi-Square Distribution
For large sample sizes, the sampling distribution of χ 2 can be closely
approximated by a continuous curve known as the Chi-square distribution. The
probability function of χ2 distribution is:
F(χ2) = C (χ2)(v/2 – 1)e – x2/2
Where e = 2.71828, v = number of degrees of freedom, C = a constant
depending only on v.
The χ2 distribution has only one parameter, v, the number of degrees of
freedom. As in case of t-distribution there is a distribution for each different
number of degrees of freedom. For very small number of degrees of freedom,
the Chi-square distribution is severely skewed to the right. As the number of
degrees of freedom increases, the curve rapidly becomes more symmetrical. For
large values of v the Chi-square distribution is closely approximated by the
normal curve.
The following diagram gives χ2 distribution for 1, 5 and 10 degrees of
freedom:
F(x2)
v=1
v=5
v = 10
0 2 4 6 8 10 12 14 16 18 20 22 χ2
χ2 Distribution
It is clear from the given diagram that as the degrees of freedom increase,
the curve becomes more and more symmetric. The Chi-square distribution is a
probability distribution and the total area under the curve in each chi-square
distribution is unity.
Properties of χ2 distribution
The main Properties of χ2 distribution are:-
(i) the mean of the χ2 distribution is equal to the number of degrees of freedom, i.e.,
X=v
(ii) the variance of the χ2 distribution is twice the degrees of freedom, Variance =
2v
(iii) µ1 = 0,
(iv) µ2 = 2v,
(v) µ3 = 8v,
(vi) µ4 = 48v + 12v2.
µ32 64v2 8
(vii) β1 = --- =-------= --
µ22 8v3 v
µ4 48v + 12v2 12
(v) β1µ3 = ---- = --------------- = 3 + ---
µ22 4v2 v
N (ab – bc - ½N)2
χ2 =
(a + c) (b + d) (c + d) (a + b)
(ii) No theoretical cell frequency should be small when the expected frequencies are
too small. If it is so, then the value of χ 2 will be overestimated and will result in
too many rejections of the null hypothesis. To avoid making incorrect
inferences, a general rule is followed that expected frequency of less than 5 in
one cell of a contingency table is too small to use. When the table contains
more than one cell with an expected frequency of less than 5 then add with the
preceding or succeeding frequency so that the resulting sum is 5 or more.
However, in doing so, we reduce the number of categories of data and will gain
less information from contingency table.
(iii) The constraints on the cell frequencies if any should be linear, i.e., they should
not involve square and higher powers of the frequencies such as ∑O = ∑E = N.
Uses of χ2 test:
(i) χ2 test as a test of independence. With the help of χ2 test, we can find
out whether two or more attributes are associated or not. Let’s assume that we
have N observations classified according to some attributes. We may ask
whether the attributes are related or independent. Thus, we can find out whether
there is any association between skin colour of husband and wife. To examine
the attributes that are associated, we formulate the null hypothesis that there is
no association against an alternative hypothesis and that there is an association
between the attributes under study. If the calculated value of χ 2 is less than the
table value at a certain level of significance, we say that the result of the
experiment provides no evidence for doubting the hypothesis. On the other
hand, if the calculated value of χ2 is greater than the table value at a certain level
of significance, the results of the experiment do not support the hypothesis.
(ii) χ2 test as a test of goodness of fit. This is due to the fact that it enables
us to ascertain how appropriately the theoretical distributions such as binomial,
Poisson, Normal, etc., fit empirical distributions. When an ideal frequency
curve whether normal or some other type is fitted to the data, we are interested
in finding out how well this curve fits with the observed facts. A test of the
concordance of the two can be made just by inspection, but such a test is
obviously inadequate. Precision can be secured by applying the χ2 test.
Solution: Let us take the hypothesis that quinine is not effective in checking
diabetes. Applying χ2 test :
(A) X (B) 240 x 812
Expectation of (AB) = ------------ =------------------= 60
N 3248
Or E1, i.e., expected frequency corresponding to first row and first column is 60.
the bale of expected frequencies shall be:
60 752 812
180 2256 2436
O E (O – E)2 (O – E)2/E
20 60 1600 26.667
220 180 1600 8.889
792 752 1600 2.218
2216 2256 1600 0.709
[∑(O – E)2/E] = 38.593
The calculated value of χ2 is greater than the table value. The hypothesis is
rejected. Hence medicine x is useful in checking malaria.
Since the calculated value of χ2 is greater than the table value the hypothesis is
not true. We, therefore, conclude the vaccine is effective in controlling
susceptibility to tuberculosis.
ANALYSIS OF VARIANCE (ANOVA)
Introduction
For managerial decision making, sometimes one has to carry out tests of
significance. The analysis of variance is an effective tool for this purpose. The
objective of the analysis of variance is to test the homogeneity of the means of
different samples.
Definition
According to R.A. Fisher, “Analysis of variance is the separation of variance
ascribable to one group of causes from the variance ascribable to other groups”.
Assumptions of ANOVA
The technique of ANOVA is mainly used for the analysis and interpretation of
data obtained from experiments. This technique is based on three important
assumptions, namely
1. The parent population is normal.
2. The error component is distributed normally with zero mean and
constant variance.
3. The various effects are additive in nature.
The technique of ANOVA essentially consists of partitioning the total variation
in an experiment into components of different sources of variation. These
sources of variations are due to controlled factors and uncontrolled factors.
Since the variation in the sample data is characterized by means of many
components of variation, it can be symbolically represented in the mathematical
form called a linear model for the sample data.
Classification of models
Linear models for the sample data may broadly be classified into three types as
follows:
1. Random effect model
2. Fixed effect model
3. Mixed effect model
In any variance components model, the error component has always
random effects, since it occurs purely in a random manner. All other
components may be either mixed or random.
Random effect model
A model in which each of the factors has random effect (including error effect)
is called a random effect model or simply a random model.
Fixed effect model
A model in which each of the factors has fixed effects, buy only the error effect
is random is called a fixed effect model or simply a fixed model.
Mixed effect model
A model in which some of the factors have fixed effects and some others have
random effects is called a mixed effect model or simply a mixed model.
In what follows, we shall restrict ourselves to a fixed effect model.
In a fixed effect model, the main objective is to estimate the effects and
find the measure of variability among each of the factors and finally to find the
variability among the error effects.
The ANOVA technique is mainly based on the linear model which
depends on the types of data used in the linear model. There are several types of
data in ANOVA, depending on the number of sources of variation namely,
One-way classified data,
Two-way classified data,
…
m-way classified data.
One-way classified data
When the set of observations is distributed over different levels of a single
factor, then it gives one-way classified data.
ANOVA for One-way classified data
Let
y denote the jth observation corresponding to the ith level of factor A and
ij
where
represents the general mean effect which is fixed and which represents
the general condition of the experimental units, ai denotes the fixed effect due
to ith level of
the factor A (i=1,2,…,k) and hence the variation due to ai
(i=1,2,…,k) is said to be control.
The last component of the model eij is the random variable. It is called the error
component and it makes the Yij a random variate. The variation in ei is due to
j
all the uncontrolled factors and eij is independently, identically and normally
y a e ⎛ i 1, 2,..., k ⎞
ij ⎜ ⎟
ij i
j 1, 2,..., n
⎝ i⎠
given by
E( yij ) for all i 1, 2,..., k
i
with
yij i eij , where eij is the random error effect due to uncontrolled factors
On substitution for i
in the above equation, the linear model reduces to
G N niai (2)
i
n a i i
i
unique solution for and ai (i = 1,2,…,k). Using this condition in equation (2),
we get
GN
G
i.e.
N
Ti
ai
ni
Ti
Hence, ai
ni
Therefore, the estimate of ai is given by
T
aµ i µ
i
ni
T G
i.e., aµ i (5) (5)
i
ni N
squares, we get
E ( yij µ $a2i )
ij
After carrying out some calculations and using the normal equations (2) and (3)
we obtain
⎛ G2 ⎞ ⎛ T 2 G2 ⎞
E ⎜ y 2 N ⎟ ⎜ in N ⎟ (6)
i
⎝ ij ⎠ ⎝ i i ⎠
The first term in the RHS of equation (6) is called the corrected total sum of
squares while y2i is called the uncorrected total sum of squares.
ij j
Proceeding as before, we get the residual sum of squares for this hypothetical
model as
⎛ 2⎞ G2
E1 ⎜ yij ⎟ (7)
N
⎝ ij ⎠
Actually,
E1 contains the variation due to both treatment and error. Therefore a
The expression in (8) is usually called the corrected treatment sum of squares
k
T2
while the term i is called uncorrected treatment sum of squares. Here it
i 1 ni
G 2
may be noted that is a correction factor (Also called a correction term).
N
Since E is based on N-k free observations, it has N - k degrees of freedom (df).
Similarly, since
E1 is based on N -1 free observation, E1 has N -1 degrees of
freedom. So
E1 has k -1 degrees of freedom.
E
When actually the null hypothesis is true, if we reject it on the basis of
the estimated value in our statistical analysis, we will be committing Type – I
er ror. The probability for committi ng this error is referred to as the level of
significance, denoted by . The t esting of the null hy pothesis Ho may be
d
F
i.e., It follows F distribution with degrees of freedom k-1 and N-k.
All these values are represented in the form of a table called ANOVA table,
furnished below.
QE :
Within the level N-k QE -
M
By subtraction E
of factor (Error) Nk
G2
Total N-1 Q yij - -
ij N
Variance ratio
The variance ratio is the ratio of the greater variance to the smaller variance. It is
also called the F-coefficient. We have
F = Greater variance / Smaller variance.
We refer to the table of F values at a desired level of significance . In general,
is taken to be 5 %. The table value is referred to as the theoretical value or the
expected value. The calculated value is referred to as the observed value.
Inference
If the observed value of F is less than the expected value of F (i.e., Fo < Fe) for
the given level of significance , then the null hypothesis H o is accepted. In
this case, we conclud
e that there is no sign ificant d ifference between the
treatment effects.
On the other hand, if t
he observed value of F isthan
greater
the expected value
of F (i.e., he given level of significance ,
Fo Fe ) for t then the null hypothesis
case, we conclude tha t all the tr
Ho is rejected. In this
eatment effects are not
equal.
Note: If the calculated value of F and the table value of F are equal, we can try
some other value of .
Problem 1
The following are the details of sales effected by three sales persons in three
door-to-door campaigns.
Sales person Sales in door – to – door campaign
A 8 9 5 10
B 7 6 6 9
C 6 6 7 5
Construct an ANOVA table and find out whether there is any significant
difference in the performance of the sales persons.
Solution:
Method I (Direct method) :
A 8 9 5 10 32
B 7 6 6 9 28
C 6 6 7 5 24
32
Sample mean for A : A 8
28
Sample mean for B : B 7
24
Sample mean for C : C 6
Total num ber of sam ple items = No. of items for A + No . of items for B + No.
of items for C
= 4
32
ll the samples X
Sum of squares of deviations for A:
A AAA8 A
A 2
8 0 0
9 1 1
5 -3 9
10 2 4
14
B BBB7 B B 2
7 0 0
6 -1 1
6 -1 1
9 2 4
6
C CCC6 C C
2
6 0 0
6 0 0
7 1 1
5 -1 1
2
Sum of squares of deviations within
varieties = A A 2 B B 2 C C 2
= 14 + 6 + 2
= 22
ANOVA Table
Source of variation Degrees of freedom Sum of squares of Variance
deviations
8
Between varieties 3–1= 2 8 4
2
22
Within varieties 12 – 3 = 9 22 2.44
9
Total 12 – 1 = 11 30
Calculation of F value:
Greater Variance 4.00
F = Smaller Variance 2.44 1.6393
Inference:
The calculated valu e of F is less than the table value of F. Therefore, the null
hypothesis is accepted. It is concluded that there is no significant difference in
the performance of the sales persons, at 5% level of significance.
Method II (Short cut method):
A = 32, B = 28, C = 24.
T= Sum of all the sample items
A B C
32 28 24
84
N = Total number of items in all the samples = 4 + 4 + 4 =12
2
2 84
Correction Factor = T 588
N 12
Calculate the sum of squares of the observed values as follows:
Sales Person X X2
A 8 64
9 81
A 5 25
A 10 100
A 7 49
B 6 36
B 6 36
B 9 81
B 6 36
C 6 36
C 7 49
C 5 25
C
618
It is to be noted that the ANOVA tables in the methods I and II are one and the
same. For the further steps of calculation of F value and drawing inference,
refer to method I.
Problem 2
The following are the details of plinth areas of ownership apartment flats offered
by 3 housing companies A,B,C. Use analysis of variance to determine whether
there is any significant difference in the plinth areas of the apartment flats.
Housing Company Plinth area of apartment flats
A 1500 1430 1550 1450
B 1450 1550 1600 1480
C 1550 1420 1450 1430
A=10+3+15+5=33
B 5+15+10+8=38
C=15+2+5+3=25
T A B C
33 38 25
96
N = Total number of items in all the samples = 4 + 4 + 4 = 12
T2 962
Correction Factor = 768
N 12
Calculate the sum of squares of the observed values as follows:
Company X X2
A 10 100
A 3 9
A 15 225
A 5 25
B 5 25
B 15 225
B 10 100
B 8 64
C 15 225
C 2 4
C 5 25
C 3 9
1036
Inference:
Since the calculated value of F is less than the table value of F, the null
hypothesis is accepted and it is concluded that there is no significant difference
in the plinth areas of ownership apartment flats offered by the three companies,
at 5% level of significance.
Problem 3
A finance manager has collected the following information on the performance
of three financial schemes.
Source of variation Degrees of Freedom Sum of squares of deviations
Treatments 5 15
Residual 2 25
Total (corrected) 7 40
df2/df1 1 2 3 4 5 6 7 8 9 10
1 647.7 799.5 864.1 899.5 921.8 937.1 948.2 956.6 963.2 968.6
2 38.5 39.0 39.1 39.2 39.2 39.3 39.3 39.3 39.3 39.3
3 17.4 16.0 15.4 15.1 14.8 14.7 14.6 14.5 14.4 14.4
4 12.2 10.6 9.9 9.6 9.3 9.1 9.0 8.9 8.9 8.8
5 10.0 8.4 7.7 7.3 7.1 6.9 6.8 6.7 6.6 6.6
6 8.8 7.2 6.5 6.2 5.9 5.8 5.6 5.5 5.5 5.4
7 8.0 6.5 5.8 5.5 5.2 5.1 4.9 4.8 4.8 4.7
8 7.5 6.0 5.4 5.0 4.8 4.6 4.5 4.4 4.3 4.2
9 7.2 5.7 5.0 4.7 4.4 4.3 4.1 4.1 4.0 3.9
10 6.9 5.4 4.8 4.4 4.2 4.0 3.9 3.8 3.7 3.7
11 6.7 5.2 4.6 4.2 4.0 3.8 3.7 3.6 3.5 3.5
12 6.5 5.0 4.4 4.1 3.8 3.7 3.6 3.5 3.4 3.3
13 6.4 4.9 4.3 3.9 3.7 3.6 3.4 3.3 3.3 3.2
14 6.2 4.8 4.2 3.8 3.6 3.5 3.3 3.2 3.2 3.1
15 6.1 4.7 4.1 3.8 3.5 3.4 3.2 3.1 3.1 3.0
16 6.1 4.6 4.0 3.7 3.5 3.3 3.2 3.1 3.0 2.9
17 6.0 4.6 4.0 3.6 3.4 3.2 3.1 3.0 2.9 2.9
18 5.9 4.5 3.9 3.6 3.3 3.2 3.0 3.0 2.9 2.8
19 5.9 4.5 3.9 3.5 3.3 3.1 3.0 2.9 2.8 2.8
20 5.8 4.4 3.8 3.5 3.2 3.1 3.0 2.9 2.8 2.7
21 5.8 4.4 3.8 3.4 3.2 3.0 2.9 2.8 2.7 2.7
df2/df1 1 2 3 4 5 6 7 8 9 10
1 39.8 49.5 53.5 55.8 57.2 58.2 58.9 59.4 59.8 60.1
2 8.5 9.0 9.1 9.2 9.2 9.3 9.3 9.3 9.3 9.3
3 5.5 5.4 5.3 5.3 5.3 5.2 5.2 5.2 5.2 5.2
4 4.5 4.3 4.1 4.1 4.0 4.0 3.9 3.9 3.9 3.9
5 4.0 3.7 3.6 3.5 3.4 3.4 3.3 3.3 3.3 3.2
6 3.7 3.4 3.2 3.1 3.1 3.0 3.0 2.9 2.9 2.9
7 3.5 3.2 3.0 2.9 2.8 2.8 2.7 2.7 2.7 2.7
8 3.4 3.1 2.9 2.8 2.7 2.6 2.6 2.5 2.5 2.5
9 3.3 3.0 2.8 2.6 2.6 2.5 2.5 2.4 2.4 2.4
10 3.2 2.9 2.7 2.6 2.5 2.4 2.4 2.3 2.3 2.3
11 3.2 2.8 2.6 2.5 2.4 2.3 2.3 2.3 2.2 2.2
12 3.1 2.8 2.6 2.4 2.3 2.3 2.2 2.2 2.2 2.1
13 3.1 2.7 2.5 2.4 2.3 2.2 2.2 2.1 2.1 2.1
14 3.1 2.7 2.5 2.3 2.3 2.2 2.1 2.1 2.1 2.0
15 3.0 2.6 2.4 2.3 2.2 2.2 2.1 2.1 2.0 2.0
16 3.0 2.6 2.4 2.3 2.2 2.1 2.1 2.0 2.0 2.0
17 3.0 2.6 2.4 2.3 2.2 2.1 2.1 2.0 2.0 2.0
18 3.0 2.6 2.4 2.2 2.1 2.1 2.0 2.0 2.0 1.9
19 2.9 2.6 2.3 2.2 2.1 2.1 2.0 2.0 1.9 1.9
20 2.9 2.5 2.3 2.2 2.1 2.0 2.0 1.9 1.9 1.9
21 2.9 2.5 2.3 2.2 2.1 2.0 2.0 1.9 1.9 1.9
22 2.9 2.5 2.3 2.2 2.1 2.0 2.0 1.9 1.9 1.9
23 2.9 2.5 2.3 2.2 2.1 2.0 1.9 1.9 1.9 1.8
24 2.9 2.5 2.3 2.1 2.1 2.0 1.9 1.9 1.9 1.8
25 2.9 2.5 2.3 2.1 2.0 2.0 1.9 1.9 1.8 1.8
26 2.9 2.5 2.3 2.1 2.0 2.0 1.9 1.9 1.8 1.8
27 2.9 2.5 2.2 2.1 2.0 2.0 1.9 1.9 1.8 1.8
28 2.8 2.5 2.2 2.1 2.0 1.9 1.9 1.9 1.8 1.8
29 2.8 2.4 2.2 2.1 2.0 1.9 1.9 1.8 1.8 1.8
30 2.8 2.4 2.2 2.1 2.0 1.9 1.9 1.8 1.8 1.8
WHAT IS A REPORT?
A report is a written document on a particular topic, which conveys
information and ideas and may also make recommendations. Reports often form
the basis of crucial decision making. Inaccurate, incomplete and poorly written
reports fail to achieve their purpose and reflect on the decision, which will
ultimately be made. This will also be the case if the report is excessively long,
jargonistic and/ or structureless. A good report can be written by keeping the
following features in mind:
1. All points in the report should be clear to the intended reader.
2. The report should be concise with information kept to a necessary
minimum and arranged logically under various headings and sub-headings.
3. All information should be correct and supported by evidence.
4. All relevant material should be included in a complete report.
Academic Reports:
A report written for an academic course can be thought of as a
simulation. We can imagine that someone wants the report for a practical
purpose, although we are really writing the report as an academic exercise for
assessment. Theoretical ideas will be more to the front in an academic report
than in a practical one. Sometimes a report seems to serve academic and
practical purposes. Students on placement with organizations often have to
produce a report for the organization and for assessment on the course.
Although the background work for both will be related, in practice, the report the
student produces for academic assessment will be different from the report
produced for the organization, because the needs of each are different.
Headings and sub-headings used in the report should be given with their
page numbers. Each chapter should begin on a new page. Use a consistent
system in dividing the report into parts. The simplest may be to use chapters for
each major part and subdivide these into sections and sub-sections. 1, 2, 3 etc.
can be used as the numbers for each chapter. The sections of chapter 3 (for
example) would be 3.1, 3.2, 3.3, and so on. For further sub-division of a sub-
section you may use 3.2.1, 3.2.2, and so on.
Abstract or Summary or Executive Summary or Introduction:
This presents an overview of the whole report. It should let the reader see
in advance, what is in the report. This includes what you set out to do, how
review of literature is focused and narrowed in your research, the relation of the
methodology you chose to your objectives, a summary of your findings and
analysis of the findings
BODY
Aims and Purpose or Aims and Objectives:
Why did you do this work? What was the problem you were
investigating? If you are not including review of literature, mention the specific
research/es which is/are relevant to your work.
Review of Literature
This should help to put your research into a background context and to
explain its importance. Include only the books and articles which relate directly
to your topic. You need to be analytical and critical, and not just describe the
works that you have read.
Methodology
Methodology deals with the methods and principles used in an activity,
in this case research. In the methodology chapter, explain the method/s you used
for the research and why you thought they were the appropriate ones. You may,
for example, be depending mostly upon secondary data or you may have
collected your own data. You should explain the method of data collection,
materials used, subjects interviewed, or places you visited. Give a detailed
account of how and when you carried out your research and explain why you
used the particular method/s, rather than other methods. Included in this chapter
should be an examination of ethical issues, if any.
Results or Findings
What did you find out? Give a clear presentation of your results. Show
the essential data and calculations here. You may use tables, graphs and figures.
Analysis and Discussion
Interpret your results. What do you make out of them? How do they
compare with those of others who have done research in this area? The accuracy
of your measurements/results should be discussed and deficiencies, if any, in the
research design should be mentioned.
Conclusions
What do you conclude? Summarize briefly the main conclusions which
you discussed under "Results." Were you able to answer some or all of the
questions which you raised in your aims and objectives? Do not be tempted to
draw conclusions which are not backed up by your evidence. Note the
deviation/s from expected results and any failure to achieve all that you had
hoped.
Recommendations
Make your recommendations, if required. The suggestions for action and
further research should be given.
Appendix
You may not need an appendix, or you may need several. If you have
used questionnaires, it is usual to include a blank copy in the appendix. You
could include data or calculations, not given in the body, that are necessary, or
useful, to get the full benefit from your report. There may be maps, drawings,
photographs or plans that you want to include. If you have used special
equipment, you may include information about it.
The plural of an appendix is appendices. If an appendix or appendices
are needed, design them thoughtfully in a way that your readers find it/them
convenient to use.
References
List all the sources which you referred in the body of the report. You
may use the pattern prescribed by American Psychological Association, or any
other standard pattern recognized internationally.
REVIEW OF LITERATURE
In the case of small projects, this may not be in the form of a critical review
of the literature, but this is often asked for and is a standard part of larger
projects. Sometimes students are asked to write Review of Literature on a topic
as a piece of work in its own right. In its simplest form, the review of literature
is a list of relevant books and other sources, each followed by a description and
comment on its relevance.
The literature review should demonstrate that you have read and analysed
the literature relevant to your topic. From your readings, you may get ideas
about methods of data collection and analysis. If the review is part of a project,
you will be required to relate your readings to the issues in the project, and while
describing the readings, you should apply them to your topic. A review should
include only relevant studies. The review should provide the reader with a
picture of the state of knowledge in the subject.
Your literature search should establish what previous researches have been
carried out in the subject area. Broadly speaking, there are three kinds of sources
that you should consult:
1. Introductory material;
2. Journal articles and
3. Books.
To get an idea about the background of your topic, you may consult one or
more textbooks at the appropriate time. It is a good practice to review in
cumulative stages - that is, do not think you can do it all at one go. Keep a
careful record of what you have searched, how you have gone about it, and the
exact citations and page numbers of your readings. Write notes as you go along.
Record suitable notes on everything you read and note the methods of
investigations. Make sure that you keep a full reference, complete with page
numbers. You will have to find your own balance between taking notes that are
too long and detailed, and ones too brief to be of any use. It is best to write your
notes in complete sentences and paragraphs, because research has shown that
you are more likely to understand your notes later if they are written in a way
that other people would understand. Keep your notes from different sources
and/or about different points on separate index cards or on separate sheets of
paper. You will do mainly basic reading while you are trying to decide on your
topic. You may scan and make notes on the abstracts or summaries of work in
the area. Then do a more thorough job of reading later on, when you are more
confident of what you are doing. If your project spans several months, it would
be advisable towards the end to check whether there are any new and recent
references.
REFERENCES
There are many methods of referencing your work; some of the most
common ones are the Numbered Style, American Psychological Association
Style and the Harvard Method, with many other variations. Just use the one
you are most familiar and comfortable with. Details of all the works referred
by you should be given in the reference section.
THE PRESENTATION OF REPORT