B128 Expt9 Sem 2
B128 Expt9 Sem 2
B128 Expt9 Sem 2
Experiment No. 9
1) List out various applications of different statistical techniques are given below:
UNIFORM DISTRIBUTION :
The concepts of discrete uniform distribution and continuous uniform distribution, as well as
the Random variables they describe, are the foundations of statistical analysis and probability
Theory.
The binomial distribution is a probability distribution that summarizes the likelihood that a
value .Will take one of two independent values under a given set of parameters or
assumptions. The Underlying assumptions of the binomial distribution are that there is only
one outcome for each Trial, that each trial has the same probability of success, and that each
trial is mutually exclusive, Or independent of each other.
BERNOULLI DISTRIBUTION
A discrete probability distribution describes the probability that each possible value of a
discrete Random variable will occur—for example, the probability of getting a six when
rolling a die. When Dealing with discrete variables, the probability of each value falls
between 0 and 1, and the sum of all The probabilities is equal to 1. So, in the die example,
assuming we’re using a standard die, the Probability of rolling a six is 0.167, or 16.7%. This
is based on dividing 1 (the sum of all probabilities) by 6 (the number of possible outcomes)
POISSON DISTRIBUTION :
A Poisson distribution is a probability distribution that can be used to show how many times
an event is Likely to occur within a specified period of time. In other words, it is a count
distribution. Poisson Distributions are often used to understand independent events that occur
at a constant rate within a Given interval of time. It was named after French mathematician
Siméon Denis Poisson.
The Poisson distribution is a discrete function, meaning that the variable can only take
specific values in a (potentially infinite) list. Put differently, the variable cannot take all
values in any continuous range.For the Poisson distribution (a discrete distribution), the
variable can only take the values 0, 1, 2, 3, etc., With no fractions or decimals.
Points to remember :
➢ A Poisson distribution can be used to measure how many times an event is likely to occur
within
“X” period of time, named after mathematician Siméon Denis Poisson.
➢ Poisson distributions, therefore, are used when the factor of interest is a discrete count
variable.
➢ Many economic and financial data appear as count variables, such as how many times a
person
becomes unemployed in a given year, thus lending itself to analysis with a Poisson
distribution.
• A Poisson distribution can be used to estimate how likely it is that something will happen
“X”
number of times. For example, if the average number of people who rent movies on a Friday
night at a single video store location is 400, a Poisson distribution can answer such questions
as,
“What is the probability that more than 600 people will rent movies?” Therefore, the
application
of the Poisson distribution enables managers to introduce optimal scheduling systems that
would not work with, say, a normal distribution.
GEOMETRIC DISTRIBUTION :
The geometric distribution, intuitively speaking, is the probability distribution of the number
of tails one
Must flip before the first head using a weighted coin. It is useful for modeling situations in
which it is Necessary to know how many attempts are likely necessary for success, and thus
has applications to Population modeling, econometrics, return on investment (ROI) of
research, and so on.
The geometric distribution is useful for determining the likelihood of a success given a
limited number of
Trials, which is highly applicable to the real world in which unlimited (and unrestricted)
trials are rare.
➢ In cost-benefit analyses, such as a company deciding whether to fund research trials that,
if Successful, will earn the company some estimated profit, the goal is to reach a success
before The cost outweighs the potential gain.
➢ In time management, the goal is to complete a task before some set amount of time
✓ Measures of Dispersion: Measures of dispersion provide information about the spread of
a variable's values. There are four key measures of dispersion:
• Range
• Variance
• Standard Deviation
• Skewness
• Kurtosis
Measures of Dispersion and Applications are as following:
a) Range: Range is simply the difference between the smallest and largest values in the
data. The interquartile range is the difference between the values at the 75 th percentile and
the 25th percentile of the data. It is defined as the difference between smallest and the
largest value of the complete data set.
c) Standard Deviation: Standard deviation, another commonly used statistic, is the square
root of the variance. It is the measurement of average distance between each quantity and
means which is how the set of data spreads out from the mean. A high standard deviation
means that the data points are spread at wider ranges of values, whereas a low standard
deviation means that the data points are close to the mean.
d) Skewness: It is a measure of the asymmetry in the distribution of the set of data. For
example, the income is skewed as some people might be earning between a standard
range while others will be way higher or lower to that range. The skewness can be
positive, negative or undefined.
- Positive Skew: When a distribution is skewed to the right, that is, the tail of the
curve on the right-hand side is longer than that to the left, and the mean value > the
mode, this is called a Positive Skew.
- Negative Skew: When a distribution is skewed to the left, that is, the tail of the
curve on the left-hand side is longer than that to the right, and the mean value < the
mode, this is called a Negative Skew and for a perfect normal distribution, the tails on
each side of the curve are exactly the same.
e) Kurtosis: The sample kurtosis is a useful measure of weather there is a problem with the
outliers in a data set. Larger kurtosis indicates a more serious outlier problem, and may
lead the researcher to choose alternative statistical methods.
Computer Science:
Computer science, the study of computer and computing, including their theoretical and
algorithmic foundations, software, and their uses for processing information. The discipline
of computer science includes the study of algorithms and data structures, computer and
network design, modeling data and information processes, and artificial intelligence.
Computer science draws some of its foundations mathematics and engineering therefore
incorporates techniques from areas such as queueing theory, probability and statistics and
electronic circuit design. Computer science also makes heavy use of hypothesis testing and
experimentation during the conceptualization, design, measurement, and refinement of new
algorithms, information structures, and computer.
Statistics plays an intrinsic role in computer science and vice versa. Statistics is used for data
mining, speech recognition, vision and image analysis, data compression, artificial
intelligence, and network and traffic modeling. A statistical background is essential for
understanding algorithms and statistical properties that from the backbone of computer
science. Computer science tend to focus on data acquisition/cleaning, retrieval, mining, and
reporting. They are often tasked with the development of algorithms for prediction and
system efficiency. Focus is also placed one machine learning (an aspect of artificial
intelligence), particularly for the purpose of data (finding patterns and associations in data)
for a variety of purposes, such as marketing and finance.
For all above tasks, the statistical techniques like Linear regression models, Discriminant
analysis, Resampling methods, Principal components analysis, Cluster analysis etc. are
useful.
Computer science and information technology (IT) are two distinct subjects, despite their
many similarities. Generally, computer science refers to designing and building computers
and computer programs. Information technology, on the other hand, refers to maintaining
and troubleshooting those computers and their networks, systems, and databases to ensure
they run smoothly
The line between IT and computer science is often blurry. Sometimes working in IT will
require knowledge of computer science principles, and being a programmer will mean
carrying out tasks that are considered more IT-related. And sometimes jobs that are
computer-science heavy, like web development or software engineering, are referred to as IT
work. All of this can lead to very understandable confusion
Common in both:
• Health and safety when using a computers.
• Environmental Impacts/Sustainability.
• Computer Legislation(Copyright, Computer Misuse, Data Protection).
• Ethical, Moral and Social implications of computers.