Вы находитесь на странице: 1из 17

QT – (Al Jamia Arts and Science College, Poopalam)

QUANTITATIVE TECHNIQUES different variables connected with a number


Quantitative techniques may be defined as of simultaneous equations.
those techniques which provide the decision 5. Differentiation: It is a mathematical
makes a systematic and powerful means of process of finding out changes in the
analysis, based on quantitative data. It is a dependent variable with reference to a small
scientific method employed for problem change in the independent variable.
solving and decision making by the 6. Integration: Integration is the reverse
management. With the help of quantitative process of differentiation.
techniques, the decision maker is able to 7.Differential Equation: It is a
explore policies for attaining the mathematical equation which involves the
predetermined objectives. In short, differential coefficients of the dependent
quantitative techniques are inevitable in variables.
decision-making process. Statistical Quantitative Techniques:
Classification of Quantitative Techniques: Statistical techniques are those techniques
There are different types of quantitative which are used in conducting the statistical
techniques. We can classify them into three enquiry concerning to certain Phenomenon.
categories. They are: They include all the statistical methods
1. Mathematical Quantitative Techniques beginning from the collection of data till
2. Statistical Quantitative Techniques interpretation of those collected data.
3. Programming Quantitative Techniques Statistical techniques involve:
Mathematical Quantitative Techniques: 1. Collection of data: One of the important
A technique in which quantitative data are statistical methods is collection of data.
used along with the principles of There are different methods for collecting
mathematics is known as mathematical primary and secondary data.
quantitative techniques. Mathematical 2. Measures of Central tendency,
quantitative techniques involve: dispersion, skewness and Kurtosis:
1. Permutations and Combinations: Measures of Central tendency is a method
Permutation means arrangement of objects used for finding he average of a series while
in a definite order. The number of measures of dispersion used for finding out
arrangements depends upon the total the variability in a series. Measures of
number of objects and the number of objects Skewness measures asymmetry of a
taken at a time for arrangement. distribution while measures of Kurtosis
Combination means selection or grouping measures the flatness of peakedness in a
objects without considering their order. distribution.
2. Set Theory:- Set theory is a modern 3. Correlation and Regression Analysis:
mathematical device which solves various Correlation is used to study the degree of
types of critical problems. relationship among two or more variables.
3. Matrix Algebra: Matrix is an orderly On the other hand, regression technique is
arrangement of certain given numbers or used to estimate the value of one variable for
symbols in rows and columns. It is a a given value of another.
mathematical device of finding out the 4. Index Numbers: Index numbers measure
results of different types of algebraic the fluctuations in various Phenomena like
operations on the basis of the relevant price, production etc over a period of time,
matrices. They are described as economic barometers.
4. Determinants: It is a powerful device 5. Time series Analysis: Analysis of time
developed over the matrix algebra. This series helps us to know the effect of factors
device is used for finding out values of which are responsible for changes:
QT – (Al Jamia Arts and Science College, Poopalam)

6. Interpolation and Extrapolation: 5. Inventory Theory: Inventory theory


Interpolation is the statistical technique of helps for optimizing the inventory levels. It
estimating under certain assumptions, the focuses on minimizing cost associated with
missing figures which may fall within the holding of inventories.
range of given figures. Extrapolation 6. Network programming: It is a technique
provides estimated figures outside the range of planning, scheduling, controlling,
of given data. monitoring and co-coordinating large and
7. Statistical Quality Control Statistical complex projects comprising of a number of
quality control is used for ensuring the activities and events. It serves as an
quality of items manufactured. The instrument in resource allocation and
variations in quality because of assignable adjustment of time and cost up to the
causes and chance causes can be known with optimum level. It includes CPM, PERT etc.
the help of this tool. Different control charts 7. Simulation: It is a technique of testing a
are used in controlling the quality of model which resembles a real life situations
products. 8. Replacement Theory: It is concerned
8. Ratio Analysis: Ratio analysis is used for with the problems of replacement of
analyzing financial statements of any machines, etc due to their deteriorating
business or industrial concerns which help efficiency or breakdown. It helps to
to take appropriate decisions. determine the most economic replacement
9. Probability Theory: Theory of policy.
probability provides numerical values of the 9. Non Linear Programming: It is a
likely hood of the occurrence of events. programming technique which involves
10. Testing of Hypothesis Testing of finding an optimum solution to a problem in
hypothesis is an important statistical tool to which some or all variables are non-linear.
judge the reliability of inferences drawn on 10. Sequencing: Sequencing tool is used to
the basis of sample studies. determine a sequence in which given jobs
Programming Techniques: should be performed by minimizing the total
Programming techniques are also called efforts.
operations research techniques. 11. Quadratic Programming: Quadratic
Programming techniques are model building programming technique is designed to solve
techniques used by decision makers in certain problems, the objective function of
modern times. Programming techniques which takes the form of a quadratic equation.
involve: 12. Branch and Bound Technique : This is
1. Linear Programming: Linear designed to solve the combinational
programming technique is used in finding a problems of decision making where there
solution for optimizing a given objective are large number of feasible solutions.
under certain constraints. Problems of plant location, problems of
2. Queuing Theory: Queuing theory deals determining minimum cost of production
with mathematical study of queues. It aims etc. are examples of combinational problems.
at minimizing cost of both servicing and Functions of Quantitative Techniques:
waiting. The following are the important functions of
3. Game Theory: Game theory is used to quantitative techniques:
determine the optimum strategy in a 1. To facilitate the decision-making process
competitive situation. 2. To provide tools for scientific research
4. Decision Theory: This is concerned with 3. To help in choosing an optimal strategy
making sound decisions under conditions of 4. To enable in proper deployment of
certainty, risk and uncertainty. resources
QT – (Al Jamia Arts and Science College, Poopalam)

5. To help in minimizing costs Classification of Correlation


6. To help in minimizing the total processing Correlation can be classified in different
time required for performing a set of jobs ways. The following are the most important
7. helping in capital resource allocation classifications
Limitations of Quantitative Techniques: 1. Positive and Negative correlation
1. Quantitative techniques involves 2. Simple, partial and multiple correlation
mathematical models, equations and other 3. Linear and Non-linear correlation
mathematical expressions Positive and Negative correlation
2. Quantitative techniques are based on Positive Correlation: When the variables
number of assumptions. are varying in the same direction, it is called
3. Quantitative techniques are very positive correlation. In other words, if an
expensive. increase in the value of one variable is
4. Quantitative techniques do not take into accompanied by an increase in the value of
consideration intangible facts like skill, other variable or if a decrease in the value of
attitude etc. one variable is accompanied by a decree se
5. Quantitative techniques are only tools for in the value of other variable, it is called
analysis and decision-making. They are not positive correlation.
decisions itself. Negative Correlation: When the variables
CORRELEATION ANALYSIS are moving in opposite direction, it is called
Introduction: negative correlation. In other words, if an
In practice, we may come across with lot of increase in the value of one variable is
situations which need statistical analysis of accompanied by a decrease in the value of
either one or more variables. The data other variable or if a decrease in the value of
concerned with one variable only is called one variable is accompanied by an increase
univariate data. The analysis of such data is in the value of other variable, it is called
called univariate analysis. The data negative correlation.
concerned with two variables are called Simple, Partial and Multiple correlation
bivariate data. The analysis of these two sets Simple Correlation In a correlation
of data is called bivariate analysis. The date analysis, if only two variables are studied it
concerned with three or more variables are is called simple correlation.
called multivariate date. Multiple correlation In a correlation
The statistical technique which can be used analysis, if three or more variables are
to study the relationship between two or studied simultaneously, it is called multiple
more variables is called correlation analysis. correlation.
Definition: Two or more variables are said Partial correlation In a correlation analysis,
to be correlated if the change in one variable we recognize more than two variable, but
results in a corresponding change in the consider one dependent variable and one
other variable. According to Simpson and independent variable and keeping the other
Kafka, “Correlation analysis deals with the Independent variables as constant.
association between two or more variables”. Linear and Non-linear correlation
Lun chou defines, “ Correlation analysis Linear Correlation: In a correlation
attempts to determine the degree of analysis, if the ratio of change between the
relationship between variables”. two sets of variables is same, then it is called
Boddington states that “Whenever some linear correlation.
definite connection exists between two or Non-linear correlation: In a correlation
more groups or classes of series of data, analysis if the amount of change in one
there is said to be correlation.” variable does not bring the same ratio of
QT – (Al Jamia Arts and Science College, Poopalam)

change in the other variable, it is called non- method all the values of the two variable are
linear correlation. plotted in a chart in the form of dots.
Degrees of correlation: Therefore, it is also known as dot chart. By
Correlation exists in various degrees observing the scatter of the various dots, we
1. Perfect positive correlation: If an can form an idea that whether the variables
increase in the value of one variable is are related or not. A scatter diagram
followed by the same proportion of increase indicates the direction of correlation and
in other related variable or if a decrease in tells us how closely the two variables under
the value of one variable is followed by the study are related. The greater the scatter of
same proportion of decrease in other related the dots, the lower is the relationship
variable, it is perfect positive correlation. Merits of Scatter Diagram method
2. Perfect Negative correlation: If an 1.it is easy to plot the points
increase in the value of one variable is 2. simple to understand
followed by the same proportion of decrease 3. abnormal values in the data can be easily
in other related variable or if a decrease in detected
the value of one variable is followed by the 4. extreme value do not affect it.
same proportion of increase in other related Demerits of Scatter diagram method
variably it is Perfect Negative Correlation. 1.the degree of correlation cannot be easily
3. Limited Degree of Positive correlation:: estimated
When an increase in the value of one 2. algebraic treatment is not possible
variable is followed by a non-proportional Correlation graph Method
increase in other related variable, or when a Under correlation graph method the
decrease in the value of one variable is individual values of the two variables are
followed by a non proportional decrease in plotted on a graph paper. Then dots relating
other related variable, it is called limited to these variables are joined separately so as
degree of positive correlation. to get two curves. By examining the direction
4. Limited degree of Negative correlation: and closeness of the two curves, we can infer
When an increase in the value of one whether the variables are related or not. If
variable is followed by a non-proportional both the curves are moving in the same
decrease in other related variable, or when a direction( either upward or downward)
decrease in the value of one variable is correlation is said to be positive. If the
followed by a non proportional increase in curves are moving in the opposite directions,
other related variable, it is called limited correlation is said to be negative.
degree of negative correlation. Correlation Coefficient:
5. Zero Correlation/ no correlation: If Correlation analysis is actually an attempt to
there is no correlation between variables it is find a numerical value to express the extent
called zero correlation. In other words, if the of relationship exists between two or more
values of one variable cannot be associated variables. The numerical measurement
with the values of the other variable, it is showing the degree of correlation between
zero correlation. two or more variables is called correlation
Methods of measuring correlation coefficient. Correlation coefficient ranges
1) Scatter Diagram between -1 and +1. Coefficient of correlation
2) Correlation graph can be computed by applying the methods
3) coefficient of correlation given below
Scatter Diagram 1) Karl Pearson’s Co-efficient of correlation
This is the simplest method for ascertaining 2) Spearman’s Rank correlation method
the correlation between variables. Under this 3) Concurrent deviation method
QT – (Al Jamia Arts and Science College, Poopalam)

Karl Pearson’s Co-efficient of Correlation Probable Error


Karl Pearson’s Coefficient of Correlation is Probable error (PE) of the Co-efficient of
the most popular method among the correlation is a statistical device which
algebraic methods for measuring correlation. measures the reliability and dependability of
This method was developed by Prof. Karl the value of co-efficient of correlation.
Pearson. If the value of coefficient of correlation ( r) is
Pearson’s coefficient of correlation is defined less than the PE, then there is no evidence of
as the ratio of the covariance between X and correlation. If the value of ‘r’ is more than 6
Y to the product of their standard deviations. times of PE, the correlation is certain and
This is denoted by ‘r’ significant. By adding and submitting PE
interpretation of Co-efficient of from coefficient of correlation, we can find
Correlation out the upper and lower limits within which
Pearson’s Co-efficient of correlation always the population coefficient of correlation may
lies between +1 and -1. The following be expected to lie.
general rules will help to interpret the Co- Spearman’s Rank Correlation Method
efficient of correlation: Pearson’s coefficient of correlation method is
1. When r = +1, It means there is perfect applicable when variables are measured in
positive relationship between variables. quantitative form. But there were many
2. When r = -1, it means there is perfect cases where measurement is not possible
negative relationship between variables. because of the qualitative nature of the
3. When r = 0, it means there is no variable. For example, we cannot measure
relationship between the variables. the beauty, morality, intelligence, honesty etc
4. When ‘r’ is closer to +1, it means there is in quantitative terms. However it is possible
high degree of positive correlation between to rank these qualitative characteristics in
variables. some order. The correlation coefficient
5. When ‘r’ is closer to – 1, it means there is obtained from ranks of the variables instead
high degree of negative correlation between of their quantitative measurement is called
variables. rank correlation. This was developed by
6. When ‘r’ is closer to ‘O’, it means there is Charles Edward Spearman in 1904.
less relationship between variables. Spearman’s coefficient correlation (R) = 1-
Properties of Co-efficient of Correlation ∑
1.correlation coefficient has a well-defined
formula Where D = difference of ranks between the
2. correlation coefficient is a pure number two variables
and is independent. N = number of pairs
3. it lies between -1 and + 1. Computation of Rank Correlation
4. correlation coefficient do not change with Coefficient when Ranks are Equal
reference to change. There may be chances of obtaining same
5. coefficient of correlation between X and Y rank for two or more items. In such a
is same as that between Y and X. situation, it is required to give average rank
Computation of Pearson’s Co-efficient of for all. Such items.
correlation: Merits of Rank Correlation method
Pearson’s correlation co-efficient can be 1.it is easy to calculate
computed in different ways. They are: 2. simple to understand
a Arithmetic mean method 3. can be applied both quantitative and
b Assumed mean method qualitative data.
c Direct method
QT – (Al Jamia Arts and Science College, Poopalam)

Demerits of Rank Correlation method meaning of the term ‘regression” is “stepping


1. Rank correlation coefficient is only an back” to the average.
approximate measure as the actual values Definition:
are not used for calculations. According to M.M Blair “Regression is the
2. It is not convenient when number of pairs measure of the average relationship between
(ie. N) is large two or more variables in terms of the
3. Further algebraic treatment is not original units of the date”.
possible. Types of Regression:-
Concurrent Deviation Method: There are two types of regression. They are
Concurrent deviation method is a very linear regression and multiple regression.
simple method of measuring correlation. Linear Regression: It is a type of regression
Under this method, we consider only the which uses one independent variable to
directions of deviations. The magnitudes of explain and/or predict the dependent
the values are completely ignored. variable.
Therefore, this method is useful when we are Multiple Regression: It is a type of
interested in studying correlation between regression which uses two or more
two variables in a casual manner and not independent variable to explain and/or
interested in degree (or precision). Under predict the dependent variable.
this method, the nature of correlation is Regression Lines:
known from the direction of deviation in the Regression line is a graphic technique to
values of variables. If deviations of 2 show the functional relationship between
variables are concurrent, then they move in the two variables X and Y. It is a line which
the same direction, otherwise in the opposite shows the average relationship between two
direction. The formula for computing the variables X and Y. If there is perfect positive
coefficient of concurrent deviation is: - correlation between 2 variables, then the
two regression lines are winding each other
r= √ and to give one line. There would be two
Where N = No. of pairs of symbol regression lines when there is no perfect
C = No. of concurrent deviations (ie, No. of + correlation between two variables. The
signs in ‘dx dy’ column) nearer the two regression lines to each
Co-efficient of determination other, the higher is the degree of correlation
co-efficient determination gives the and the farther the regression lines from
percentage variation in the dependent each other, the lesser is the degree of
variable in relation with the independent correlation.
variable. In other words co-efficient of Properties of Regression lines:-
determination gives the ratio of the 1. The two regression lines cut each other at
explained variable to the total variance. The the point of average of X and average of Y
co-efficient of determination is the square of 2. When r = 1, the two regression lines
the correlation coefficient. coincide each other and give one line.
REGRESSION ANALYSIS 3. When r = 0, the two regression lines are
Introduction:- mutually perpendicular.
Regression analysis means to analyse the Regression Equations (Estimating
average relationship between two variables Equations)
and thereby provides a mechanism for Regression equations are algebraic
estimation or predication or forecasting. The expressions of the regression lines. Since
term ‘Regression” was firstly used by Sir there are two regression lines, therefore two
Francis Galton in 1877. The dictionary regression equations. They are :-
QT – (Al Jamia Arts and Science College, Poopalam)

1.Regression Equation of X on Y:- This is is a number lying between ‘0’ and ‘1’ ‘0’
used to describe the variations in the values denotes the even which cannot occur, and ‘1’
of X for given changes in Y. denotes the event which is certain to occur
here equation is X = a + by Terms use in Probability.
∑x = na +b ∑y Random Experiment: A random
∑xy = a∑y + b ∑y2 experiment is an experiment that has two or
more outcomes which vary in an
2. Regression Equation of Y on X :- This is unpredictable manner from trial to trail
used to describe the variations in the value when conducted under uniform conditions.
of Y for given changes in X. In a random experiment, all the possible
here equation is Y = a + bx outcomes are known in advance but none of
∑y = na +b ∑x the outcomes can be predicted with
∑xy = a∑x + b ∑x2 certainty. For example, tossing of a coin is a
Difference between correlation and random experiment because it has two
regression outcomes (head and tail), but we cannot
Correlation Regression predict any of them which certainty.
It studies degree of It studies the nature Sample Point: Every indecomposable
relationship of relationship outcome of a random experiment is called a
between variables between variables sample point. It is also called simple event or
It is not used for It is basically used elementary outcome. Eg. When a die is
prediction purposes for prediction thrown, getting ‘3’ is a sample point.
purposes Sample space: Sample space of a random
It is basically used It is basically used experiment is the set containing all the
as a tool for as a tool for sample points of that random experiment.
determining the studying cause and Eg:- When a coin is tossed, the sample space
degree of effect relationship is (Head, Tail)
relationship Event: An event is the result of a random
There may be There is no such experiment. It is a subset of the sample space
nonsense nonsense regression of a
correlation between random experiment.
two variables Sure Event (Certain Event): An event
There is no question There must be whose occurrence is inevitable is called sure
of dependent and dependent and even. Eg:- Getting a white ball from a box
independent independent containing all while balls.
variables variables Impossible Events: An event whose
THEORY OF PROBABILITY occurrence is impossible, is called impossible
INTRODUCTION event. Eg:- Getting a white ball from a box
Probability refers to the chance of happening containing all red balls.
or not happening of an event. Any problem Uncertain Events: An event whose
which contains uncertainty about the occurrence is neither sure nor impossible is
happening of the event is the problem of called uncertain event. Eg:- Getting a white
probability. ball from a box containing white balls and
Definition of Probability black balls.
The probability of given event may be Equally likely Events: Two events are said
defined as the numerical value given to the to be equally likely if anyone of them cannot
likely hood of the occurrence of that event. It. be expected to occur in preference to other.
For example, getting herd and getting tail
QT – (Al Jamia Arts and Science College, Poopalam)

when a coin is tossed are equally likely tendency of frequency ratio to stabilize, is
events. common in all random experiments. This
Mutually exclusive events: A set of events phenomenon is known as statistical
are said to be mutually exclusive of the regularity.
occurrence of one of them excludes the Relative Frequency Definition or
possibility of the occurrence of the others. Empirical Approach
Exhaustive Events:: A group of events is According to Relative Frequency definition,
said to be exhaustive when it includes all the probability of an event can be defined as
possible outcomes of the random experiment the relative frequency with which it occurs in
under consideration. an indefinitely large number of trials.
Dependent Events: Two or more events are Here, probability has between 0 and 1, i.e. 0
said to be dependent if the happening of one ≤ P(A) ≤ 1
of them affects the happening of the other. Axiomatic Approach (Modern Approach)
DIFFERENT SCHOOLS OF THOUGHT ON to Probability
PROBABILITY Let ‘S’ be the sample space of a random
Different Approaches/Definitions of experiment, and ‘A’ be an event of the
Probability random experiment, so that ‘A’ is the subset
There are 4 important schools of thought on of ‘S’. Then we can associate a real number to
probability :- the event ‘A’. This number will be called
1. Classical or Priori Approach Objective probability of ‘A’ if it satisfies the following
Probability three axioms or postulates :-
2. Relative frequency or Empirical Approach 1.P (A) is a real number such that P (A) ≥ 0
3. Modern or Axiomatic Approach for every subset of S
1. Classical or Priori Approach 2. P (S) = 1 where ‘S’ is the sample space
If out of ‘n’ exhaustive, mutually exclusive 3. P (A ⋃ B) = P(A) + P (B), where A and B
and equally likely outcomes of an are two non-intersecting subset of S
experiment; ‘m’ are favourable to the THEOREMS OF PROBABILITY
occurrence of an event ‘A’, then the There are two important theorems of
probability of ‘A’ is defined as to be P(A) = probability. They are :
1. Addition Theorem
Frequency ratio : consider a random
2. Multiplication Theorem
experiment. Let ‘A’ be an event associated
Addition Theorem
with the random experiment. Let us repeat
Here, there are 2 situations.
the experiment ‘n’ times. Let the event A
(a) Events are mutually exclusive
happen ‘f ‘ out of ‘n’ repetition of the
(b) Events are not mutually exclusive
experiment. Then f/n is called frequency
(a) Addition theorem (Mutually Exclusive
ratio.
Events): If two events, ‘A’ and ‘B’, are
Law of statistical regularity : the outcomes
mutually exclusive the probability of the
of a random experiment vary in an
occurrence of either ‘A’ or ‘B’ is the sum of
unpredictable manner. But if we consider
the individual probability of A and B.
sequence of frequency ratios, we can observe
P(A or B) = P(A) + P(B)
a regularity. We notice that these frequency
i.e., P(A⋃B) = P(A) + P(B)
ratios have a tendency to become more and
(b)Addition theorem (Not mutually
more stable as the number of repetitions of
exclusive events): If two events, A and B are
the experiment increases. The frequency
not mutually exclusive the probability of the
ratio appears to approach a fixed number as
occurrence of either A or B is the sum of
the number repetitions is increased. This
their individual probability minus
QT – (Al Jamia Arts and Science College, Poopalam)

probability for both to happen. Distribution) can be defined as a distribution


P(A or B) = P(A) + P(B) – P(A and B) obtained for a random variable on the basis
i.e., P(A⋃B) = P(A) + P(B) – P(A∩B) of a mathematical model. It is obtained not
MULTIPLICATION THEOREM on the basis of actual observation or
Here there are two situations: experiments, but on the basis of probability
(a) Events are independent law.
(b) Events are dependent Random variable
(a)Multiplication theorem (independent Random variable is a variable who value is
events): If two events are independent, then determined by the outcome of a random
the probability of occurring both will be the experiment. Random variable is also called
product of the individual probability chance variable or stochastic variable.
P(A and B) = P(A) X P(B) Classification of Probability Distribution
i.e., P(A⋂B) = P(A) X P(B) Discrete probability Distribution
(b)Multiplication theorem (dependent Continuous Probability Distribution
Events):-If two events, A and B are Binomial Distribution
dependent, the probability of occurring 2nd Poisson distribution
event will be affected by the outcome of the Normal Distribution
first. Discrete Probability Distribution: If the
P(A⋂B) = P(A) X P(B/A) random variable of a probability distribution
CONDITIONAL PROBABILITY assumes specific values only, it is called
Multiplication theorem states that if two discrete probability distributions. Binomial
events, A and B, are dependent events then, distribution and poisson distribution are
the probability of happening both will be the discrete probability distributions.
product of P(A) and P(B/A). Continuous Probability Distributions:- If
i.e., P(A and B) or P(A ⋂ B ) = P(A) X P(B/A) the random variable of a probability
Here, P (B/A) is called Conditional distribution assumes any value in a given
probability interval, then it is called continuous
Inverse Probability probability distributions. Normal
If an event has happened as a result of distributions is a continuous probability
several causes, then we may be interested to distribution.
find out the probability of a particular cause BIONOMIAL DISTRIBUTION
of happening that events. This type of Meaning & Definition:
problem is called inverse probability. Baye’s Binomial Distribution is associated with
theorem is based upon inverse probability. James Bernoulli, a Swiss Mathematician.
BAYE’S THEOREM: Therefore, it is also called Bernoulli
Baye’s theorem is based on the proposition distribution. Binomial distribution is the
that probabilities should revised on the basis probability distribution expressing the
of all the available information. The revision probability of one set of dichotomous
of probabilities based on available alternatives, i.e., success or failure. In other
information will help to reduce the risk words, it is used to determine the probability
involved in decision-making. The of success in experiments on which there are
probabilities before revision is called priori only two mutually exclusive outcomes.
probabilities and the probabilities after Binomial distribution is discrete probability
revision is called posterior probabilities. distribution.
PROBABILITY DISTRIBUTION Binomial Distribution can be defined as
DEFINITION follows: “A random variable x is said to
Probability distribution (Theoretical follow Binomial Distribution with
QT – (Al Jamia Arts and Science College, Poopalam)

parameters n and p if its probability function Fitting a Binomial Distribution


is: Steps:
f (x) = nC x px qn-x 1. Find the value of n, p and q
Where, P = probability of success in a single 2. Substitute the values of n, p and q in the
trial Binomial Distribution function of f (x) = nC
q=1–p x px qn-x
n = number of trials 3. Put x = 0, 1, 2, ……….. in the function f (x)
x = number of success in ‘n’ trials. = nC x px qn-x
Assumption of Binomial Distribution OR 4. Multiply each such terms by total
(Situations where Binomial Distribution frequency (N) to obtain the expected
can be applied) frequency.
Binomial distribution can be applied when:- POISSON DISTRIBUTION
1. The random experiment has two outcomes Meaning and Definition:
i.e., success and failure. Poisson Distribution is a limiting form of
2. The probability of success in a single trial Binomial Distribution. In Binomial
remains constant from trial to trial of the Distribution, the total number of trials are
experiment. known previously. But in certain real life
3. The experiment is repeated for finite situations, it may be impossible to count the
number of times. total number of times a particular event
4. The trials are independent. occurs or does not occur. In such cases
Properties (features) of Binomial Poisson Distribution is more suitable. Poison
Distribution: Distribution is a discrete probability
1. It is a discrete probability distribution. distribution. It was originated by Simeon
2. The shape and location of Binomial Denis Poisson. The Poisson Distribution is
distribution changes as ‘p’ changes for a defined as:-
given ‘n’.
3. The mode of the Binomial distribution is p (x) =
equal to the value of ‘r’ which has the largest Where x = random variable (i.e., number of
probability. success in ‘n’ trials.
4. Mean of the Binomial distribution e = 2.7183
increases as ‘n’ increases with ‘p’ remaining m = mean of Poisson distribution
constant. Properties of Poisson Distribution
5. The mean of Binomial distribution is np. 1. Poisson Distribution is a discrete
6. The Standard deviation of Binomial probability distribution.
distribution is √ 2. Poisson Distribution has a single
7. If ‘n’ is large and if neither ‘p’ nor ‘q’ is too parameter ‘m’. When ‘m’ is known all the
close zero, Binomial distribution may be terms can be found out.
approximated to Normal Distribution. 3. It is a positively skewed distribution.
8. If two independent random variables 4. Mean and Variance of Poisson Distribution
follow Binomial distribution, their sum also are equal to ‘m’.
follows Binomial distribution. 5. In Poisson Distribution, the number of
Mean and Standard Deviation of Binomial success is relatively small.
Distribution 6. The standard deviation of Poisson
Mean of Binomial Distribution = np Distribution is √m.
Standard Deviation of Binomial Distribution
=√
QT – (Al Jamia Arts and Science College, Poopalam)

Practical situations where Poisson 4. Normal curve is a bell shaped curve.


Distribution can be used 5. Mean, Median and Mode coincide at the
1. To count the number of telephone calls centre of the curve.
arising at a telephone switch board in a unit 6. Quantities are equi-distant from median.
of time. Q3 – Q2 = Q2 – Q1
2. To count the number of customers arising 7. Normal curve is asymptotic to the base
at the super market in a unit of time. line.
3. To count the number of defects in 8. Total area under a normal curve is 100%.
Statistical Quality Control. 9. The ordinate at the mean divide the whole
4. To count the number of bacteria’s per unit. area under a normal curve into two equal
5. To count the number of defectives in a parts. (50% on either side).
park of manufactured goods. 10. The height of normal curve is at its
6. To count the number of persons dying due maximum at the mean.
to heart attack in a year. 11. The normal curve is unimodel, i.e., it has
7. To count the number of accidents taking only one mode.
place in a day on a busy road. 12. Normal curve is mesokurtic.
NORMAL DISTRIBUTION 13. No portion of normal curve lies below the
The normal distribution is a continuous x-axis.
probability distribution. It was first 14. Theoretically, the range of normal curve
developed by De-Moivre in 1733 as limiting is – α to + α . But practically the range is
form of binomial distribution. Fundamental Fitting of a Normal Distribution
importance of normal distribution is that Procedure :
many populations seem to follow 1. Find the mean and standard deviation of
approximately a pattern of distribution as the given distribution. (i.e., μ and σ)
described by normal distribution. Numerous 2. Take the lower limit of each class.
phenomena such as the age distribution of 3. Find Z value for each of the lower limit. Z =
any species, height of adult persons,
intelligent test scores of students, etc. are
considered to be normally distributed. 4. Find the area for z values from the table.
Definition of Normal Distribution The first and the last values are taken as 0.5.
A continuous random variable ‘X’ is said to5. Find the area for each class. Take
follow Normal Distribution if its probability
difference between 2 adjacent values if same
function is: signs and take total of adjacent values if
opposite signs.
6. Find the expected frequency by
multiplying area for each class by N.
π = 3.146 TESTING OF HYPOTHESIS
e = 2.71828 Statistical Inference:
μ = mean of the distribution Statistical inference refers to the process of
σ = standard deviation of the distribution selecting and using a sample statistic to draw
Properties of Normal Distribution conclusions about the population parameter.
1. Normal distribution is a continuous Statistical inference deals with two types of
distribution. problems.
2. Normal curve is symmetrical about the They are:-
mean. 1. Testing of Hypothesis
3. Both sides of normal curve coincide 2. Estimation
exactly.
QT – (Al Jamia Arts and Science College, Poopalam)

Hypothesis: Hypothesis is a statement Generally, the level of significance is fixed at


subject to verification. More precisely, it is a 1% or 5%.
quantitative statement about a population, 3. Decide a test criterion: The third step in
the validity of which remains to be tested. In testing of hypothesis is to select an
other words, hypothesis is an assumption appropriate test criterion. Commonly used
made about a population parameter. tests are z-test, t-test, X2 – test, F-test, etc.
Testing of Hypothesis: Testing of 4. Calculation of test statistic: The next
hypothesis is a process of examining step is to calculate the value of the test
whether the hypothesis formulated by the statistic using appropriate formula. The
researcher is valid or not. The main objective general form for computing the value of test
of hypothesis testing is whether to accept or statistic is:-
reject the hypothesis. Value of Test statistic = Difference Standard
Procedure for Testing of Hypothesis: Error
The various steps in testing of hypothesis 5. Making Decision: Finally, we may draw
involves the following :- conclusions and take decisions. The decision
1. Set Up a Hypothesis:: The first step in may be either to accept or reject the null
testing of hypothesis is to set up a hypothesis hypothesis. If the calculated value is more
about population parameter. Normally, the than the table value, we reject the null
researcher has to fix two types of hypothesis. hypothesis and accept the alternative
They are null hypothesis and alternative hypothesis. If the calculated value is less
hypothesis. than the table value, we accept the null
Null Hypothesis:-Null hypothesis is the hypothesis.
original hypothesis. It states that there is no Sampling Distribution
significant difference between the sample The distribution of all possible values which
and population regarding a particular matter can be assumed by some statistic, computed
under consideration. The word “null” means from samples of the same size randomly
‘invalid’ of ‘void’ or ‘amounting to nothing’. drawn from the same population is called
Null hypothesis is denoted by Ho. Sampling distribution of that statistic.
Alternative Hypothesis:-Any hypothesis Standard Error (S.E)
other than null hypothesis is called Standard Error is the standard deviation of
alternative hypothesis. When a null the sampling distribution of a statistic.
hypothesis is rejected, we accept the other Standard error plays a very important role in
hypothesis, known as alternative hypothesis. the large sample theory. The following are
Alternative hypothesis is denoted by H1. the important uses of standard errors:-
2. Set up a suitable level of significance: 1. Standard Error is used for testing a given
After setting up the hypothesis, the hypothesis
researcher has to set up a suitable level of 2. S.E. gives an idea about the reliability of a
significance. The level of significance is the sample, because the reciprocal of S.E. is a
probability with which we may reject a null measure of reliability of the sample.
hypothesis when it is true. For example, if 3. S.E. can be used to determine the
level of significance is 5%, it means that in confidence limits within which the
the long run, the researcher is rejecting true population parameters are expected to lie.
null hypothesis 5 times out of every 100 Test Statistic
times. Level of significance is denoted by α The decision to accept or to reject a null
(alpha). hypothesis is made on the basis of a statistic
α = Probability of rejecting H0 when it is computed from the sample. Such a statistic is
true. called the test statistic. There are different
QT – (Al Jamia Arts and Science College, Poopalam)

types of test statistics. All these test statistics test statistic follows t-distribution. Uses of t-
can be classified into two groups. They are test are:-
a. Parametric Tests 1. To test the population mean when the
b. Non-Parametric Tests sample is small and the population s.D.is
PARAMETRIC TESTS unknown.
The statistical tests based on the assumption 2. To test the equality of two sample means
that population or population parameter is when the samples are small and population
normally distributed are called parametric S.D. is unknown.
tests. The important parametric tests are:- 3. To test the difference in values of two
1. z-test dependent samples.
2. t-test 4. To test the significance of correlation
3. f-test coefficients.
Z-test: The following are the important assumptions
Z-test is applied when the test statistic in t-test:-
follows normal distribution. It was 1. The population from which the sample
developed by Prof.R.A.Fisher. The following drawn is normal.
are the important uses of z-test:- 2. The sample observations are independent.
1. To test the population mean when the 3. The population S.D.is known.
sample is large or when the population 4. When the equality of two population
standard deviation is known. means is tested, the samples are assumed to
2. To test the equality of two sample means be independent and the population variance
when the samples are large or when the are assumed to be equal and unknown.
population standard deviation is known. F-test:
3. To test the population proportion. F-test is used to determine whether two
4. To test the equality of two sample independent estimates of population
proportions. variance significantly differ or to establish
5. To test the population standard deviation both have come from the same population.
when the sample is large. For carrying out the test of significance, we
6. To test the equality of two sample calculate a ration, called F-ratio. F-test is
standard deviations when the samples are named in honour of the great statistician
large or when population standard R.A.Fisher. It is also called Variance Ration
deviations are known. Test.
7. To test the equality of correlation Uses of F-distribution:-
coefficients. 1. To test the equality of variances of two
Z-test is used in testing of hypothesis on the populations.
basis of some assumptions. The important 2. To test the equality of means of three or
assumptions in z-test are:- more populations.
1. Sampling distribution of test statistic is 3. To test the linearity of regression
normal. Assumptions of F-distribution:-
2. Sample statistics are dose the population 1. The values in each group are normally
parameter and therefore, for finding distributed.
standard error, sample statistics are used in 2. The variance within each group should be
place where population parameters are to be equal for all groups.
used. 3. The error (Variation of each value around
T-test: its own group mean) should be independent
t-distribution was originated by W.S.Gosset for each value.
in the early 1900. t-test is applied when the
QT – (Al Jamia Arts and Science College, Poopalam)

TYPES OF ERRORS IN TESTING OF area which corresponds to the


HYPOTHESIS: predetermined level of significance. If the
In any test of hypothesis, the decision is to calculated value of the test statistic falls in
accept or reject a null hypothesis. The four the rejection region, we reject the null
possibilities of the decision are:- hypothesis. Rejection region is also called
1. Accepting a null hypothesis when it is true. critical region. It is denoted by α (alpha).
2. Rejecting a null hypothesis when it is false. Acceptance Region: Acceptance region is
3. Rejecting a null hypothesis when it is true. the area which corresponds to 1 – α.
4. Accepting a null hypothesis when it is Acceptance region = 1 – rejection region = 1-
false. α.
Out of the above 4 possibilities, 1 and 2 are If the calculated value of the test statistic
correct, while 3 and 4 are errors. The error falls in the acceptance region, we accept the
included in the above 3rd possibility is called null hypothesis.
type I error and that in the 4th possibility is TWO TAILED AND ONE TAILED TESTS:
called type II error. A two tailed test is one in which we reject the
Type I Error null hypothesis if the computed value of the
The error committed by rejecting a null test statistic is significantly greater or lower
hypothesis when it is true, is called Type I than the critical value (table value) of the
error. The probability of committing Type I test statistic. Thus, in two tailed test the
error is denoted by α (alpha). critical region is represented by both tails of
α = Prob. (Type I error) the normal curve. If we are testing
= Prob. (Rejecting H0 when it is true) hypothesis at 5 % level of significance, the
Type II Error size of the acceptance region is 0.95 and the
The error committed by accepting a null size of the rejection region is 0.05 on both
hypothesis when it is false is called Type II sides together. (i.e. 0.025 on left side and
error. The probability of committing Type II 0.025 on right side of the curve).
error is denoted by β (beta). TESTING OF EQUALITY OF TWO SAMPLE
β = Prob. (Type II error) MEANS
β = Prob. (Accepting H0 when it is false) This test is used to test whether there is
Small and Large samples significant difference between two sample
The size of sample is 30 or less than 30, the means. It there is no significant difference,
sample is called small sample. When the size we can consider the samples are drawn from
of sample exceeds 30, the sample is called the same population.
large sample. Procedure:
Degree of freedom 1. Set up null hypothesis that there is no
Degree of freedom is defined as the number significant difference between the two
of independent observations which is means.
obtained by subtracting the number of H0 : μ1 = μ2
constraints from the total number of H1 : μ1 ≠ μ2
observations. Degree of freedom = Total 2. Decide the test criterion:
number of observations – Number of • If sample is large, apply z – test
constraints. • If sample is small, but population S.D. is
Rejection region and Acceptance region known, apply z-test.
The entire area under a normal curve may be • If sample is small and population S.D. is
divided into two parts. They are rejection unknown, apply t-test.
region and acceptance region. 3. Apply the formula:
Rejection Region: Rejection region is the
QT – (Al Jamia Arts and Science College, Poopalam)

between observed and expected frequencies.


Z or t = χ2-test is a statistical test which tests the
4. Fix the degree of freedom: significance of difference between observed
5. Obtain the table value. frequencies and corresponding theoretical
6. Decide whether to accept or reject the H0. frequencies of a distribution without any
TESTING OF EQUALITY OF TWO SAMPLE assumption about the distribution of the
STANDARD DEVIATIONS population. This is one of the simplest and
This test is used to test whether there is any most widely used nonparametric test in
significant difference between the standard statistical work. This test was developed by
between the standard deviation of two Prof. Karl Pearson in 1990.
samples. Uses of χ2 - test
Procedure: The uses of chi-square test are:-
1. Set the null hypothesis that there is no 1. Useful for the test of goodness of fit:- χ2 -
significant difference between two standard test can be used to test whether there is
deviations. goodness of fit between the observed
2. Decide the test criterion: frequencies and expected frequencies.
If sample is large, apply Z – test 2. Useful for the test of independence of
If sample is sample, apply F – test attributes:- χ2 test can be used to test
3. Apply the formula: whether two attributes are associated or not.
4. Fix the degree of freedom 3. Useful for the test of homogeneity:- χ2 -
5. Obtain the table value. test is very useful t5o test whether two
6. Decide whether to accept or reject the null attributes are homogeneous or not.
hypothesis. 4. Useful for testing given population
NON-PARAMETRIC TESTS variance:- χ2-test can be used for testing
A non-parametric test is a test which is not whether the given population variance is
concerned with testing of parameters. acceptable on the basis of samples drawn
Nonparametric tests do not make any from that population.
assumption regarding the form of the χ2 -test as a test of goodness of fit:
population. Therefore, non-parametric tests As a non-parametric test, χ2-test is mainly
are also called distribution free tests. used to test the goodness of fit between the
Following are the important non- observed frequencies and expected
parametric tests:- frequencies.
1. Chi-square test (χ2 – test) Procedure:-
2. Sign test 1. Set up mull hypothesis that there is
3. Signed rank test (Wilcoxon matched pairs goodness of fit between observed and
test) expected frequencies.
4. Rank sum test (Mann-whitney U-test and 2. Find the χ2 value using the following
Kruskal-Wallis H test) formula:
CHI-SQUARE TEST
The value of chi-square describes the
magnitude of difference between observed
frequencies and expected frequencies under Where O = Observed frequencies E =
certain assumptions. Χ2 value (χ2 quantity) Expected frequencies
ranges from zero to infinity. It is zero when 3. Compute the degree of freedom. d. f. = n – r
the expected frequencies and observed –1
frequencies completely coincide. So greater Where ‘r’ is the number of independent
the value of χ2, greater is the discrepancy constraints to be satisfied by the frequencies
QT – (Al Jamia Arts and Science College, Poopalam)

4. Obtain the table value corresponding to 4. Obtain the table value


the lord of significance and degrees of 5. Decide whether to accept or reject the null
freedom. hypothesis.
5. Decide whether to accept or reject the null χ2 – test for Population Variance:
hypothesis. If the calculated value is less χ2 – test can be used for testing the given
than the table value, we accept the null population when the sample is small.
hypothesis and conclude that there is Steps:-
goodness of fit. If the calculated value is 1. Set up null and alternative hypotheses:
more than the table value we reject the null H0: There is no significant difference
hypothesis and conclude that there is no between sample variance and population
goodness of fit. variance.
χ2 - test as a test of independence: H0: There is significant difference between
χ2 – test is used to find out whether one or sample variance and population variance.
more attributes are associated or not. 2. Find the χ2 value.
Procedure:-
1. Set up null and alternative hypothesis.
H0: Two attributes are independent (i.e.,
there is no association between the Where S2 = Sample variance , =
attributes)
Population variance
H1: Two attributes are dependent (i.e., there
3. Find the degree of freedom d.f. = n-1
is an association between the attributes)
4. Obtain the table value
2. Find the χ2 value. 5. Decide whether to accept or reject the null
hypothesis
Limitations of Chi-square tests:-
3. Find the degree of freedom d.f. = (r-1)(c-1) 1. It is not as reliable as a parametric test.
Where r = Number of rows Hence it should be used only when
c = Number of columns parametric tests cannot be used.
4. Obtain table value corresponding to the 2. χ2 value cannot be computed when the
level of significance and degree of freedom. given values are proportions or percentages.
5. Describe whether to accept or reject the wilcoxon matched pairs test (signed rank
H0 . If the calculated value is less than the test)
table value, we accept the H0 and conclude Signed rank test was developed by Frank
that the attributes are independent. Wilcoxon. It is an important non-parametric
χ2 – test as a test of homogeneity test. This method is used when we can
χ2 – test is used to find whether the samples determine both direction and magnitude of
are homogeneous as far as a particular difference between matched values.
attribute is concerned. Here there are two cases:-
Steps: a) When the number of matched pairs are
1. Set up null and alternative hypotheses: less than or equal to 25.
H0: There is homogeneity. b) When the number of matched pairs are
H1: There is no homogeneity (heterogeneity) more than 25.
2. Find the χ2 value. ANALYSIS OF VARIANCE
Definition of Analysis of Variance
Analysis of variance may be defined as a
technique which analyses the variance of
two or more comparable series (or samples)
3. Find the degree of freedom d.f. = (r-1)(c-1)
QT – (Al Jamia Arts and Science College, Poopalam)

for determining the significance of


differences in their arithmetic means and for
determining whether different samples
under study are drawn from same

Study
population or not, with the of the statistical
technique, called F – test.
Characteristics of Analysis of Variance:
1. It makes statistical analysis of variance of
two or more samples.
2. It tests whether the difference in the

well…
means of different sample is due to chance or
due to any significance cause.
3. It uses the statistical test called, F – Ratio.
Types of Variance Analysis:
There are two types of variance analysis.
They are:-
1. One way Analysis of Variance
2. Two way analysis of Variance
One way Analysis of Variance:
In one way analysis of variance, observations MUHAMMED RIYAS N
are classified into groups on the basis of a ASST. PROF.
single criterion. For example, yield of a crop AL JAMIA ARTS AND SCIENCE COLLEGE
is influenced by quality of soil, availability of POOPALAM
rainfall, quantity of seed, use of fertilizer, etc. PERINTHALMANNA
It we study the influence of one factor, It is PH : 9747799772
called one way analysis of variance. E-mail- riyasmuhammed89@gmail.com
If we want to study the effect of fertilizer of
yield of crop, we apply different kinds of
fertilizers on different paddy fields and try to
find out the difference in the effect of these
different kinds of fertilizers on yield.
two way analysis of variance
Two way analysis of variance is used to test
the effect of two factors simultaneously on a
particular variable.

Вам также может понравиться