Академический Документы
Профессиональный Документы
Культура Документы
Project Report on
Submitted By
Saurabh Sharma (Roll No.10506058)
Dewakar Prasad(Roll No.10506060)
Manish Tripathy(Roll No 10406023)
Session: 2008-09
Submitted By
Saurabh Sharma (Roll No.10506058)
Dewakar Prasad(Roll no.10506060)
Manish tripathy(Roll No.10406023)
Session: 2008-09
2
Department of Computer Science & Engineering
National Institute of Technology
Rourkela-769008
Orissa
This is to certify that that the work in this thesis report entitled Image compression using
wavelet transform submitted by Saurabh Sharma ,Dewakar Prasad and Manish tripathy in
partial fulfillment of the requirements for the degree of Bachelor of Technology in Computer
Science & Engineering Session 2005-2009 in the department of Computer Science &
Engineering, National Institute of Technology Rourkela, This is an authentic work carried out by
them under my supervision and guidance.
To the best of my knowledge the matter embodied in the thesis has not been submitted to any
other University /Institute for the award of any degree.
3
ACKNOWLEDGEMENT
Saurabh Sharma(10506058)
Dewakar Prasad(10506060)
Manish tripathy(10406023)
4
CONTENTS
Chapter Page No.
Certificate 3
Acknowledgement 4
Abstract 7
Chapter 1 Literature Review 8-12
1.1 Introduction 9
1.2 Why compression is needed 10
1.3 Fundamental of image compression technique 11
1.4 Objective 12
1.5 Organization of Report 12
Chapter 2 Image compression methodology 13-25
2.1 Overview 14
2.2 Different types of transform used for coding 15
2.3 Entropy coding 24
Chapter 3 Wavelet Transform 23-47
3.1 Overview 23
3.2 What are basis function 26
3.3 Fourier Analysis 30
3.4 Similarities between Fourier and wavelet transform 32
3.5 Dissimilarities between Fourier and wavelet transform 33
3.6 List of Wavelet transform 34
Chapter 4 Results and Dicussion 48-58
4.2 Results 49
4.3 Conclusion 58
NOMENCLATURE 59
REFERENCES 60-61
5
6
ABSTRACT
7
CHAPTER 1
LITURATURE REVIEW
8
CHAPTER 1: LITURATURE REVIEW
1.1 INTRODUCTION
9
Here one more factor, which is not considered, is the processing power
requirement at both the ends i.e. at the Server/Mobile to Mobile/Server.
Thus in this paper we have considered all of these parameters like the
processing power required in the mobile handset which is limited and
also the processing time considerations at the server/mobile ends which
will handle all the loads.
Since images will constitute a large part of future wireless data, we focus
in this paper on developing energy efficient, computing efficient and
adaptive image compression and communication techniques. Based on a
popular image compression algorithm, namely, wavelet image
compression, we present an Implementation of Advanced Image
Compression Algorithm Using Wavelet Transform.
11
Lossy compression often follows the same pattern as lossless, but with
one or more quantization steps somewhere in (A). Sometimes clever
designers may defer the loss until suggested by statistics detected in (C);
an example of this would be modern zero tree image coding.
(B) Organization by context often means data reordering, for
which a simple but good example is JPEG's "Zigzag" ordering.
The purpose of this step is to improve the estimates found by the
next step.
(C) A probability estimate (or its heuristic equivalent) is formed
for each token to be encoded. Often the estimation formula will
depend on context found by (B) with separate 'bins' of state
variables maintained for each conditioned class.
Coding Redundancy
If the gray levels of an image are coded in a way that uses more code
symbols than absolutely necessary to represent each gray level, the
resulting image is said to contain coding redundancy. It is almost always
present when an images gray levels are represented with a straight or
natural binary code. Let us assume that a random variable r K lying in the
13
interval [0, 1] represents the gray levels of an image and that each r K
occurs with probability Pr (r K).
Pr (r K) = N k / n where k = 0, 1, 2 L-1
L = No. of gray levels.
N k =No. of times that gray appears in that image
N = Total no. of pixels in the image
If no. of bits used to represent each value of r K is l (r K), the average no.
of bits required to represent each pixel is
L avg = l (r K) Pr (r K)
That is average length of code words assigned to the various gray levels
is found by summing the product of the no. of bits used to represent each
gray level and the probability that the gray level occurs. Thus the total
no. of bits required to code an MN image is MN L avg.
The Information of any given pixel can be reasonably predicted from the
value of its neighbouring pixel. The information carried by an individual
pixel is relatively small.
In order to reduce the inter pixel redundancies in an image, the 2-D pixel
array normally used for viewing and interpretation must be transformed
into a more efficient but usually non visual format. For example, the
differences between adjacent pixels can be used to represent an image.
These types of transformations are referred as mappings. They are called
reversible if the original image elements can be reconstructed from the
transformed data set.
14
Psycho visual Redundancy
15
Image Compression Model
16
Lossy Coding Techniques:
1. Image Modeling:
It is aimed at the exploitation of statistical characteristics of the
image (i.e. high correlation, redundancy). It defines such things as
the transformation to be applied to the Image.
2. Parameter Quantization:
The aim of Quantization is to reduce the amount of data used to
represent the information within the new domain.
3. Encoding:
Here a code is generated by associating appropriate code words to
the raw produced by the Quantizer. Encoding is usually error free. It
optimizes the representation of the information and may introduce
some error detection codes.
17
that must be preserved, the depth of field that must be captured, whether
they can be placed on a glass platen or require a custom book-edge
scanner, whether they can tolerate exposure to high light intensity, and
whether specular reflections must be captured or minimized. A detailed
examination of some of the originals, perhaps with a magnifier or
microscope, may be necessary to determine the level of detail within the
original that might be meaningful for a researcher or scholar. For
example, in drawings or paintings it may be important to preserve
stippling or other techniques characteristic
18
CHAPTER 1: LITURATURE REVIEW
1.4 OBJECTIVE
19
CHAPTER 2
Image Compression Methodology
20
2.1 Overview
1. FT (Fourier Transform)
2. DCT (Discrete Cosine Transform)
3. DWT (Discrete Wavelet Transform)
21
2.2.2 The Discrete Cosine Transform (DCT):
The discrete cosine transform (DCT) helps separate the image into parts
(or spectral sub-bands) of differing importance (with respect to the
image's visual quality). The DCT is similar to the discrete Fourier
transform: it transforms a signal or image from the spatial domain to the
frequency domain.
22
image of stars on a night sky. This means that the transient elements of a
data.
23
2.3 Quantization:
24
According to Shannon's source coding theorem, the optimal code length
for a symbol is logbP, where b is the number of symbols used to make
output codes and P is the probability of the input symbol.
Three of the most common entropy encoding techniques are Huffman
coding, range encoding, and arithmetic coding. If the approximate
entropy characteristics of a data stream are known in advance (especially
for signal compression), a simpler static code such as unary coding,
Elias gamma coding, Fibonacci coding, Golomb coding, or Rice coding
may be useful.
There are three main techniques for achieving entropy coding:
Huffman Coding - one of the simplest variable length coding
schemes.
Run-length Coding (RLC) - very useful for binary data
containing long runs of ones of zeros.
Arithmetic Coding - a relatively new variable length coding
scheme that can combine the best features of Huffman and run-
length coding, and also adapt to data with non-stationary statistics.
25
26
Chapter 3
WAVELET TRANSFORM
27
CHAPTER 3: NUMERICAL MODELING
3.1 OVERVIEW
28
original signal or function can be represented in terms of a wavelet
expansion (using coefficients in a linear combination of the wavelet
functions), data operations can be performed using just the
corresponding wavelet coefficients. And if you further choose the best
wavelets adapted to your data, or truncate the coefficients below a threshold,
your data is sparsely represented. This sparse coding makes wavelets
an excellent tool in the field of data compression.
The best basis vectors have the valuable extra property that the
vectors are perpendicular, or orthogonal to each other. For the basis
(1,0) and (0,1), this criteria is satisfied.
Now let's go back to the analog world, and see how to relate these
concepts to basis functions. Instead of the vector (x,y), we have a
function f(x). Imagine that f(x) is a musical tone, say the note A in a
particular octave. We can construct A by adding sines and cosines
using combinations of amplitudes and frequencies. The sines and
cosines are the basis functions in this example, and the elements of
29
Fourier synthesis. For the sines and cosines chosen, we can set the
additional requirement that they be orthogonal. How? By choosing
the appropriate combination of sine and cosine function terms whose
inner product add up to zero. The particular set of functions that are
orthogonal and that construct f(x) are our orthogonal basis functions
for this problem.
30
3.3 Fourier analysis
FOURIER TRANSFORM
The Fourier transform's utility lies in its ability to analyze a signal in
the time domain for its frequency content. The transform works by
first translating a function in the time domain into a function in the
frequency domain. The signal can then be analyzed for its frequency
content because the Fourier coefficients of the transformed function
represent the contribution of each sine and cosine function at each
frequency. An inverse Fourier transform does just what you'd expect,
transform data from the frequency domain into the time domain.
DISCRETE FOURIER TRANSFORM
The discrete Fourier transform (DFT) estimates the Fourier transform
of a function from a finite number of its sampled points. The sampled
points are supposed to be typical of what the signal looks like at all
other times.
The DFT has symmetry properties almost exactly the same as the
continuous Fourier transform. In addition, the formula for the inverse
discrete Fourier transform is easily calculated using the one for the
discrete Fourier transform because the two formulas are almost
identical.
31
signal has sharp transitions, we window the input data so that the
sections converge to zero at the endpoint. This windowing is
accomplished via a weight function that places less emphasis near the
interval's endpoints than in the middle. The effect of the window is to
localize the signal in time.
The fast Fourier transform (FFT) and the discrete wavelet transform
(DWT) are both linear operations that generate a data structure that
contains segments of various lengths, usually filling and
transforming it into a different data vector of length .
32
Both transforms have another similarity. The basis functions are
localized in frequency, making mathematical tools such as power
spectra (how much power is contained in a frequency interval) and
scale grams (to be defined later) useful at picking out frequencies and
calculating power distributions.
33
3.6 LIST OF WAVELET RELATED TRANSFORM
1. Continuous wavelet transform
A continuous wavelet transform is used to divide a continuous-time
function into wavelets. Unlike Fourier transform, the continuous
wavelet transform possesses the ability to construct a time frequency
represented of a signal that offers very good time and frequency
localization.
2 .Multiresolution analysis
A multiresolution analysis (MRA) or multiscale approximation
(MSA) is the design methods of most of the practically relevant
discrete wavelet transform (DWT) and the justification for the
algorithm of the fast Fourier wavelet transform (FWT)
34
35
36
3.2 HAAR WAVELET
37
Wavelets are mathematical functions that were developed by
scientists working in several different fields for the purpose of sorting
data by frequency. Translated data can then be sorted at a resolution
which matches its scale. Studying data at different levels allows for
the development of a more complete picture. Both small features and
large features are discernable because they are studied separately.
Unlike the discrete cosine transform, the wavelet transform is not
Fourier-based and therefore wavelets do a better job of handling
discontinuities in data.
The Haar wavelet operates on data by calculating the sums and
differences of adjacent elements. The Haar wavelet operates first on
adjacent horizontal elements and then on adjacent vertical elements.
The Haar transform is computed using:
38
CHAPTER 4
RESULTS AND DISCUSSION
39
CHAPTER 4 : RESULTS AND DISCUSSION
4.1 RESULTS
The image on the left is the original image and the image on
the right is the compressed one
(The point is that the image on the left you are right now
viewing is compressed using Haar wavelet method and the
loss of quality is not visible. Of course, image compression
using Haar Wavelet is one of the simplest ways.)
40
4.2 CONCLUSION
41
BIBILOGRAPHY :- [1
42
43