Вы находитесь на странице: 1из 5

The rest of this paper is organized as follows.

Section II presents the procedure of the proposed moving objects detection method. In Section III, experimental results are discussed. Finally, Section IV presents concluding remarks. [3] The video surveillance market is increasingly moving toward cheaper, efficient, portable, and high resolution systems. a need for better video resolution has generated an interest in using high definition (HD) rates for video surveillance. [2] Our prototyped architecture is more hardware-efficient than previous FPGA-based architectures in terms of power consumption, area, throughput, and memory utilization. THERE IS AN increasing need for high quality, real-time video on low power, mobile devices such as mobile phone cameras, bio-medical devices, security and sensor cameras, and so on. Motion estimation (ME) is the most powerful compression tool in the H.264/AVC [1], but it is also the most time consuming. In order to meet low power and real-time constraints, it is critical to efficiently speed up ME via fast algorithms and hardware acceleration.
[4]

This paper deals with an intelligent image processing method for the video surveillance systems. We propose a technology detecting and tracking multiple moving objects, which can be applied to consumer electronics such as home and business surveillance systems. To detect and track the specific moving objects only, it is important to eliminate the environmental disturbances such as light scattering, leaves, birds and so on from input images. To do this, two methods are mainly used. One is to use the Bayesian method such as the particle filter (PF) or the extended Kalman filter (EKF), the other is to use the difference image methods such as the background modeling (BM) or the Gaussian mixture model (GMM). [1] Motion detection is the first essential process in the extraction of information regarding moving objects and makes use of stabilization in functional areas, such as tracking, classification, recognition, and so on. Our method achieves complete detection of moving objects by involving three significant proposed modules: a background modeling (BM) module. The three major classes of methods for motion detection are background subtraction, temporal differencing, and optical flow [13]. Background subtraction [14][23], [33] is the most popular motion detection method and consists of the differentiation of moving objects from a maintained and updated background model, which can be further grouped into parametric type and non-parametric type [33]. Based on the implicit assumption along with the choice of parameters, the parametric model may achieve perfect performance corresponding to the real data along with parametric information. The currently implemented method for background subtraction accomplishes its objective by subtracting each pixel of the incoming video frame from the background model, thus generating an absolute difference. Although the currently implemented background subtraction method is convenient for implementation, the noise tolerance in the video frame relies on the determined threshold. Functionalities such as object classification, tracking, behaviour, and identification are then performed on the regions where moving objects have been detected.

A two-phase background matching procedure is used to select suitable background candidates for generation of an updated background model. The major purpose of background subtraction is to generate a reliable background model and thus significantly improve the detection of moving objects.

Fig. 3. Flowchart for our proposed background modeling procedure

A. Simple Background Subtraction Both the reference image B(x, y) and the incoming video frame It(x, y) are obtained from the video sequence. A binary motion detection mask D(x, y) is calculated as follows:
D(x, y) = 1, if |It(x, y) B(x, y)| > 0, if |It(x, y) B(x, y)|

where is the predefined threshold which designates pixels as either the background or the moving objects in a video frame. If the absolute difference between a reference image and an incoming video frame does not exceed , the pixels of the detection mask are labeled 0, which means it contains background, otherwise, active ones are labeled 1, which designates it as containing moving objects. A significant problem experienced by the SBS method in most real video sequences is that it fails to respond precisely when noise occurs in the incoming video frame It(x, y) and static objects occur in the reference image B(x, y) [20]. Note that the reference image B(x, y) represents the fixed background model, which is selected from the test frames [20]. [9] The aim of motion detection is to identify moving objects from a sequence of image frames. Identifying and localizing moving objects such as humans or vehicles on an image sequence is not straightforward due to the sensitivity on environmental changes. Moving objects in an image sequence generate image difference and detecting each connected component in the difference regarding the intensity and size is the key issue in many motion detection

algorithms. Thus, one may view the problem as segmentation or clustering problem. In this section, we discuss the related models to our new model and address challenges.

[15] To reduce power consumption, there was another way to design a motion detection system. [14] Motion estimation (ME) has been widely used in many video applications such as video compression, video segmentation, and video tracking. [13] Motion estimation (ME) has a key role in compression of video sequences. Motion estimation and motion compensation have been widely used in many video coding standards, such as MPEG-1/2/3 and H.261/263/264, to remove the temporal redundancy between successive frames in video sequences. Many different motion estimation methods are investigated and reported in the literature. [12] Human-behavior analysis based on video surveillance has been investigated worldwide. As a consumer video application, automatic surveillance requires a sufficiently high accuracy and the computation complexity should enable a real-time performance. With the continuous improvements in video analysis techniques, automatic low-cost video surveillance gradually emerges for consumer applications. Video surveillance can contribute to the safety of people in the home and ease control of home-entrance and equipment usage functions. [19] Motion detection plays an essential role in any surveillance system, occupancy-detection applications and moving object systems. It is also widely used in real-time image processing applications such as wireless security systems and occupancy-detection systems, medical instruments and toys. These applications extract the movement of objects and detect motion from currently acquired images. The surveillance and occupancy-detection can be achieved, if the change of motion can be detected if differences exist between the current image and the previous image. We designed a user interface program to provide seamless integration with the surveillance camera system. After motion detection, this system stores the image data into flash memory. At the same time, the size and information of the detected images are also saved in the flash memory. When motion is detected, the operation mode is switched into capture mode and the full image is shown on the screen and is stored into the flash memory. The transfer mode then transfers the data to the PC. [20] Generally, there are three types of moving objects detection methods. They are optical flow method, temporal difference method and background subtraction method Compared with other types of methods, background subtraction method is more practical and effective. It uses the distortion between current frame and background frame to detect moving objects. [21] On the other hand, with the improved fabrication technology and architecture, CMOS Image Sensors (CIS) have significant advantages : (1) Low voltage operation and low power consumption, (2) compatibility with integrated functional blocks, such as, control blocks, analog signal processors and Analog to Digital Converters (ADC).

3. CMOS IMAGE SENSOR DESIGN 3.1 GENERAL SENSOR ARCHITECTURE A general CMOS Image Sensor (CIS) is shown in Figure 3. The CIS is generally composed of four building blocks : (1) Pixel array absorbing light flux and generating photocurrent. Since photocurrent is very small, it is collected into a capacitor and amplified. (2) Analog Signal Processor (ASP) processing an analog signal of the acquired voltage value of each pixel in the column line. This block reduces the noise using the double sampling technique. (3) Analog to Digital Converter (ADC) converting analog signal into digital data of each pixel data. (4) Control and Timing block generating timing signals to control the entire circuits.

[5]
In centralized transportation surveillance systems, video is captured and compressed at low processing power remote nodes and transmitted to a central location for processing. Such compression can reduce the accuracy of centrally run automated object tracking algorithms. Transmitting high-quality video requires expensive wired communication links, while relying on lower cost wireless links requires heavy compression and results in reduced video quality. Most video compression systems use block-based motion compensation, where temporal redundancy is eliminated via the use of block motion vectors and frequency-transformed residuals. For typical transportation surveillance the camera is stationary, and the majority of bitrate is spent representing temporal changes to the scene due primarily to acquisition noise or small changes to lighting. [10] On the other hand, most current watermarking schemes focus on the frequency domain rather than the spatial domain because the characteristics of the former are more robust, invisible, and stable. In the frequency domain, a watermark is inserted into coefficients obtained by using an image transform process. The common well-known transform methods are the discrete cosine transform (DCT), discrete Fourier transform (DFT), and discrete wavelet transform (DWT). [16] Two key challenges in video communications for both wired and wireless environments are efficient compression and effective error protection against channel errors. We consider the problem of error protection for transmitting compressed video data. Specifically, we consider unequal error protection when compression is based on the three dimensional discrete cosine transform. A major problem in video coding at very low bit rates is how to protect the highly compressed bit streams from the significant channel errors during transmission. The general method used to handle channel errors is to trade some redundancy for data protection, using forward error correction methods. This is achieved by a controlled introduction of some redundant bits to the compressed video stream, via channel coding.

Вам также может понравиться