Speed Determination of a Moving object from a Video Sequence using Bounding Box Approach and Region Based Segmentation

DOI : 10.17577/IJERTV3IS20537

Download Full-Text PDF Cite this Publication

Text Only Version

Speed Determination of a Moving object from a Video Sequence using Bounding Box Approach and Region Based Segmentation

Prodip Kumar Sarker

Lecturer

Department of Computer Science & Engineering

Bangabandhu Sheikh Mujibur Rahman Science and Technology University, Gopalganj, Bangladesh

Md. Jamal Uddin

Lecturer

Department of Computer Science & Engineering Gono Bishwabidyalay (University), Dhaka, Bangladesh

Abstract-Traveling speed determination of a moving object from a video sequence is a critical task in real world applications. This paper concerned with the determination of the traveling speed of a moving object from a video sequence based on subsequent object detection techniques. After pre- processing of the original video sequence, which is sampled from the video camera, the target moving object is detected with the improved algorithm in which the moving object region can be extracted completely through several processing of feature extraction and region based. Among the multiple moving objects of the video, the target object has been detected based on particular criteria of region that it occupies. The traveling speed of the target moving object from changes of its coordinate position from the video frames is determined by calculating the distance that the object moved in a sequence of frames with respect to the frame rate that the video is recorded.

Keyword-Object detection, feature extraction, region based segmentation, moving object and traveling speed.

  1. INTRODUCTION

    The traveling speed determination of a selected moving object from a suitable video sequence is an important aspect in some real world activity. Information about the speed of regions in an image can be also important to the scene interpretation. Object trajectories may be calculated, structural ambiguities may be solved, important events noted, and future situations predicted. At lower levels, information about speed of small structures in an image sequence can be used as an important cue for image segmentation. A discontinuity in the speed of small image areas strongly implies the existence of an object boundary. Visually dissimilar image regions with the same speed may either be part of the same object or semantically linked in the overall scene interpretation. If an observer is moving, perceived speed represents an important depth cue. Because of the value of speed information, increasing attention is being paid to methods analyzing time- varying visual environments.

    However, several methods for speed determination of some customized moving object from video sequence have been developed to date. All of the methods required to detect the image object due to the positional shift in each frame in the given video sequence. In this paper our proposed method is quite simple and efficient to determine the traveling speed of the moving object from a suitable video sequence. In this method, firstly, we need to detect the target object that moves from initial frame to the last frame in the given video sequence. An alternate approach is directly calculating the speed information from spatial gradient of the images and local intensity changes over time due to speed. The speed of a moving object is calculated by the distance it traveled with respect to the time. Euclidean distance formula is used to calculate the distance between the sequences of frames. By using the values of distance with respect to frame rate, the speed of the object is defined. The defined is of 2- dimensional.

    Each method has limitations. Obviously, procedures which require an accurate segmentation cannot be used to assist in the segmentation process itself. Methods dependent on cross correlation are computationally expensive. While some efficiency are possible, much effort still required to perform the matching at a dense enough sampling of points to be useful for tasks such as segmentation. Techniques using frame-to-frame differences provide efficiency, but require sophisticated higher-level processing to deal effectively with occlusions and situations in which object boundaries are not completely within the image frame. They relate intensity changes over time at appoint to the spatial intensity gradient in the neighborhood of the point. The frame rate is assumed to be sufficiently rapid such that the value of the gradient at a point does not change significantly between frames. Speed along the x and y axes are treated independently.

  2. PROPOSED ALGORITHM FOR SPEED

    DETERMINATION

    The procedure first taking a video sequence that contain moving object then process the video sequence to get all the frames and then convert video frames to image format. To reducing noise from image use median filter. Segmentation technique is applied for separation of foreground and background objects. Frame difference algorithm is used for separation of two sequential frames [6]. After performing segmentation operation a rectangular bounding box method is applied for feature extraction. After performing feature extraction operation to detect the moving object from the background scene and calculating the traveling speed between each adjacent pair of frames. Finally calculate the average traveling speed that is the required speed. To calculate traveling speed at first calculates distance between two consecutive frames.

    Taking input video sequence that

    contain moving object

    Process the video sequence to get all the frames and convert video

    Process each frame to detect moving object from the background scene

    Detect the target moving object and find out the coordinate position of the object

    Feature Extraction

    Calculation the distance between two adjacent frames

    Calculation the traveling speed of moving object in each adjacent pair of frames

    Now the average traveling speed is the required speed

    Figure 1: Flowchart of overall process for traveling speed determination from a video sequence

  3. PROCESSING VIDEO FOR SPEED

    DETERMINATIONS

    Video processing is a particular case of signal processing, which often employs video filters and where the input and output signals are video files or video streams. A video signal is a sequence of two dimensional (2D) images projected from a dynamic three dimensional (3D) scene on to the image plane of a video camera. The color value at any point in a video frame records the emitted or reflected light at a particular 3D point in the observed scene. Video sequences from the video camera is taken and process it as needed to convert AVI format and get all the frames of that video sequence which are inputted to the next phase of this work.

  4. DETECTION OF ALL MOVING OBJECTS Detection of all moving objects is composed of

    the procedure Bounding box extraction and Frame difference Segmentation which is the most important part of this work. Segmentation and feature extraction procedures are given bellow:

    Feature Extraction

    For detecting object in Speed analysis can be viewed as three different problems [3].

    * The first is the case when the camera is moving and the objects in the world are stationary. In this case, the extraction of camera motion is a challenge.

    • In the second case, the camera is stationary, and objects in the world are moving.

    • It is the combination of the two, where both the camera and some objects in the world are moving.

    As, in our work the camera is stationary, so second case is applicable to this point. Different algorithm is usually applied in the second case. In this case, difference algorithm can be divided into two types: one is difference between continuous images; the other is difference between current image and background images. For difference between current image and background image, suppose that the gray value of current image at position (x, y) is f (x, y), the gray value of background image at position (x, y) is b(x, y), the difference between images can be written as: d(x, y) = f(x, y) b(x, y) .. (1)

    For difference between continuous images, suppose that the gray value of image at position (x, y) at time t is f (x, y, t), the gray value of image at position (x, y) at time t+1 is f(x, y,t+1), the difference between images can be written as:

    d(x, y) = f(x, y, t+1) b(x, y, t) ..(2)

    Every object has a specific feature which is used to visualize the object and used for tracking. In this paper to extracted specific feature of an object for tracking, bounding

    box algorithm is used. Algorithm for bounding box is as follows:

    1. Take the input video sequence

    2. Extract frames from video sequence

    3. Convert frame to image sequence

    4. Read the image difference

    5. for(present position=initial value: final value) of Y resolution

    6. for(present position=initial value: final value) of X resolution

      1. calculate the sharp change in intensity of image from top and bottom

      2. store the values in an array

    7. Height of the bounding box is = bottom value top value

    8. for(present position=initial value: final value) of X resolution

    9. for(present position=initial value: final value) of Y resolution

      1. calculate the sharp change in intensity of image from left and right

      2. store the values in an array

    10. Width of bounding box = right value left value

    11. Using the dimensions, draw boundary to the image Initial value: the starting position of the pixel in an

      However, it also has drawbacks:

      • Decisions about region membership are often more difcult than applying edge detectors.

      • It cant nd objects that span multiple disconnected regions. (Whereas edge-based method can be designed to handle gaps produced by occlusionthe Hough transform is one example.)

        The objectives of region-based approaches can be summarized as follows:

      • Produce regions that are as large as possible (i.e., produce as few regions as possible).

      • Produce coherent regions, but allow some exibility for variation within the region. Notice the inherent tradeoffs here. If we require that the pixels in a region be too similar, we get great coherency and probably wont span separate objects, but we over segment the image into regions much smaller than the actual objects. If we allow more exibility, we can produce larger regions that more likely ll entire objects, but they may cross multiple objects and leak across what should otherwise be boundaries. Remember: the real goal is to nd regions that correspond to objects as a person sees themnot an easy goal [2].

        The basic formulation for Region-Based segmentation is: (a) = .

        image. Final value: the ending the starting position

        of the pixel in an image

        (b)

        =1

        is a connected region, i= 1, 2,n

        =

        =

        . .

        2

        . .

        2

        (c) = = 1,2, , .

        (d) = = 1,2, , .

        (e) ( ) = for any adjacent is a

        logical predicate defined over the points in the

        is the null set.

    12. Add the height value with top value and store it in a variable like mid top

    13. Add the width value to the left value and store it in a variable like mid left

    14. Assign the max intensity to the pixel at pixel value at (mid top, mid left).

    Object Segmentation

    The objective of segmentation is to partition an image into regions. When a moving object is segmented, a region of pixels assigned to the object is available. This region can be tracked using approaches like cross- correlation. The location of the region in the next frame is to be determined. A moving object usually corresponds to one or several tracked regions. Combination of several regions to one object is then performed at a higher level of abstraction [1].To perform segmentation operation, region based method is used because it takes less processing time. We now turn from segmentation by nding boundaries (pixel differences) to segmentation by nding coherent regions (pixel similarities). This approach has specic advantages over boundary based methods:

    • It is guaranteed (by denition) to produce coherent regions. Linking edges, gaps produced by missing edge pixels, etc. are not an issue.

    • It works from the inside out, instead of the outside in. The question of which object a pixel belongs to is immediate, not the result of point-in-contour tests.

    1. Means that the segmentation must be complete; that is, every pixel must be in a region.

    2. Requires that points in a region must be connected in some predefined sense.

    3. Indicates that the regions must be disjoint.

    4. Deals with the properties that must be satisfied by the pixels in a segmented region. For example

      = if all pixels in have the same gray level.

    5. Indicates that region and are different in the sense of predicate [4].

  5. OBJECT DETECTION AND FIND OUT ITS POSITION

    To identify a single object as target object with its 2Dcoordinate position from multiple object in each frame from a video sequence, our algorithm always detect the object that is occupied the maximum region [5]. So, when we will take the video sequence for speed determination of the target object, we will focus on the target object as much as possible that the object will occupy maximum region compared to the other moving object. And of course the camera must static. To identify the position of the target object in each frame of input video sequence the centred point of the total region that is occupied by the object have been considered as reference point.

    In the similar way, the reference point of target object in each frame of the video is find out and stores these

    positions. Finally from these positions, the movement of target object is measured and the traveling speed is calculated according to the speed calculation procedure.

    Procedure for object detection: for i=0 to (totalFrame-1) do

    1. Read frame[i],

    2. Update frame[i] using Extract bounding box,

    3. process frame[i] as follows :

      1. Determine the connected components.

        1. Run-length encodes the input image.

      1. Calculate change in distance by distance

        = 2 1 2 + 2 1 2

        Where 1 = previous pixel position and 2 = present pixel position in width

        1 = previous pixel position and 2 = present pixel position in height.

      2. Store all the distance values in an array.

        1. Calculate the speed between the frames and

          1. Scan the runs, assigning preliminary labels and recording label equivalences in a local

            +1

            as

            =

            equivalence table.

          2. Resolve the equivalence classes. Relabel the runs based on the resolved equivalence classes.

            1. Compute the area of each component

        2. Repeat step 6 to 7 for i= 0 to 2to determinate all the speed between the frames.

        3. Calculate the average value of speed as () for

        1

        i=0 to 2

        1. Remove small objects bellow a threshold.

      10.

      = ( )

      ( )

    4. Create morphological structuring element.

    5. Close the binary image by the structuring element.

    6. Measure image regions

    7. Find the maximum region

    8. Identify the centered location (x, y) of that region.

    9. Return x-coordinate value and y-coordinate value

    10. End

  6. Speed Determination of A Selected Moving Object

    Several methods for speed determination of some customized moving object from video sequence have developed to date. All of the methods required to detect the image object due to the positional shift in each frame in the given video clip. In our work our proposed method is quite simple and efficient to determinate the traveling speed of the moving objects from video sequence. In this method, firstly, we need to detect the target object that moves from initial frame to the last frame in the given video clip that has already been discussed above. The speed of a moving object is defined as the total amount of distance traveling in unit time.

    Algorithm/procedure for speed determination of a selected moving object

    1. Load the input video file containing moving objects.

    2. Process the video file to get the required information about the video file.

    3. Find the number of frames of the video

    4. Find the frame rate of the video

    5. Calculate the total duration of the video as

    11. is the real speed (meter/second) of the moving object.

  7. EXPERIMENTAL RESULTS AND

    DISCUSSION

    Our approach is implemented on a PC under Windows operating system using MATLAB to execute all the required source code. The technique discussed above is applied on various video files like as inputvideo1.avi, inputvideo2.avi etc. The first code allows the user to select a video file is executed. The results of the traveling speed of a selected target moving object from a video sequence are given bellow: Several attributes such as file size, number of frames, frame rate are find out from selected input video sequences.

    Several frames of inputvideo1.avi are given bellow:

    Fig 2a: 1st frame Fig 2b: 2nd frame

    second and unit time =

    1

    6. Determinate the displacement Di of the object

    between the i-th frame and (i+1) the frame

    1. Read the centroid position of each frame.

    2. Calculate the distance between two centroid frames.

    3. for (present position = initial value : final value) of X resolution.

    4. for (present position = initial value : final value) of Y resolution.

    Fig 2c: 4th frame 2d: Last frame

    In first input video the file size is 73728 Kb, number of frames 100 and frame rates per second is 15. After getting the object position in each frame of the input video the traveling speed of that object have calculated with the speed calculation procedure. Using the proposed speed determination algorithm the resultant traveling speed of the moving object is 30 meters per second.Several frames of inputvideo1.avi are given bellow:

    Fig 3a: 1st frame Fig 3b: 2 nd frame

    Fig 3c: 5th frame 3d: Last frame

    In second input video the file size 23020 Kb, number of frames 140 and frame per second is 25. After getting the object position in each frame of the input video the traveling speed of that object have calculated with the speed calculation procedure. Using the proposed speed determination algorithm the resultant traveling speed of the moving object is 1.52 meter per second.

  8. CONCLUSION

    In this paper, an attempt has been made to develop a virtual system for determination the traveling speed of a selectable moving object of a suitable video clip using subsequent object detection technique based on feature extraction and region based segmentation near to the real time [7]. Feature extraction using bounding box approach and the region based segmentation techniques are relevant to detect multiple moving object from a video clip to determinate the traveling speed of a target moving. As we know that object detection technique is not completely efficient for all kinds of objects which are available presently all over the world, so this work has experimentally been shown to be quite accurate and effective in detecting the target moving object from a video sequence and also determined the traveling speed. After all, for the test bench for this work, the traveling speed of a selected moving object of a suitable video clip has been determined at a satisfactory level. In this paper, the primary works are the video processing as well as image processing for the detection of moving object within the video clip, but it focuses on the detection of multiple objects from images in the video sequences and detecting the target object based on region that it occupies to determine the traveling speed of the moving object.

  9. REFERENCES

  1. Alper Yilmaz, Omar Javed, and Mubark Shah, Object tracking: A survey, ACM Comput. Surv., 38(4): 13, 2006.

  2. Bryan S. Morse, Segmentation (Region Based), Brigham Young University,19982000, Last modied on March 3, 2000.

  3. Yong Fan1, Zhengyu Zhang2, Journal of Communication and Computer, ISSN1548-7709, USA Jul. 2006, Volume, No.7 (Serial No.20).

  4. http://en.wikipedia.org/wiki/Region_growingRegion growing From Wikipedia, the free encyclopedia

  5. Shuan Wang, Haizhou Ai, Kezhong He, Difference-image- basedMultiple Motion Targets Detection and Tracking, Journal of Image andGraphics, Vol. 4, No. 6(A), Jun., 1999: pp. 270-273.[8]

  6. Gonzaleg, R.C. Woods, R.E [1992], Digital Image Processing.

  7. Murat Tekalp, Digital Video Processing, Tsinghua University Press andPrentice Hall, Beijing, 1998.[7]

Leave a Reply