SlideShare a Scribd company logo
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 230
INTEGRATION OF POSES TO ENHANCE THE SHAPE OF THE
OBJECT TRACKING FROM A SINGLE VIEW VIDEO
Hemavathy J1
, Shahul Hameed. C2
, Ani Bernish. S3
1
Asst.Prof, Information Technology, Panimalar Engineering College, Tamil Nadu, India
2
Asst.Prof, Information Technology, Panimalar Engineering College, Tamil Nadu, India
3
Asst.Prof, Information Technology, Panimalar Engineering College, Tamil Nadu, India
Abstract
In computer vision, tracking human pose has received a growing attention in recent years. The existing methods used multi-view
videos and camera calibrations to enhance the shape of the object in 3D view. In this paper, tracking and partial reconstruction
of the shape of the object from a single view video is identified. The goal of the proposed integrated method is to detect the
movement of a person more accurately in 2D view. The integrated method is a combination of Silhouette based pose estimation
and Scene flow based pose estimation. The silhouette based pose estimation is used to enhance the shape of the object for 3D
reconstruction and scene flow based pose estimation is used to capture the size as well as the stability of the object. By integrating
these two poses, the accurate shape of the object has been calculated from a single view video.
Keywords: Pose Estimation, optical Flow, Silhouette, Object Reconstruction, 3D Objects
--------------------------------------------------------------------***----------------------------------------------------------------------
1. INTRODUCTION
Object reconstruction in three dimensional views is an
emerging topic in computer vision. This is mainly used in
computer graphics, animation, biomedical, surveillance and
for many other analysis purposes. 3D object reconstruction
is the construction of 3D shape of the object from a video,
which is having the two dimensional information about the
3D objects in sequence of frames. For object tracking,
different types of removable physical markers are used on
the moving object in the recorded video. Using the
transformation of markers or sensors, the movement or
motion trajectory of the object is tracked. They are very
expensive and used only to track the object which is already
having the sensors or markers in the body.
In Adaboost edge detection technique [14], multi-view 3D
human pose estimation is still a difficult task in uninhibited
environments. Several associated approaches are still
dependent on silhouettes of the topic obtained by
background subtraction. Background subtraction is complex
if the cameras are moving or the background is vibrant.
Scene flow [15] is the three dimensional action field of
position in the world coordinate; just as optical flow is the
two dimensional action field of position in an image. It
presents a structure for the calculation of dense, non-rigid
scene flow from optical flow. This method does include the
knowledge of structure computed independently at the next
time instant and also the optical flow which results in less
accuracy.
The proposed method is marker less and we have used
single view camera to capture the movement of a person. To
reconstruct the shape of the object from the 2D sequence of
images without the camera calibration, estimate the initial
pose and its registration is the main task. The rest of paper is
organized as follows. We discuss the related work in section
II, Section III explains in detail about the proposed work,
results and evaluation metrics are discussed in Section IV
and Section V provides the conclusion of the paper.
2. RELATED WORKS
3D shape reconstruction is the developed in computer vision
field, which is mainly used for visual effects and the
analysis purposes. To construct the 3D objects from a 2D
video sequence, pose estimation, silhouette extraction and
3D scene flow estimation are the important tasks. The
various pose estimation methods are proposed in [1], [2],
[3], and [4].
Yogesh Rathi et al. [1] and Jehoon Lee et al. [2] have
proposed particle filtering method for pose estimation. [1]
Suggested to use particle filtering in static images for
segmentation and pose registration problems. But it is
producing major occlusion effects between objects. [2]
Track the rigid body object in a 2D image sequence and to
estimate the pose in 3D space. Different 3D model
reconstruction methods are proposed in [3], [4] and [11].
Soon-Yong Park [3] et al comes with the suggestion for
automatic three-dimensional model reconstruction
technique, which acquire the complete and closed 3D
models for real object.
Youngbae Hwang et al. [11] developed a method for 3D
reconstruction using Silhouette Extraction which suggests
three refinements to get the accurate silhouettes of the
moving objects. Christoph Vogel et al. [5] has given one
novel method for 3D scene flow estimation, its regularizing
3D scene flow computation by penalizing deviations from
local rigidity of motion and integrated into an energy
minimization framework. Jung-Ho Ahn [6] et al, Matt Olson
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 231
et al. [7], Bruce Gooch [8], Alexandre Alahi et al. [10],
Youngbae Hwang et al. [11] and Robert T. Collins et al.
[12] proposed various methods for silhouette extraction. [6],
[8], [10] and [12] implemented the simple background
subtraction algorithms to compute the silhouettes of the
moving objects. [7] Using octree data structure to capture
the silhouettes.
3. PROPOSED WORK
Initial pose estimation is the task of registering the shape of
the object being tracked. It is used to adjust the position and
orientation of the object. The proposed integrated method is
used to capture the accurate shape of the object. Depending
upon the accurate result from the pose estimation method
the 3D shape can be reconstructed from a single view video.
In this paper, the proposed method is used to estimate the
accurate object from the human motion. First method, Scene
Flow based Pose Estimation method is used to capture the
scale and stability of the object and the second method,
Silhouette based pose registration is used to enhance the
shape of the object from the 2D video. By integrating these
methods, the shape of the object can be registered more
accurately. Fig.1 shows the overview of the proposed
method. Initial pose of the moving object is registered first,
which is computed by the active contour method. This initial
pose will be used for comparison with further movement of
the object.
3.1. Scene Flow Based Pose Estimation Method
Scene flow is the structure of evident movement of objects,
edges and surfaces. Optical flow is the motion points of an
object in a two dimensional space coordinates. Scene flow is
the motion points of moving object in three dimensional
world coordinates. It is used to estimate the motion of
moving objects from the series of frames. It is computed
between two consecutive frames in all the video frames.
Scene flow composes of two components namely Normal
Flow and Parallel Flow. The direct method without any
constraints is called as normal flow. We have used
hierarchical version of Lucas-Kanade optical flow algorithm
for parallel flow. This standard algorithm is a predictor for
efficient and robust stereo. Optical flow vector is the
temporal rate of change of the image plane coordinates (u,
v) at a particular point (x, y) is determined by the spatial-
temporal variations of the intensity pattern I(x, y). The input
frames are shifted to the world coordinates from the special
coordinates and the gradient function is computed by using
the gradient formula for the input frames.
(1)
The linear system is updated with its displacement values,
before value updating the error frame (T(x, y) - I(x, y)) and
Hessian matrix are generating. The transformation of the
moving object is then calculated using (2). The two frames
are separated by a small time increment ∆t and the
movement vector (u, v) obtained by comparing the two
consecutive images. Ix(x,y) is the increase in brightness per
pixel at position (x, y) in x direction. Iy(x,y) is the increase
in brightness per pixel at position (x, y) in y direction. The
optical flow is defined as motion field in 2D dimensional
view. Scene flow is defined as motion field of 3D
dimensional.
....... (2)
By combining the nearest neighbour pixel‟s information
Lucas-Kanade algorithm solves the optical flow equation.
Lucas-Kanade optical flow algorithm divides the frame or
image into smaller blocks and assuming the constant
velocity in each block.
Fig-1.: Architectural Diagram of Integrated Method
3.2. Silhouette based Pose Estimation Method
A silhouette may be the image of a person, an object or a
scene and it represented as a solid shape of a single colour.
Normally silhouette edges are represent in black colour to
match the outline of the subject. Foreground detection [8]
methods are used to extract the silhouettes from the video.
The mixture of Gaussian models both background and the
foreground from consecutive frames like {{X1, X2, ... Xt} =
I (x, y, i): 1 ≤ i ≤ t} any time t, at a particular pixel ( x, y).
Model the values of the particular pixel as a mixture of
Gaussians. The probability of observing the current pixel
value is,
(3)
Where, is the weight of the ith
Gaussian in the mixture
at the time of t. µi,t is the mean value of the ith
Gaussian in
the mixture, at time t. is the covariance matrix of the ith
Gaussian in the mixture, at time t. is the Gaussian
probability density function. Weights are updated or
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 232
normalized for every new frame. Every new pixel value
is checked against the existing k Gaussian distributions until
a match is found. The match is defined as a pixel value
within 2.5 standard deviation of a distribution.
Detect which Gaussians may correspond to background
colours based on the persistence and the variance of each of
the Gaussians.
Pixels that do not match with the background pixel value,
Gaussians are grouped using connected components,
represented as the foreground. The extracted silhouette
sequences of frames are integrated with the active contour
model using gradient function.
3.3. Integrated Model
In this section, the resulting sequence is produced by the
integration of scene flow based and silhouette based pose
estimation. The moving object details from both the
methods are integrated based on time by using the Boolean
operator Logical OR. Silhouette extraction is used to extract
the shape of the moving object. It starts with the background
estimation and subtraction. Background subtraction is used
to calculate the foreground moving objects. In video, it
combines the process of extraction and separation of the
foreground objects. Then the video frames are segmented
from the foreground objects.
The input video stream sequence is denoted by I(x,y). Initial
pose of the moving object is denoted as PI,t(x,y) at the time
of t. generally the moving object is mentioned as (obj ti(x,y))
at the time of t and in the spatial cooardinate (x,y).
The output of Scene flow based pose estimation is
mentioned as (obj1 ti(x,y)) at the time t as well as the output
of Silhouette based pose estimation of the moving object is
denoted as (obj ti(x,y)) at the time of t and in the position
(x,y) in fig. 2. In proposed method, without any smoothing
function the evident tracked the object without any noisy
data. It produces more information than the scene flow
based pose estimation like the size and shape of the tracked
object.
1. Start
2. Read the input video as frame by frame I(x, y) .
3. Register the initial pose PI,t(x,y) of the moving
object (obj ti(x,y)) at the time t.
4. Compute the Scene Flow based pose estimation
for the moving object (obj1 ti(x,y))
5. Calculate the silhouette based pose estimation of
the moving object (obj2 ti(x,y))
6. Repeat
a. Integrate (obj1 ti(x,y)) and (obj2 ti(x,y)) over
time using Logical OR operator.
7. Till the end of sequence.
8. End
Fig - 2: Algorithm for Integrated Method
4. RESULTS AND DISCUSSION
The proposed method has been implemented in MATLAB
7.5 tool on Windows 7 Platform. A common video „Denis
walk‟ is taken as the input video. The input video is in „avi‟
format. Proposed experiments and results are carried out as
shown Fig. 3, Fig. 4, Fig. 5, Fig. 6 and Fig. 7
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 233
Fig -3: Sample Frames from the Input video
Fig.-4: Selective frames after Pose Estimation
Fig.-5: Output Frames of Scene Flow based Pose Estimation
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 234
Fig.-6: Output frames of Silhouette based Pose Estimation.
Fig-7: Integrated Output
Table –1: Comparison of Execution Time
Methods Execution time(sec)
Silhouette based pose
Estimation
10.486
Scene flow based pose
Estimation
52.1525
Integrated model 131.799
The execution time of silhouette based pose estimation,
scene flow based pose estimation and integrated model are
presented in Table I. The Integrated system is able to track
the movement of the person more accurately when
compared to the silhouette based pose estimation and scene
flow based pose estimation. Fig 7 represents the output of
the proposed system.
4.1. Comparative Analysis
The evaluation of proposed methods for the moving object
datasets are calculated using the following metrics (4).
(4)
True positive, True negative, false positive and false
negative is denoted as TP, TN, FP, FN respectively.
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 235
Table -2: Comparative Analysis of Methods
Evaluation
Metrics
Silhouette
based Pose
estimation
Scene flow
based pose
estimation
Integrated
Model
Accuracy 24% 51% 52%
The accuracy of the proposed method is compared with the
existing method and the results are shown in Table 2. The
object‟s pixel wise accuracy is calculated between pose
estimation and the silhouette based pose estimation. It
produces 24% of accuracy. The scene flow based accuracy
is calculated between pose estimation and scene flow based
pose estimation. It produces 51% of accuracy, but it is
having some noisy output. The integrated method is again
compared with the pose estimation results by pixel wise.
The proposed Integration model produces 52% of accuracy
without any smoothing or filtering technique.
5. CONCLUSION AND FUTURE WORK
Normally pose estimation methods are mainly used for 3D
reconstruction of 2D objects. Reconstruction of 3D shape
from a single view video is a difficult task because it doesn‟t
provide enough information as much as the multi view
videos. The proposed method provides more exact
information about the shape and size of the objects using
silhouette based pose estimation and scene flow based pose
estimation. It can be applied to multiple moving object
videos and dynamic background outdoor videos for future
enhancement. The performance analysis has concluded that
the integrated system tracks the movement of the person
more accurately when compared to the silhouette based pose
estimation and scene flow based pose estimation. The future
work of this paper is to reconstruct the shape of the object in
3D view.
REFERENCES
[1] S.C.Bilir, Y.Yemez,”Non-rigid 3D shape Tracking
from Multiview video”, Computer Vision and
image Understanding 116(2012)1121-1134.
[2] Ugur Gudukby,Lbrahim Demir,Yigithan
Dedeoglu,
[3] ”Motion Capture and human Pose Reconstruction
from a single-view Video Sequence”, Digital
Signal Processing 23(2013)1441-1450.
[4] ShoheiNobuhara , Takashi Matsuyama
,”Heterogeneous Deformation Model for 3D Shape
and Motion Recovery from Multi-Viewpoint
Images”, Proceedings of the 2nd
International
Symposium on 3D dataProcessing, Visualization
and Transmission(2004).
[5] E. de Aguiar, C. Theobalt, C. Stoll, H.-P. Seidel,
“Marker-less deformable mesh tracking for human
shape and motion capture”, in: Proc. Computer
Vision and Pattern Recognition (CVPR), 2007.
[6] J. Starck, A. Hilton, “Surface capture for
performance-based animation”, IEEE Comput.
Graph. Appl. 27 (3) (2007) 21–31.
[7] T. Matsuyama, X. Wu, T. Takai, S. Nobuhara,
“Real-time 3D shape reconstruction, dynamic 3D
mesh deformation, and high fidelity visualization
for 3D video”, Comput. Vis. Image Understand. 96
(3) (2004) 393–434.
[8] K.Varanasi, A. Zaharescu, E. Boyer, R. Horaud,
“Temporal surface tracking using mesh evolution”,
in: Proc. European Conference on Computer Vision
(ECCV), 2008, pp. 30.
[9] A Alahi, L Bagnato, D Matti,” Foreground
silhouette extraction robust to sudden changes of
background appearance Image Processing (ICIP
2012.
[10] Park, M.Subbarao, “ Pose Estimation and
Integration of Complete 3D Model Reconstruction”
Computer Vision, 2002,(WACV) 2002.
[11] Y Rathi, N Vaswani, A Tannenbaum,,”Particle
Filtering for Geometric Active Contours with
Application to Tracking Moving and Deforming
Objects” Computer Vision, 2005.
[12] Youngbae Hwang, Jun-Sik Kim, Inso Kweon,”
Silhouette Extraction for Visual Hull
Reconstruction “, Conference on Machine Vision
Applications(MVA2005 IAPR) May 16-18, 2005,
Japan.
[13] C.H. Esteban, F. Schmitt, “Silhouette and stereo
fusion for 3D object modeling”, Comput. Vis.
Image Understand. 96 (3) (2004) 367–392.
[14] Y. Yemez, Y. Sahillioglu, “Shape from silhouette
using topology-adaptive mesh deformation”,
Pattern Recognition. Lett. 30 (2009) 1198–1207.
[15] Hunsung Kim,Ryuleki Sakamoto,Itaru
Kitahara,tomji Toriyama and Kiyoshi
Kogure,”Robust Silhouette Extraction Technique
using Background subtraction” (2007) 552-557.
[16] Shih-Chia Huang,” An advanced motion detection
algorithm with video quality analysis for video
surveillance systems”, IEEE Transactions On
Circuits And Systems For Video Technology, Vol.
21, No. 1, January 2011.
[17] C.Sminchisesu,A.Telea, “ Human Pose Estimation
from Silhouettes a Consistent Approach using
Distance Level Sets”, International Conference in
Central Europe on Computer Graphics 2002,
Visualization and Computer Vision.
BIOGRAPHY
Mrs.Hemavathy received M.E degree in
Information Technology from Madras
Institute of Technology, Chennai,
TamilNadu, India in 2013. Presently she is
working Panimalar Engineering
College, Chennai , India.

More Related Content

PDF
An interactive image segmentation using multiple user input’s
PDF
Tracking and counting human in visual surveillance system
PDF
Preliminary study of multi view imaging for accurate 3 d reconstruction using...
PDF
Preliminary study of multi view imaging for accurate
PDF
CHARACTERIZING HUMAN BEHAVIOURS USING STATISTICAL MOTION DESCRIPTOR
PDF
3 video segmentation
PDF
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
PDF
Multi modal medical image fusion using weighted
An interactive image segmentation using multiple user input’s
Tracking and counting human in visual surveillance system
Preliminary study of multi view imaging for accurate 3 d reconstruction using...
Preliminary study of multi view imaging for accurate
CHARACTERIZING HUMAN BEHAVIOURS USING STATISTICAL MOTION DESCRIPTOR
3 video segmentation
Automatic 3D view Generation from a Single 2D Image for both Indoor and Outdo...
Multi modal medical image fusion using weighted

What's hot (14)

PDF
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
PDF
K-Means Clustering in Moving Objects Extraction with Selective Background
PDF
Image compression using sand algorithm
PDF
Simulation of collision avoidance by navigation
PDF
I0343065072
PDF
Development of Human Tracking System For Video Surveillance
PDF
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
PDF
IMAGE RECOGNITION USING MATLAB SIMULINK BLOCKSET
PDF
A novel predicate for active region merging in automatic image segmentation
PDF
Applying edge density based region growing with frame difference for detectin...
PDF
Real time implementation of object tracking through
PDF
Stereo Correspondence Estimation by Two Dimensional Real Time Spiral Search A...
PDF
C0365025029
PDF
HOL, GDCT AND LDCT FOR PEDESTRIAN DETECTION
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
K-Means Clustering in Moving Objects Extraction with Selective Background
Image compression using sand algorithm
Simulation of collision avoidance by navigation
I0343065072
Development of Human Tracking System For Video Surveillance
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
IMAGE RECOGNITION USING MATLAB SIMULINK BLOCKSET
A novel predicate for active region merging in automatic image segmentation
Applying edge density based region growing with frame difference for detectin...
Real time implementation of object tracking through
Stereo Correspondence Estimation by Two Dimensional Real Time Spiral Search A...
C0365025029
HOL, GDCT AND LDCT FOR PEDESTRIAN DETECTION
Ad

Viewers also liked (18)

PDF
Inpainting scheme for text in video a survey
PDF
Services, security challenges and security policies in cloud computing
PDF
A case study on performance evaluation of grid connected pv generation system...
PDF
Application of energy harvesting system in hybrid automobiles
PDF
Standardization of punica granatum explant and callus induction through micro...
PDF
Nutrients retention in functional beef burgers with especial emphasis on lipi...
PDF
Qo s parameters for obs network
PDF
Heat transefer of a dusty fluid over a stretching sheet with internal heat ge...
PDF
Dynamic texture based traffic vehicle monitoring system
PDF
Coin based mobile charger using rfid with pv for public usage
PDF
Study on degradation of fertilizer and pesticides using lysimeter
PDF
A review on various types of software defined radios (sdrs) in radio communic...
PDF
Optimizing the process parameters of friction stir butt welded joint on alumi...
PDF
Automatic room light controller using microcontroller and visitor counter
PDF
Arm based human machine interface of plastic extrusion blow molding system
PDF
Modeling, simulation and optimization analysis of steering knuckle component ...
PDF
A survey on encryption algorithms for data security
PDF
Design and analysis of punching die
Inpainting scheme for text in video a survey
Services, security challenges and security policies in cloud computing
A case study on performance evaluation of grid connected pv generation system...
Application of energy harvesting system in hybrid automobiles
Standardization of punica granatum explant and callus induction through micro...
Nutrients retention in functional beef burgers with especial emphasis on lipi...
Qo s parameters for obs network
Heat transefer of a dusty fluid over a stretching sheet with internal heat ge...
Dynamic texture based traffic vehicle monitoring system
Coin based mobile charger using rfid with pv for public usage
Study on degradation of fertilizer and pesticides using lysimeter
A review on various types of software defined radios (sdrs) in radio communic...
Optimizing the process parameters of friction stir butt welded joint on alumi...
Automatic room light controller using microcontroller and visitor counter
Arm based human machine interface of plastic extrusion blow molding system
Modeling, simulation and optimization analysis of steering knuckle component ...
A survey on encryption algorithms for data security
Design and analysis of punching die
Ad

Similar to Integration of poses to enhance the shape of the object tracking from a single view video (20)

PDF
Development of Human Tracking in Video Surveillance System for Activity Anal...
PDF
D018112429
PDF
A Robust Method for Moving Object Detection Using Modified Statistical Mean M...
PDF
Motion Human Detection & Tracking Based On Background Subtraction
PDF
Survey on video object detection & tracking
PDF
Enhanced target tracking based on mean shift
PDF
Enhanced target tracking based on mean shift algorithm for satellite imagery
DOCX
Motion Object Detection Using BGS Technique
DOCX
Motion Object Detection Using BGS Technique
PDF
3 video segmentation
PDF
Moving object detection using background subtraction algorithm using simulink
PDF
ROBUST STATISTICAL APPROACH FOR EXTRACTION OF MOVING HUMAN SILHOUETTES FROM V...
PDF
Object extraction using edge, motion and saliency information from videos
PDF
Intelligent two axis dual-ccd image-servo shooting platform design
PDF
Intelligent two axis dual-ccd image-servo shooting platform design
PDF
Background differencing algorithm for moving object detection using system ge...
PDF
IRJET- A Review Analysis to Detect an Object in Video Surveillance System
PDF
Stereo Vision Human Motion Detection and Tracking in Uncontrolled Environment
PDF
A STOCHASTIC STATISTICAL APPROACH FOR TRACKING HUMAN ACTIVITY
PDF
Schematic model for analyzing mobility and detection of multiple
Development of Human Tracking in Video Surveillance System for Activity Anal...
D018112429
A Robust Method for Moving Object Detection Using Modified Statistical Mean M...
Motion Human Detection & Tracking Based On Background Subtraction
Survey on video object detection & tracking
Enhanced target tracking based on mean shift
Enhanced target tracking based on mean shift algorithm for satellite imagery
Motion Object Detection Using BGS Technique
Motion Object Detection Using BGS Technique
3 video segmentation
Moving object detection using background subtraction algorithm using simulink
ROBUST STATISTICAL APPROACH FOR EXTRACTION OF MOVING HUMAN SILHOUETTES FROM V...
Object extraction using edge, motion and saliency information from videos
Intelligent two axis dual-ccd image-servo shooting platform design
Intelligent two axis dual-ccd image-servo shooting platform design
Background differencing algorithm for moving object detection using system ge...
IRJET- A Review Analysis to Detect an Object in Video Surveillance System
Stereo Vision Human Motion Detection and Tracking in Uncontrolled Environment
A STOCHASTIC STATISTICAL APPROACH FOR TRACKING HUMAN ACTIVITY
Schematic model for analyzing mobility and detection of multiple

More from eSAT Journals (20)

PDF
Mechanical properties of hybrid fiber reinforced concrete for pavements
PDF
Material management in construction – a case study
PDF
Managing drought short term strategies in semi arid regions a case study
PDF
Life cycle cost analysis of overlay for an urban road in bangalore
PDF
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
PDF
Laboratory investigation of expansive soil stabilized with natural inorganic ...
PDF
Influence of reinforcement on the behavior of hollow concrete block masonry p...
PDF
Influence of compaction energy on soil stabilized with chemical stabilizer
PDF
Geographical information system (gis) for water resources management
PDF
Forest type mapping of bidar forest division, karnataka using geoinformatics ...
PDF
Factors influencing compressive strength of geopolymer concrete
PDF
Experimental investigation on circular hollow steel columns in filled with li...
PDF
Experimental behavior of circular hsscfrc filled steel tubular columns under ...
PDF
Evaluation of punching shear in flat slabs
PDF
Evaluation of performance of intake tower dam for recent earthquake in india
PDF
Evaluation of operational efficiency of urban road network using travel time ...
PDF
Estimation of surface runoff in nallur amanikere watershed using scs cn method
PDF
Estimation of morphometric parameters and runoff using rs & gis techniques
PDF
Effect of variation of plastic hinge length on the results of non linear anal...
PDF
Effect of use of recycled materials on indirect tensile strength of asphalt c...
Mechanical properties of hybrid fiber reinforced concrete for pavements
Material management in construction – a case study
Managing drought short term strategies in semi arid regions a case study
Life cycle cost analysis of overlay for an urban road in bangalore
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
Laboratory investigation of expansive soil stabilized with natural inorganic ...
Influence of reinforcement on the behavior of hollow concrete block masonry p...
Influence of compaction energy on soil stabilized with chemical stabilizer
Geographical information system (gis) for water resources management
Forest type mapping of bidar forest division, karnataka using geoinformatics ...
Factors influencing compressive strength of geopolymer concrete
Experimental investigation on circular hollow steel columns in filled with li...
Experimental behavior of circular hsscfrc filled steel tubular columns under ...
Evaluation of punching shear in flat slabs
Evaluation of performance of intake tower dam for recent earthquake in india
Evaluation of operational efficiency of urban road network using travel time ...
Estimation of surface runoff in nallur amanikere watershed using scs cn method
Estimation of morphometric parameters and runoff using rs & gis techniques
Effect of variation of plastic hinge length on the results of non linear anal...
Effect of use of recycled materials on indirect tensile strength of asphalt c...

Recently uploaded (20)

PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPTX
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
PPT
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PPTX
additive manufacturing of ss316l using mig welding
PPTX
Foundation to blockchain - A guide to Blockchain Tech
PPT
Project quality management in manufacturing
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PPTX
OOP with Java - Java Introduction (Basics)
PPTX
Geodesy 1.pptx...............................................
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PPTX
CH1 Production IntroductoryConcepts.pptx
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PPTX
Lecture Notes Electrical Wiring System Components
R24 SURVEYING LAB MANUAL for civil enggi
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
additive manufacturing of ss316l using mig welding
Foundation to blockchain - A guide to Blockchain Tech
Project quality management in manufacturing
Automation-in-Manufacturing-Chapter-Introduction.pdf
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
OOP with Java - Java Introduction (Basics)
Geodesy 1.pptx...............................................
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
CH1 Production IntroductoryConcepts.pptx
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
bas. eng. economics group 4 presentation 1.pptx
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Embodied AI: Ushering in the Next Era of Intelligent Systems
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
Lecture Notes Electrical Wiring System Components

Integration of poses to enhance the shape of the object tracking from a single view video

  • 1. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 230 INTEGRATION OF POSES TO ENHANCE THE SHAPE OF THE OBJECT TRACKING FROM A SINGLE VIEW VIDEO Hemavathy J1 , Shahul Hameed. C2 , Ani Bernish. S3 1 Asst.Prof, Information Technology, Panimalar Engineering College, Tamil Nadu, India 2 Asst.Prof, Information Technology, Panimalar Engineering College, Tamil Nadu, India 3 Asst.Prof, Information Technology, Panimalar Engineering College, Tamil Nadu, India Abstract In computer vision, tracking human pose has received a growing attention in recent years. The existing methods used multi-view videos and camera calibrations to enhance the shape of the object in 3D view. In this paper, tracking and partial reconstruction of the shape of the object from a single view video is identified. The goal of the proposed integrated method is to detect the movement of a person more accurately in 2D view. The integrated method is a combination of Silhouette based pose estimation and Scene flow based pose estimation. The silhouette based pose estimation is used to enhance the shape of the object for 3D reconstruction and scene flow based pose estimation is used to capture the size as well as the stability of the object. By integrating these two poses, the accurate shape of the object has been calculated from a single view video. Keywords: Pose Estimation, optical Flow, Silhouette, Object Reconstruction, 3D Objects --------------------------------------------------------------------***---------------------------------------------------------------------- 1. INTRODUCTION Object reconstruction in three dimensional views is an emerging topic in computer vision. This is mainly used in computer graphics, animation, biomedical, surveillance and for many other analysis purposes. 3D object reconstruction is the construction of 3D shape of the object from a video, which is having the two dimensional information about the 3D objects in sequence of frames. For object tracking, different types of removable physical markers are used on the moving object in the recorded video. Using the transformation of markers or sensors, the movement or motion trajectory of the object is tracked. They are very expensive and used only to track the object which is already having the sensors or markers in the body. In Adaboost edge detection technique [14], multi-view 3D human pose estimation is still a difficult task in uninhibited environments. Several associated approaches are still dependent on silhouettes of the topic obtained by background subtraction. Background subtraction is complex if the cameras are moving or the background is vibrant. Scene flow [15] is the three dimensional action field of position in the world coordinate; just as optical flow is the two dimensional action field of position in an image. It presents a structure for the calculation of dense, non-rigid scene flow from optical flow. This method does include the knowledge of structure computed independently at the next time instant and also the optical flow which results in less accuracy. The proposed method is marker less and we have used single view camera to capture the movement of a person. To reconstruct the shape of the object from the 2D sequence of images without the camera calibration, estimate the initial pose and its registration is the main task. The rest of paper is organized as follows. We discuss the related work in section II, Section III explains in detail about the proposed work, results and evaluation metrics are discussed in Section IV and Section V provides the conclusion of the paper. 2. RELATED WORKS 3D shape reconstruction is the developed in computer vision field, which is mainly used for visual effects and the analysis purposes. To construct the 3D objects from a 2D video sequence, pose estimation, silhouette extraction and 3D scene flow estimation are the important tasks. The various pose estimation methods are proposed in [1], [2], [3], and [4]. Yogesh Rathi et al. [1] and Jehoon Lee et al. [2] have proposed particle filtering method for pose estimation. [1] Suggested to use particle filtering in static images for segmentation and pose registration problems. But it is producing major occlusion effects between objects. [2] Track the rigid body object in a 2D image sequence and to estimate the pose in 3D space. Different 3D model reconstruction methods are proposed in [3], [4] and [11]. Soon-Yong Park [3] et al comes with the suggestion for automatic three-dimensional model reconstruction technique, which acquire the complete and closed 3D models for real object. Youngbae Hwang et al. [11] developed a method for 3D reconstruction using Silhouette Extraction which suggests three refinements to get the accurate silhouettes of the moving objects. Christoph Vogel et al. [5] has given one novel method for 3D scene flow estimation, its regularizing 3D scene flow computation by penalizing deviations from local rigidity of motion and integrated into an energy minimization framework. Jung-Ho Ahn [6] et al, Matt Olson
  • 2. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 231 et al. [7], Bruce Gooch [8], Alexandre Alahi et al. [10], Youngbae Hwang et al. [11] and Robert T. Collins et al. [12] proposed various methods for silhouette extraction. [6], [8], [10] and [12] implemented the simple background subtraction algorithms to compute the silhouettes of the moving objects. [7] Using octree data structure to capture the silhouettes. 3. PROPOSED WORK Initial pose estimation is the task of registering the shape of the object being tracked. It is used to adjust the position and orientation of the object. The proposed integrated method is used to capture the accurate shape of the object. Depending upon the accurate result from the pose estimation method the 3D shape can be reconstructed from a single view video. In this paper, the proposed method is used to estimate the accurate object from the human motion. First method, Scene Flow based Pose Estimation method is used to capture the scale and stability of the object and the second method, Silhouette based pose registration is used to enhance the shape of the object from the 2D video. By integrating these methods, the shape of the object can be registered more accurately. Fig.1 shows the overview of the proposed method. Initial pose of the moving object is registered first, which is computed by the active contour method. This initial pose will be used for comparison with further movement of the object. 3.1. Scene Flow Based Pose Estimation Method Scene flow is the structure of evident movement of objects, edges and surfaces. Optical flow is the motion points of an object in a two dimensional space coordinates. Scene flow is the motion points of moving object in three dimensional world coordinates. It is used to estimate the motion of moving objects from the series of frames. It is computed between two consecutive frames in all the video frames. Scene flow composes of two components namely Normal Flow and Parallel Flow. The direct method without any constraints is called as normal flow. We have used hierarchical version of Lucas-Kanade optical flow algorithm for parallel flow. This standard algorithm is a predictor for efficient and robust stereo. Optical flow vector is the temporal rate of change of the image plane coordinates (u, v) at a particular point (x, y) is determined by the spatial- temporal variations of the intensity pattern I(x, y). The input frames are shifted to the world coordinates from the special coordinates and the gradient function is computed by using the gradient formula for the input frames. (1) The linear system is updated with its displacement values, before value updating the error frame (T(x, y) - I(x, y)) and Hessian matrix are generating. The transformation of the moving object is then calculated using (2). The two frames are separated by a small time increment ∆t and the movement vector (u, v) obtained by comparing the two consecutive images. Ix(x,y) is the increase in brightness per pixel at position (x, y) in x direction. Iy(x,y) is the increase in brightness per pixel at position (x, y) in y direction. The optical flow is defined as motion field in 2D dimensional view. Scene flow is defined as motion field of 3D dimensional. ....... (2) By combining the nearest neighbour pixel‟s information Lucas-Kanade algorithm solves the optical flow equation. Lucas-Kanade optical flow algorithm divides the frame or image into smaller blocks and assuming the constant velocity in each block. Fig-1.: Architectural Diagram of Integrated Method 3.2. Silhouette based Pose Estimation Method A silhouette may be the image of a person, an object or a scene and it represented as a solid shape of a single colour. Normally silhouette edges are represent in black colour to match the outline of the subject. Foreground detection [8] methods are used to extract the silhouettes from the video. The mixture of Gaussian models both background and the foreground from consecutive frames like {{X1, X2, ... Xt} = I (x, y, i): 1 ≤ i ≤ t} any time t, at a particular pixel ( x, y). Model the values of the particular pixel as a mixture of Gaussians. The probability of observing the current pixel value is, (3) Where, is the weight of the ith Gaussian in the mixture at the time of t. µi,t is the mean value of the ith Gaussian in the mixture, at time t. is the covariance matrix of the ith Gaussian in the mixture, at time t. is the Gaussian probability density function. Weights are updated or
  • 3. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 232 normalized for every new frame. Every new pixel value is checked against the existing k Gaussian distributions until a match is found. The match is defined as a pixel value within 2.5 standard deviation of a distribution. Detect which Gaussians may correspond to background colours based on the persistence and the variance of each of the Gaussians. Pixels that do not match with the background pixel value, Gaussians are grouped using connected components, represented as the foreground. The extracted silhouette sequences of frames are integrated with the active contour model using gradient function. 3.3. Integrated Model In this section, the resulting sequence is produced by the integration of scene flow based and silhouette based pose estimation. The moving object details from both the methods are integrated based on time by using the Boolean operator Logical OR. Silhouette extraction is used to extract the shape of the moving object. It starts with the background estimation and subtraction. Background subtraction is used to calculate the foreground moving objects. In video, it combines the process of extraction and separation of the foreground objects. Then the video frames are segmented from the foreground objects. The input video stream sequence is denoted by I(x,y). Initial pose of the moving object is denoted as PI,t(x,y) at the time of t. generally the moving object is mentioned as (obj ti(x,y)) at the time of t and in the spatial cooardinate (x,y). The output of Scene flow based pose estimation is mentioned as (obj1 ti(x,y)) at the time t as well as the output of Silhouette based pose estimation of the moving object is denoted as (obj ti(x,y)) at the time of t and in the position (x,y) in fig. 2. In proposed method, without any smoothing function the evident tracked the object without any noisy data. It produces more information than the scene flow based pose estimation like the size and shape of the tracked object. 1. Start 2. Read the input video as frame by frame I(x, y) . 3. Register the initial pose PI,t(x,y) of the moving object (obj ti(x,y)) at the time t. 4. Compute the Scene Flow based pose estimation for the moving object (obj1 ti(x,y)) 5. Calculate the silhouette based pose estimation of the moving object (obj2 ti(x,y)) 6. Repeat a. Integrate (obj1 ti(x,y)) and (obj2 ti(x,y)) over time using Logical OR operator. 7. Till the end of sequence. 8. End Fig - 2: Algorithm for Integrated Method 4. RESULTS AND DISCUSSION The proposed method has been implemented in MATLAB 7.5 tool on Windows 7 Platform. A common video „Denis walk‟ is taken as the input video. The input video is in „avi‟ format. Proposed experiments and results are carried out as shown Fig. 3, Fig. 4, Fig. 5, Fig. 6 and Fig. 7
  • 4. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 233 Fig -3: Sample Frames from the Input video Fig.-4: Selective frames after Pose Estimation Fig.-5: Output Frames of Scene Flow based Pose Estimation
  • 5. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 234 Fig.-6: Output frames of Silhouette based Pose Estimation. Fig-7: Integrated Output Table –1: Comparison of Execution Time Methods Execution time(sec) Silhouette based pose Estimation 10.486 Scene flow based pose Estimation 52.1525 Integrated model 131.799 The execution time of silhouette based pose estimation, scene flow based pose estimation and integrated model are presented in Table I. The Integrated system is able to track the movement of the person more accurately when compared to the silhouette based pose estimation and scene flow based pose estimation. Fig 7 represents the output of the proposed system. 4.1. Comparative Analysis The evaluation of proposed methods for the moving object datasets are calculated using the following metrics (4). (4) True positive, True negative, false positive and false negative is denoted as TP, TN, FP, FN respectively.
  • 6. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 03 Issue: 12 | Dec-2014, Available @ http://www.ijret.org 235 Table -2: Comparative Analysis of Methods Evaluation Metrics Silhouette based Pose estimation Scene flow based pose estimation Integrated Model Accuracy 24% 51% 52% The accuracy of the proposed method is compared with the existing method and the results are shown in Table 2. The object‟s pixel wise accuracy is calculated between pose estimation and the silhouette based pose estimation. It produces 24% of accuracy. The scene flow based accuracy is calculated between pose estimation and scene flow based pose estimation. It produces 51% of accuracy, but it is having some noisy output. The integrated method is again compared with the pose estimation results by pixel wise. The proposed Integration model produces 52% of accuracy without any smoothing or filtering technique. 5. CONCLUSION AND FUTURE WORK Normally pose estimation methods are mainly used for 3D reconstruction of 2D objects. Reconstruction of 3D shape from a single view video is a difficult task because it doesn‟t provide enough information as much as the multi view videos. The proposed method provides more exact information about the shape and size of the objects using silhouette based pose estimation and scene flow based pose estimation. It can be applied to multiple moving object videos and dynamic background outdoor videos for future enhancement. The performance analysis has concluded that the integrated system tracks the movement of the person more accurately when compared to the silhouette based pose estimation and scene flow based pose estimation. The future work of this paper is to reconstruct the shape of the object in 3D view. REFERENCES [1] S.C.Bilir, Y.Yemez,”Non-rigid 3D shape Tracking from Multiview video”, Computer Vision and image Understanding 116(2012)1121-1134. [2] Ugur Gudukby,Lbrahim Demir,Yigithan Dedeoglu, [3] ”Motion Capture and human Pose Reconstruction from a single-view Video Sequence”, Digital Signal Processing 23(2013)1441-1450. [4] ShoheiNobuhara , Takashi Matsuyama ,”Heterogeneous Deformation Model for 3D Shape and Motion Recovery from Multi-Viewpoint Images”, Proceedings of the 2nd International Symposium on 3D dataProcessing, Visualization and Transmission(2004). [5] E. de Aguiar, C. Theobalt, C. Stoll, H.-P. Seidel, “Marker-less deformable mesh tracking for human shape and motion capture”, in: Proc. Computer Vision and Pattern Recognition (CVPR), 2007. [6] J. Starck, A. Hilton, “Surface capture for performance-based animation”, IEEE Comput. Graph. Appl. 27 (3) (2007) 21–31. [7] T. Matsuyama, X. Wu, T. Takai, S. Nobuhara, “Real-time 3D shape reconstruction, dynamic 3D mesh deformation, and high fidelity visualization for 3D video”, Comput. Vis. Image Understand. 96 (3) (2004) 393–434. [8] K.Varanasi, A. Zaharescu, E. Boyer, R. Horaud, “Temporal surface tracking using mesh evolution”, in: Proc. European Conference on Computer Vision (ECCV), 2008, pp. 30. [9] A Alahi, L Bagnato, D Matti,” Foreground silhouette extraction robust to sudden changes of background appearance Image Processing (ICIP 2012. [10] Park, M.Subbarao, “ Pose Estimation and Integration of Complete 3D Model Reconstruction” Computer Vision, 2002,(WACV) 2002. [11] Y Rathi, N Vaswani, A Tannenbaum,,”Particle Filtering for Geometric Active Contours with Application to Tracking Moving and Deforming Objects” Computer Vision, 2005. [12] Youngbae Hwang, Jun-Sik Kim, Inso Kweon,” Silhouette Extraction for Visual Hull Reconstruction “, Conference on Machine Vision Applications(MVA2005 IAPR) May 16-18, 2005, Japan. [13] C.H. Esteban, F. Schmitt, “Silhouette and stereo fusion for 3D object modeling”, Comput. Vis. Image Understand. 96 (3) (2004) 367–392. [14] Y. Yemez, Y. Sahillioglu, “Shape from silhouette using topology-adaptive mesh deformation”, Pattern Recognition. Lett. 30 (2009) 1198–1207. [15] Hunsung Kim,Ryuleki Sakamoto,Itaru Kitahara,tomji Toriyama and Kiyoshi Kogure,”Robust Silhouette Extraction Technique using Background subtraction” (2007) 552-557. [16] Shih-Chia Huang,” An advanced motion detection algorithm with video quality analysis for video surveillance systems”, IEEE Transactions On Circuits And Systems For Video Technology, Vol. 21, No. 1, January 2011. [17] C.Sminchisesu,A.Telea, “ Human Pose Estimation from Silhouettes a Consistent Approach using Distance Level Sets”, International Conference in Central Europe on Computer Graphics 2002, Visualization and Computer Vision. BIOGRAPHY Mrs.Hemavathy received M.E degree in Information Technology from Madras Institute of Technology, Chennai, TamilNadu, India in 2013. Presently she is working Panimalar Engineering College, Chennai , India.