SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 03 | Mar 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 267
Human Pose Detection using Deep Learning
Deepak Maurya1, Grandel Dsouza2, Anoop Patel3
1,2,3 Student, Dept. of Information Technology, Xavier Institute of Engineering, Mahim, Mumbai,
Maharashtra, India.
---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract - Human pose detection is a crucial problem within
the field of Computer Vision. Imagine having the ability to
trace a person’s every small movement and do a
biomechanical analysis in real-time. This technology has a
huge implication both now and in the future. Applications
based on Human pose detection may including video
surveillance, assisted living, healthy lifestyle and sports
analysis. Formally speaking, Pose Estimation or detection is
predicting the part or joint positions of an individual from a
picture or a video. This technology is made possible by
combining two major computer technologies i.e. Artificial
Intelligence (A.I.) and Computer Graphics
Key Words: Human pose detection, Computer vision,
biomechanical analysis, video surveillance, assisted living,
sport analysis, Artificial intelligence, computer graphics.
1. INTRODUCTION
Human body pose estimation or detection in computer
vision/graphics is that the study of algorithms, systems,
and pre-trained models that recover the pose of an
articulated body, which consists of joints and rigid parts
using image-based observations. it's one of the longest-
lasting prevalent problems in computer vision the reason
being the complexity of the models that relate observation
with the pose, and since of the variability of situations
during which it'd beuseful.
Perception of the citizens in their neighbouring
environment is a crucial capability that robots must
possess. If an individual uses gestures to point to a specific
object, then the interacting machine should be ready to
understand things in the world context. Thus, pose
estimation is a crucial and challenging problem in computer
vision, and lots of algorithms are deployed in solving this
problem over the last 25 years. Many solutions involve
training complex models with large sample data sets and
hence improvising and progressing further.
Pose estimation may be a difficult problem and a lively
subject of research because the physical body has 244
degrees of freedom with 230 joints. Although not all
movements between joints are evident, the physical body
consists of 10 large parts with 20 degrees of freedom.
Algorithms must account for giant variability introduced by
differences in appearance thanks to clothing, body shape,
size, and hairstyles. Additionally, the results could also be
ambiguous thanks to partial occlusions from self-
articulation, like an individual's hand covering their face,
or occlusions from external objects. Finally, most
algorithms estimate pose from monocular (two-
dimensional) images, taken from a traditional camera.
Other issues include varying lighting and camera
configurations. The challenges are made even more complex
to match up to additional performance requirements. These
images lack the three-dimensional (3-D) information of an
actual body pose, resulting in further ambiguities. there's
recent add this area wherein images from RGBD cameras
provide information about colour and depth.
2. How exactly did we achieve human pose detection
using deep learning?
Firstly, to implement Human pose detection a computer
system must precisely distinguish between a human being
and the surrounding objects and the system must recognize
the different body parts and differ each one from the other
without any faults. This is where deep learning comes into
the picture. What exactly is done is that the computer system
is trained with different sample data in this context the
system will be trained with different images like head,
shoulders, arms etc. The training is performed using the
Neural network concept of machine learning like CNN
(Convolutional neural network) and Deep learning. Thus,
different images of the different body parts and at different
angles are fed on to the system and thus a trained model is
achieved. Here it must be noted that the more sample data set
the system is trained with the better performance the system
shall yield. Also every body part here in the trained model
will have a unique identity co-ordinate. This trained model is
termed as COCO (common objects in context) model and MPII
model. Functionalities of these models are: a) Image
classification b) Object localization c) Semantic segmentation
d) Segmenting individual object instances.
Upon achieving the above models, we now have an
intelligent system that can identify and distinguish a human
body from the surrounding. Now is where the computer
graphics comes into picture. For computer graphics we have
used OpenCV which we have used as a python library for
image processing as well as video processing. Also, we have
use Flask for designing a website (i.e. for Graphical User
Interface)
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 03 | Mar 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 268
3. Implementation Methodology
We approached the project in seven phases which are
mentioned below:
1. Firstly, we downloaded the model weights
2. Then we loaded it into the network
3. Read video as a set of frames which we took as
an input to the network
4. Predictions were made and key points were
parsed
5. By using the key points, the skeletal
structure was drawn
6. The skeletal structure gave the various angles that
were being made
7. Finally using the angles, we got a result wherein
certain instructions were given and then we plotted
the graph
4. Details of Hardware and Software
Hardware:
a. CPU: 4 core processor
b. RAM: 4 GB
Software:
i. Python 3.6
ii. OpenCV
iii. Visual Studio Code
iv. Operating System: Windows 10
5. Results
Figure-1: Skeletal structure of a human being and
instructions corresponding to it
Here we can see the skeletal structure of a human being.
Whenever a human being is doing certain exercises, there
are predefined angles at which the movements should take
place. Based on one’s action, one would get the
instructions. If one is not doing the movements properly,
then one would be cautioned about that so that he can
make the required improvements.
Figure-2: Graph describing the movements in the skeletal
structure
6. Acknowledgement
We express our sincere gratitude to our Honorable
Principal of Xavier institute of Engineering Mahim, Mumbai
Dr. Y.D.Venkatesh for encouragement and facilities
provided to us. We would like to place on record our deep
sense of gratitude to Prof. Chhaya Narvekar, Head of Dept
Of Information Technology, Xavier Institute of Engineering,
Mahim, Mumbai, for her generous guidance help and useful
suggestions. We would also like to thank our entire
Information Technology staff who have willingly
cooperated with us in resolving our queries and providing
us all the required facilities on time.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 03 | Mar 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 269
7. Conclusion
We designed this project with an aim in mind to develop a low-cost software model for identifying the objects and so that
we can carry out estimation of motion analysis of the objects and after trying various approaches we have successfully
accomplished it.
8. References
[1] A survey on Model based approaches for 2D and3D Visual Human Pose Recovery Xavier Perez-Sala 1, Sergio Escalera 2,
Cecilio Angulo 3 and Jordi Gonz`alez
[2] Integrated recognition, localization, and detection using convolutional networks. L. Pishchulin, M. Andriluka, P.
Gehler, and B. Schiele
[3] 2017 IEEE 12th International Conference on Automatic Face & Gesture Recognition Realtime Multiperson 2D Pose
Estimation using Part Affinity Fields Zhe Cao Tomas Simon Shih-En Wei Yaser Sheikh, The Robotics Institute, Carnegie
Mellon University
[4] Articulated bodyposeestimation- Wikipedia.html
[5]https://www.learnopencv.com/de ep-learning-based-human-pose- estimation- using -opencv -cpp- python/

More Related Content

PDF
Saksham seminar report
PDF
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
PPTX
Saksham presentation
PDF
IRJET- A Deep Learning based Approach for Automatic Detection of Bike Rid...
PDF
IRJET- Sign Language Interpreter
PDF
IRJET- Tracking and Recognition of Multiple Human and Non-Human Activites
PDF
IRJET- IoT based Facial Recognition Biometric Attendance
PDF
IRJET- Survey Paper on Vision based Hand Gesture Recognition
Saksham seminar report
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
Saksham presentation
IRJET- A Deep Learning based Approach for Automatic Detection of Bike Rid...
IRJET- Sign Language Interpreter
IRJET- Tracking and Recognition of Multiple Human and Non-Human Activites
IRJET- IoT based Facial Recognition Biometric Attendance
IRJET- Survey Paper on Vision based Hand Gesture Recognition

What's hot (20)

PDF
IRJET- Confidential Data Access through Deep Learning Iris Biometrics
PDF
A COMPARATIVE STUDY ON HUMAN ACTION RECOGNITION USING MULTIPLE SKELETAL FEATU...
PDF
IRJET - Autonomous Navigation System using Deep Learning
PDF
IRJET- Development of a Face Recognition System with Deep Learning and Py...
PDF
IRJET- Face Recognition using Deep Learning
PDF
A Simulation Method of Soft Tissue Cutting In Virtual Environment with Haptics
PDF
Vehicle Monitoring System based On IOT, Using 4G/LTE
PDF
IRJET- Classification of Assembly (W-Section) using Artificial Intelligence
PDF
IRJET- Spot Me - A Smart Attendance System based on Face Recognition
PDF
IRJET- Survey on Face Recognition using Biometrics
PDF
An Efficient VLSI Design of AES Cryptography Based on DNA TRNG Design
PDF
IRJET- Navigation and Camera Reading System for Visually Impaired
PDF
IRJET - Automatic Attendance Provision using Image Processing
DOCX
Multiple object detection report
DOC
Augmented Reality for Robotic Surgical Dissection - Final Report
PDF
IRJET- Deep Feature Fusion for Iris Biometrics on Mobile Devices
PPTX
Automatic Attendance System using Deep Learning
PDF
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
PDF
IRJET - Chatbot with Gesture based User Input
PDF
A survey paper on various biometric security system methods
IRJET- Confidential Data Access through Deep Learning Iris Biometrics
A COMPARATIVE STUDY ON HUMAN ACTION RECOGNITION USING MULTIPLE SKELETAL FEATU...
IRJET - Autonomous Navigation System using Deep Learning
IRJET- Development of a Face Recognition System with Deep Learning and Py...
IRJET- Face Recognition using Deep Learning
A Simulation Method of Soft Tissue Cutting In Virtual Environment with Haptics
Vehicle Monitoring System based On IOT, Using 4G/LTE
IRJET- Classification of Assembly (W-Section) using Artificial Intelligence
IRJET- Spot Me - A Smart Attendance System based on Face Recognition
IRJET- Survey on Face Recognition using Biometrics
An Efficient VLSI Design of AES Cryptography Based on DNA TRNG Design
IRJET- Navigation and Camera Reading System for Visually Impaired
IRJET - Automatic Attendance Provision using Image Processing
Multiple object detection report
Augmented Reality for Robotic Surgical Dissection - Final Report
IRJET- Deep Feature Fusion for Iris Biometrics on Mobile Devices
Automatic Attendance System using Deep Learning
IRJET - Creating a Security Alert for the Care Takers Implementing a Vast Dee...
IRJET - Chatbot with Gesture based User Input
A survey paper on various biometric security system methods
Ad

Similar to Human pose detection using machine learning by Grandel (20)

PDF
HUMAN IDENTIFIER WITH MANNERISM USING DEEP LEARNING
PDF
A SURVEY ON HUMAN POSE ESTIMATION AND CLASSIFICATION
PPTX
Human Pose estimation project for computer vision
PDF
”YOGA WITH AI”
PDF
hpe3d_report.pdf
PPTX
Human Pose Estimation by Deep Learning
PDF
Dance With AI – An interactive dance learning platform
PDF
AI Personal Trainer Using Open CV and Media Pipe
PPTX
[Mmlab seminar 2016] deep learning for human pose estimation
PDF
Study of AI Fitness Model Using Deep Learning
PDF
AI Personal Trainer Using Open CV and Media Pipe
PDF
Yoga Posture Classification using Computer Vision
PDF
Dragos_Papava_dissertation
PDF
Human Behavior Understanding: From Human-Oriented Analysis to Action Recognit...
PPTX
Towards Accurate Multi-person Pose Estimation in the Wild (My summery)
PPTX
IATMSI 2022 Presentation Format.pptx
PDF
Pose Trainer: “An Exercise Guide and Assessment in Physiotherapy”
PDF
Virtual Yoga System Using Kinect Sensor
PPTX
Deep learning-for-pose-estimation-wyang-defense
PPTX
Human pose estimation with deep learning
HUMAN IDENTIFIER WITH MANNERISM USING DEEP LEARNING
A SURVEY ON HUMAN POSE ESTIMATION AND CLASSIFICATION
Human Pose estimation project for computer vision
”YOGA WITH AI”
hpe3d_report.pdf
Human Pose Estimation by Deep Learning
Dance With AI – An interactive dance learning platform
AI Personal Trainer Using Open CV and Media Pipe
[Mmlab seminar 2016] deep learning for human pose estimation
Study of AI Fitness Model Using Deep Learning
AI Personal Trainer Using Open CV and Media Pipe
Yoga Posture Classification using Computer Vision
Dragos_Papava_dissertation
Human Behavior Understanding: From Human-Oriented Analysis to Action Recognit...
Towards Accurate Multi-person Pose Estimation in the Wild (My summery)
IATMSI 2022 Presentation Format.pptx
Pose Trainer: “An Exercise Guide and Assessment in Physiotherapy”
Virtual Yoga System Using Kinect Sensor
Deep learning-for-pose-estimation-wyang-defense
Human pose estimation with deep learning
Ad

Recently uploaded (20)

PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PPTX
sap open course for s4hana steps from ECC to s4
PDF
Electronic commerce courselecture one. Pdf
PPTX
Spectroscopy.pptx food analysis technology
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
cuic standard and advanced reporting.pdf
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PPTX
Cloud computing and distributed systems.
PDF
Spectral efficient network and resource selection model in 5G networks
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
KodekX | Application Modernization Development
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Encapsulation theory and applications.pdf
PDF
NewMind AI Weekly Chronicles - August'25 Week I
Diabetes mellitus diagnosis method based random forest with bat algorithm
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
Network Security Unit 5.pdf for BCA BBA.
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
sap open course for s4hana steps from ECC to s4
Electronic commerce courselecture one. Pdf
Spectroscopy.pptx food analysis technology
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Mobile App Security Testing_ A Comprehensive Guide.pdf
cuic standard and advanced reporting.pdf
MIND Revenue Release Quarter 2 2025 Press Release
Cloud computing and distributed systems.
Spectral efficient network and resource selection model in 5G networks
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
KodekX | Application Modernization Development
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
Digital-Transformation-Roadmap-for-Companies.pptx
Unlocking AI with Model Context Protocol (MCP)
Encapsulation theory and applications.pdf
NewMind AI Weekly Chronicles - August'25 Week I

Human pose detection using machine learning by Grandel

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 03 | Mar 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 267 Human Pose Detection using Deep Learning Deepak Maurya1, Grandel Dsouza2, Anoop Patel3 1,2,3 Student, Dept. of Information Technology, Xavier Institute of Engineering, Mahim, Mumbai, Maharashtra, India. ---------------------------------------------------------------------***---------------------------------------------------------------------- Abstract - Human pose detection is a crucial problem within the field of Computer Vision. Imagine having the ability to trace a person’s every small movement and do a biomechanical analysis in real-time. This technology has a huge implication both now and in the future. Applications based on Human pose detection may including video surveillance, assisted living, healthy lifestyle and sports analysis. Formally speaking, Pose Estimation or detection is predicting the part or joint positions of an individual from a picture or a video. This technology is made possible by combining two major computer technologies i.e. Artificial Intelligence (A.I.) and Computer Graphics Key Words: Human pose detection, Computer vision, biomechanical analysis, video surveillance, assisted living, sport analysis, Artificial intelligence, computer graphics. 1. INTRODUCTION Human body pose estimation or detection in computer vision/graphics is that the study of algorithms, systems, and pre-trained models that recover the pose of an articulated body, which consists of joints and rigid parts using image-based observations. it's one of the longest- lasting prevalent problems in computer vision the reason being the complexity of the models that relate observation with the pose, and since of the variability of situations during which it'd beuseful. Perception of the citizens in their neighbouring environment is a crucial capability that robots must possess. If an individual uses gestures to point to a specific object, then the interacting machine should be ready to understand things in the world context. Thus, pose estimation is a crucial and challenging problem in computer vision, and lots of algorithms are deployed in solving this problem over the last 25 years. Many solutions involve training complex models with large sample data sets and hence improvising and progressing further. Pose estimation may be a difficult problem and a lively subject of research because the physical body has 244 degrees of freedom with 230 joints. Although not all movements between joints are evident, the physical body consists of 10 large parts with 20 degrees of freedom. Algorithms must account for giant variability introduced by differences in appearance thanks to clothing, body shape, size, and hairstyles. Additionally, the results could also be ambiguous thanks to partial occlusions from self- articulation, like an individual's hand covering their face, or occlusions from external objects. Finally, most algorithms estimate pose from monocular (two- dimensional) images, taken from a traditional camera. Other issues include varying lighting and camera configurations. The challenges are made even more complex to match up to additional performance requirements. These images lack the three-dimensional (3-D) information of an actual body pose, resulting in further ambiguities. there's recent add this area wherein images from RGBD cameras provide information about colour and depth. 2. How exactly did we achieve human pose detection using deep learning? Firstly, to implement Human pose detection a computer system must precisely distinguish between a human being and the surrounding objects and the system must recognize the different body parts and differ each one from the other without any faults. This is where deep learning comes into the picture. What exactly is done is that the computer system is trained with different sample data in this context the system will be trained with different images like head, shoulders, arms etc. The training is performed using the Neural network concept of machine learning like CNN (Convolutional neural network) and Deep learning. Thus, different images of the different body parts and at different angles are fed on to the system and thus a trained model is achieved. Here it must be noted that the more sample data set the system is trained with the better performance the system shall yield. Also every body part here in the trained model will have a unique identity co-ordinate. This trained model is termed as COCO (common objects in context) model and MPII model. Functionalities of these models are: a) Image classification b) Object localization c) Semantic segmentation d) Segmenting individual object instances. Upon achieving the above models, we now have an intelligent system that can identify and distinguish a human body from the surrounding. Now is where the computer graphics comes into picture. For computer graphics we have used OpenCV which we have used as a python library for image processing as well as video processing. Also, we have use Flask for designing a website (i.e. for Graphical User Interface)
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 03 | Mar 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 268 3. Implementation Methodology We approached the project in seven phases which are mentioned below: 1. Firstly, we downloaded the model weights 2. Then we loaded it into the network 3. Read video as a set of frames which we took as an input to the network 4. Predictions were made and key points were parsed 5. By using the key points, the skeletal structure was drawn 6. The skeletal structure gave the various angles that were being made 7. Finally using the angles, we got a result wherein certain instructions were given and then we plotted the graph 4. Details of Hardware and Software Hardware: a. CPU: 4 core processor b. RAM: 4 GB Software: i. Python 3.6 ii. OpenCV iii. Visual Studio Code iv. Operating System: Windows 10 5. Results Figure-1: Skeletal structure of a human being and instructions corresponding to it Here we can see the skeletal structure of a human being. Whenever a human being is doing certain exercises, there are predefined angles at which the movements should take place. Based on one’s action, one would get the instructions. If one is not doing the movements properly, then one would be cautioned about that so that he can make the required improvements. Figure-2: Graph describing the movements in the skeletal structure 6. Acknowledgement We express our sincere gratitude to our Honorable Principal of Xavier institute of Engineering Mahim, Mumbai Dr. Y.D.Venkatesh for encouragement and facilities provided to us. We would like to place on record our deep sense of gratitude to Prof. Chhaya Narvekar, Head of Dept Of Information Technology, Xavier Institute of Engineering, Mahim, Mumbai, for her generous guidance help and useful suggestions. We would also like to thank our entire Information Technology staff who have willingly cooperated with us in resolving our queries and providing us all the required facilities on time.
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 03 | Mar 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 269 7. Conclusion We designed this project with an aim in mind to develop a low-cost software model for identifying the objects and so that we can carry out estimation of motion analysis of the objects and after trying various approaches we have successfully accomplished it. 8. References [1] A survey on Model based approaches for 2D and3D Visual Human Pose Recovery Xavier Perez-Sala 1, Sergio Escalera 2, Cecilio Angulo 3 and Jordi Gonz`alez [2] Integrated recognition, localization, and detection using convolutional networks. L. Pishchulin, M. Andriluka, P. Gehler, and B. Schiele [3] 2017 IEEE 12th International Conference on Automatic Face & Gesture Recognition Realtime Multiperson 2D Pose Estimation using Part Affinity Fields Zhe Cao Tomas Simon Shih-En Wei Yaser Sheikh, The Robotics Institute, Carnegie Mellon University [4] Articulated bodyposeestimation- Wikipedia.html [5]https://www.learnopencv.com/de ep-learning-based-human-pose- estimation- using -opencv -cpp- python/