SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 509
DESIGN AND DEVELOPMENT OF TESSERACT-OCR BASED ASSISTIVE
SYSTEM TO CONVERT CAPTURED TEXT INTO VOICE OUTPUT
G.ELUMALAI1,J.SUNDAR RAJAN2, P.SURYA PRAKASAH3, V.L.SUSRUTH4,P.K.SUDHARSANAN5
1 Associate Professor, Dept. of Electronics and Communication Engineering, Panimalar Engineering College,
Tamil Nadu, India
2,3,4,5 UG students, Dept. of Electronics and Communication Engineering, Panimalar Engineering College,
Tamil Nadu, India
---------------------------------------------------------------------***---------------------------------------------------------------------
Abstract - The biggest challenge faced by the blind people,
is their inability to view real life objects and to read. The only
efficient system that exists so far, is the braille system, that
enables the blind to read. This system is time consuming and
the time taken to recognize the text is long. Our aim here is to
reduce the time taken to read. In our work, using a Raspberry
Pi, we have designed a smart reader, so that the blind people
may read. The module that we have designed, either uses a
webcam or a mobile camera that is linked with a RaspberryPi,
to focus on a range of printed text. The OCR (OpticalCharacter
Recognition) package installed in raspberry pi tests it into a
digital article which is then subjected to skew modification,
segmentation, before feature extraction to perform sorting.
Our proposed project automatically focuses the regionsoftext
in the object, after which the text characters are thenlocalized
using a localization algorithm, that uses feature recognition
and edge pixel distribution using artificialneuralnetwork.The
text characters are then binarized into a machine readable
form by using an OCR algorithm called as Tesseract. The
recognized characters of the text are then converted into
audio format, so that they may be recognized by the blind
users.
Key Words: Embedded devices, Image processing, OCR,
Neural Networks, Raspberry Pi.
1. INTRODUCTION
The importance of reading in our society, today, cannot be
ignored. We could find printed text everywhere in the form
of medical prescription, newspapers, books, handouts,
medicine bottles and tablets and the like. While video
magnifiers, optical aids, and screen readers can help with
low vision and blind users to access documents, few devices
do exist that provide optimum access to daily-life hand held
objects such as fruit juice cartons and invention packages.
The most important causes for visual degradation include
diabetic retinopathy, aging factor, eye diseases such as
ocular tumors and accidents, which lead to an increaseinthe
number of visually disabled people, nowadays. Cataract is
leading cause of blindness and visual impairment. Mobile
applications are available these days, thataidinvisualization
and they form an integral part of the blind peoples lives.
Recent advances in mobile technology, digital camera,
computer vision and camera-based application make it
possible to support visually challenged persons by
developing camera-based application that combine
computer vision with other existing technology such as
optical character recognition (OCR) system. To detect text
information from image, practical difficulties exist, such as
non-uniform backgrounds, due to the large variations in
character font, size, texture, color, background, orientations,
and many other reasons. Text detection from scene/text
camera images is possible due to high resolution camera.
2. PROJECT OBJECTIVE
The biggest challenge faced by the blind peopletodayistheir
inability to read and recognize the real-world objects. The
only efficient method that exists so far is the braille method,
which enables them to read. It is somewhat time consuming
to read the text using a braille and also it might be not be
economically feasible. The objective of the project is to
create a portable, light-weight module that enablestheblind
to read. This module uses a webcam to capture the image of
the focused text, extract separate text characters from the
scene and then provide a speech output using the Tesseract
OCR.
3. LITERATURE SURVEY
In the field of Text To Speech conversion, researchisfocused
on the identification of the input text and the ability to
express it in a natural manner. Sentimental analysis
procedure(positive/neutral/negative) is used in order to
input features for expressive speech synthesis, where
identification of the input text is a front end task. Different
combinations of classifiers and textualfeaturesareevaluated
for the determination of the most appropriate adaptation
procedure. Semeval 2007 dataset and Twitter corpus are
used for the evaluation of the effectivenessof thisscheme of
Sentiment Analysis, which is appropriate for an expressive
Text To Speech scenario. Conducted experiments validate
the procedure proposed with respect to Sentiment Analysis
[1]. MD-TTS TC not only considers not only the contents of
the text but also its structure, in contrast to the topic of text
classification tasks. The proposal is validated by the
experiments that were conducted in terms of both
subjective(synthetic speech quality) and objective(TC
efficiency) evaluation criteria. [2] For the visually disabled
people, travel aids have been developed in the recent years
in the form of stereovision-capable devices. They are
however, bulky and assist the blind person in avoiding
obstacles only and they are head-mounted. We researched
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 510
about a wearable system that can not only assist in avoiding
obstacles, but at the same time supports live streaming of
videos over the 3G network. The module is comprised of an
eye-glass and an embedded device that is power efficient.
Two miniature camerasare used for imaging, on one end,by
the eye-glass. For the combination of stereo images, FPGA
and FIFO buffers are used. Also, live stereo matching on an
embedded processor is achieved by means of parallel
processing. Live streaming over 3G network, of the video
that is captured by the system is also supported. A person
with healthy eyesight can guide a visually impaired person,
with the help of this livevideo streaming [3]. We researched
about a system that helps the blind people to read from the
labelsand product packages of common handhelddevicesin
their daily lives, by means of a camera. The Region Of
Interest(ROI) in the camera view is obtained by using an
innovative motion-based approach, where the users were
asked to shake the object. This is done in order to isolate the
intended image from cluttered backgrounds and other
objects in the camera view. A mixture of Gaussian based,
background subtraction methodswere used by thismethod.
Text information is obtained from the extracted ROI, by a
combination of localization and recognition. Feature
extraction and the distribution of edge pixels are studied,
using Adaboost model, to automatically localize the text
characters in the ROI. Localized text characters are then
identified by using the OCR algorithms. Recognized text is
conveyed to the blind users using audio output. The
effectiveness of the system hardware is evaluated by using
the dataset that is obtained from ten blind people[4]. In TTS
technology, the ability is provided to a computer to speak. A
synthesizer is used to convert the text into speech, thus
making it eligible for further processing. By learningvarious
textbooks related to languages, language mastery was
obtained, traditionally. Due to the implied cost factor,
moving along this method was very difficult. Various
dictionaries were used by the existing system, to pronounce
the identified text with correct pronunciation . Only for a set
of words which are available in the dictionary, the operation
is supported. [5] Studies in blind people have shown the
activation of cortical areas that sub serve the vision, but it
has long been controversial whether blind people have
tactile acuity. We compared the passivetactileacuityofblind
and sighted subjects on a fully automatedgratingorientation
task and used multivariate Bayesian data analysis to
determine predictors of acuity. Irrespective oftheamountof
childhood vision, braille reading or light perception level,
acuity was superior in blind people. Acuity is better in
women than in men and it also decline with age. Acuity is
chiefly dependent on the force of contact between the
stimulus surface and the subject . The difference between
blind people and healthy sighted people varied a lot in spite
of the intragroup variation: the acuity of both sighted and
blind people of the same gender remained the same, except
that the acuity of a blind person was similar to that of a
sighted person who was 23 years younger. The results
suggest that cross modal plasticity may underlie tactile
acuity enhancement in blindness. [6]
4. EXISTING SYSTEM
A few systems already exist that help the blind people have
access to daily-life hand held objects, but they are largely
ineffective with respect to the focusarea of the labelling.For
example, portable bar-code readers do exist, but it is very
difficult for a blind person to locate a bar codeandthenpoint
the laser beam on that bar code. Also, to enable the blind
people to read, braille system exists. In spite of ignoring the
manufacturing cost, the system can be said to be largely
ineffective with regard to the processing speed, that is, the
speed at which the blind subject can identify theinformation
and then assimilate it is too low.
5. PROPOSED SYSTEM
The proposed system uses a Raspberry Pi board, an
ultrasonic sensor and a webcam to recognize the text in the
scene. In this, the webcam is focussed on the scene. A video
streaming is obtained, from which the images are captured
frame by frame. The images are refined in order to eliminate
any noise that is present in it. A feature called segmentation
is used in order to separate each character from other in the
text. Graphical details such as icons or logos, if any, are
eliminated. Each obtained character are compared with the
datasets that are created as a part of the Tesseract library.
The Tesseract OCR is the most efficient algorithm available
that checks for the obtained character in ten dimensions.
Once, the character is recognized, it must be made available
as an audio output. For this, we use a software calledfestival.
The festival is used to provide the audio output for the
recognized character. Apart from these features, an extra
feature is added, that enables the blind to know the type of
object that he/she interacts with.(a menu, newspaper and
the like). An ultrasonic sensor is included as a part of the
project, that makes the project obtain charactersonlywithin
a particular distance.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 511
Fig-1: Block diagram of the proposed system
6. SYSTEM DESIGN SPECIFICATION
6.1. Hardware specification
1. RASPBERRY PI:The RaspberryPi,usingtheARM1176JZF-S
core, is a credit card sized single computer on chip or SoC.
System on chip is a method by which all the electronic
components that are necessary to run a computer systemare
placed on a single chip. In order to start up, the Raspberry Pi
needsan OS.On board non-volatilememory,whichstoresthe
bootloaders are eliminated in the Raspberry Pi, with the aim
of cost reduction. Traditional embedded systems use Linux
kernels and file system. For this reason, an SD/MMCcardslot
is provided. Raspberry Pi will be executed according to the
application program and after the completion of successful
boot loading.
2. ULTRASONIC SENSORS: Ultrasonic sensors provide cost
effective, unique sensingmethods, which arenotprovidedby
any other technology, thus serving the market at large.
Application problems that are eithercostprohibitiveorthose
that couldn’t be solved using other sensors can be solved by
an ultrasonic sensor, by applying a wide variety of ultrasonic
transducers and a wide variety of frequency ranges.
Detection over distance are required by a variety of
applications in industrial sensing. The detection range of
ultrasonic sensors is quite high, as high as up to forty feet,
which limit switches and inductive switching cannot do.
Photo electric sensorscandetectlong distances,buttheylack
a wide area coverage or a wide area coverage which can be
obtained by using a numberof sensors. Migatron’sultrasonic
sensor cover both narrow and wide areas. Proper ultrasonic
transducer selection is what it all takes. Since there are wide
range of target materials available, only ultrasonic sensors
are independent of the target material composition . The
target material may be solid, liquid, porous, soft or woody in
nature and irrespective of its target composition, it can still
be detected. Distance measurement is easy and it can be
calculated by using the time from which theultrasonic signal
leavesthe signaltothe time theultrasonicreturnsbacktothe
sensor. This method is accurate to .05% of range which
equates to +or- .002 of an inch at a distance of 4 inches.
6.2. Software specifications:
1. OPEN CV: OpenCV (Open Source Computer VisionLibrary)
is a collectionof library functions, whose main objective is to
enable computer vision. The main objective of OpenCV is to
provide a framework forcomputer vision and to improvethe
readabilityofmachinesover commercialproducts. Sinceitis
BSD licensed, businesses can easily use and modify the code.
Over2500 sophisticated algorithmsarepresentinthelibrary,
which include both machine learning and computer vision
algorithms. Using the algorithms, the computer can identify
objects, identify faces, recognize moving objects, recognize
human actions, extract 3D models of objects, combine
different images to produce an entire image scene, red eye
removal, scenery recognition and augmented reality etc. can
be performed. Around 47 thousandmemberusercommunity
for OpenCV exists and OpenCV is widely used by many
commercial bodies, governmental institution and research
organizations.
2. NEURAL-NETWORKS: Neural networksare a collectionof
hardware and software modules that are patterned around
the neural networks of the human brain, in technical terms
of information technology. They are a variety of deep
learning technology that are also termed as artificial neural
networks. Complex signal processing and complex patter
recognition tasks are some of the commercialapplicationsof
the neural networks. Image processing and text recognition,
handwriting recognition, weather prediction and facial
recognition are some examples of the applications of neural
networks since 2000. A neural network usually involves a
large number of parallel operating processors arranged in
tiers. Raw input information are received in the first tier –
analogous to optic nerves in human visual processing.
Instead of receiving the output from the raw input, each tier
receives the output from the tier that precedes it. The last
tier produces the output of the system. Every node that
processes possess its own sphere of knowledge, including
what it was originally programmed with, or what is has
learnt or those set of rules which it has created for itself.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 512
3. OCR: Optical Character Recognition is the conversion of
either printed or handwritten characters into a machine
readable format(binary 0s and 1s). The text of concern may
be obtained either from an imageof a document,orascanned
copyof the documentor from areallife scene such asthetext
on signs and billboards(Fig.6 and Fig.7) or from subtitle text
underneath an image (such asthose thatoccur ina television
broadcast). It is used to enter the information from different
types of documents such as records, passports, cheques,
invoice, bills or any other such suitable type of
documentation. OCR is a widely used method for converting
the printed data into digital format, so that they may be
edited, searched , compressed, used in line display and in
machine related activitiessuch asdata mining, TTS, machine
learning and cognitive computing.
4. ESPEAK: A job is done by a computer in three distinct
stages, wherein the first stage is called input (where
information is often fed using a computer and a mouse),
processing (where the input isrespondedtobythecomputer,
say, by subtracting the numbers that you typed in or by
enhancing the colors on a photo you scanned), and output
(where you can see how the computer has processed the
input, wherein it is obtained on a computer screen or as a
printout on a paper).Espeak uses a technology called as
speech synthesis, wherein a text material is loudly read out
by the computer using a real or simulated voice via a
loudspeaker. This technologyis referredtoasTextToSpeech.
5. PYTHON: Python is an object oriented programming
language that is widely used in Pi development boards. It is
the preferred language for data analytics and data science. It
wasfound towards the end of the 1980sasa successor to the
ABClanguage byGuido Van Rossum inthe NationalResearch
Institute for Mathematicsand Computer Science.Python has
a clear syntax and it is developed with the intention of
making the code readable. All these haspaved the way for its
popularity since its inception. Pythonisahigh-levellanguage.
The code for python is largely written in simple English, that
makes it understandable to the user. The commands
provided to the Pi board are mostly in English.This isinstark
contrast to an assembler that more or less operates in a way
that is comfortable for the machine, thus making it
cumbersome for a normal human to read.Pythonisvaluedby
those whowant tolearnprogramming,on accountofitsclear
syntax and the fact that it is a high level programming
language. Raspberry Pi recommends Python for those who
are willing to learn embedded programming from scratch.
7. IMPLEMENTATION OF THE PROPOSED SYSTEM
Themainidea behinddeveloping the device is theportability
and ease that it could offer to the end user. Hence, it is
deliberately intended to be simple, so that it could be of light
weight and hence, easy to carry. The module uses a
Raspberry Pidevelopment board, whichis a creditcardsized
SoC and is quite efficient. A webcam is interfaced with this
board, in order to enable the transmission of the captured
stream of video to the Pi board. An ultrasonic sensor is used
as a range finder, that enables the module to detect text data
only within a specified range. The intended simplicity of the
module comes from the coding part, which is towards the
software side. The most efficient OCR algorithm that is
available is the Tesseract OCR,whose accuracyofdetectionis
pretty high. Since portability is intended and the Pi board
runs either using a PCpower supplyor an AC adapter,we are
using a power banktopoweronthe Pi board. Theworkingof
the device thusgoeson asfollows: The Raspberry Pi board is
powered ON, on startup, the code for identifying and
recognizingthe text gets executed(Fig.4). The camera is then
focused towards the intended text. Ultrasonic sensor is used
to enable thewebcam toread the textonlywithinastipulated
area. The Tesseract algorithm then identifies each character
of the text and feeds it to another software module called as
Espeak. Espeak converts thedetected data into audiooutput.
The audiooutput ismade available at the3mmearphonejack
in the Pi board. The arrangement of the device(Fig.2 and
Fig.3), the scanned text and the obtainedoutput are included
below(Fig.5):
Fig-2: Top view of the proposed system
Fig-3: Side view of the proposed system
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 513
Fig-4: Scanned text
Fig-5: Obtained Output(since the audio output cannot be
included, the output is included in visual form)
Fig-6: Scanned object
The object scanned by the web camera is orange which is
identified by the name of citrus using the Imagga database
(Fig.6).Thetesseract detects thescannedobjectanddetectsit
as citrus and displays the output(as shown in Fig.7).
Fig-7: Output
8. CONCLUSION AND FUTURE SCOPE
A text detection and recognition with speech output system
was successfully demonstrated on Android platform. This
system is very handy and useful for the visually challenged
persons. Comparedwith a PC platform,themobileplatformis
portable and more convenient to use. This system will be
helpful for visuallychallengedpersons to accessinformation
in written form and in the surrounding. It is useful to
understand the written text messages, warnings, and traffic
direction in voice form by converting it from Text to voice. It
is found that this system is capable of converting the sign
boards and other text into speech. There is further scope in
the development of this project. The project may further be
developed into/embedded along with the walking stick,
which blind peopleusuallyuse tonavigate.Apartfromthis,in
the future, smaller development boards may come into
existence or smaller camerasmay be developed,thusfurther
enhancing the portability of this device. Flexible PCB boards
is the latest trending technology in the field of printed circuit
boards. If we implement all these into the project, the size of
the project will be greatlyreduced, thusbeingahandysystem
for the blind people.
9.REFERENCES
[1]. Alexandre Trilla and Francesc Alías. (2013), “Sentence
Based Sentiment Analysis for Expressive Text-to-Speech”,
IEEE Transactions on Audio, Speech, and Language
Processing, Vol. 21, Issue. 2. pp. 223-233.
[2]. Alías F, Sevillano X, Socoró J. C, Gonzalvo X. (2008),
“Towards high-quality next-generation text-to-speech
synthesis”, IEEE Trans. Audio, Speech,LanguageProcess,Vol.
16, No. 7. pp. 1340-1354.
[3]. Balakrishnan G, Sainarayanan G, Nagarajan R. and
Yaacob S. (2007) “Wearable real-time stereo vision for the
visually impaired”, Vol. 14, No. 2, pp. 6–14.
[4]. Chucai Yi, Yingli Tian, Aries Arditi. (2014), “Portable
Camera-based Assistive Text and Product Label Reading
from Hand-held Objects for Blind Persons”,IEEE/ASME
Transactions on Mechatronics, Vol. 3, No. 2, pp. 1-10.
[5]. Deepa Jose V. and Sharan R. (2014), “A Novel Model for
Speech to Text Conversion”, International Refereed Journal
of Engineering and Science (IRJES) Vol.3, Issue.1, pp. 39-41.
[6]. Goldreich D. and Kanics I. M. (2003), “Tactile Acuity is
Enhanced in Blindness”, International Journal of Research
and Science, Vol. 23, No. 8, pp. 3439–3445.

More Related Content

PDF
IRJET- Device for Location Finder and Text Reader for Visually Impaired P...
PPTX
1st review
PDF
Text Detection and Recognition with Speech Output for Visually Challenged Per...
PDF
IRJET- Indoor Shopping System for Visually Impaired People
PDF
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
PDF
IRJET- Review on Text Recognization of Product for Blind Person using MATLAB
PDF
docmentation with rules
PDF
TOP 5 Most View Article From Academia in 2019
IRJET- Device for Location Finder and Text Reader for Visually Impaired P...
1st review
Text Detection and Recognition with Speech Output for Visually Challenged Per...
IRJET- Indoor Shopping System for Visually Impaired People
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Text Recognization of Product for Blind Person using MATLAB
docmentation with rules
TOP 5 Most View Article From Academia in 2019

What's hot (20)

PDF
Eye(I) Still Know! – An App for the Blind Built using Web and AI
PDF
A review of consumer brain computer interface devices
PDF
IRJET- Wearable AI Device for Blind
PDF
Smart Tracking System
PDF
IRJET- Deep Learning Techniques for Object Detection
PDF
Modelling Framework of a Neural Object Recognition
PDF
33 102-1-pb
PDF
IRJET- Image to Text Conversion using Tesseract
PDF
IRJET- Navigation and Camera Reading System for Visually Impaired
PDF
Resume_Saifullah Abro
PDF
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
PDF
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
PDF
Application on character recognition system on road sign for visually impaire...
PDF
Design of a Communication System using Sign Language aid for Differently Able...
PDF
Communication among blind, deaf and dumb People
PDF
Multi-modal Asian Conversation Mobile Video Dataset for Recognition Task
PDF
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
PDF
Precision Physiotherapy & Sports Training: Part 1
PDF
IRJET- Facial Expression Recognition using Deep Learning: A Review
PPTX
Braille Technology
Eye(I) Still Know! – An App for the Blind Built using Web and AI
A review of consumer brain computer interface devices
IRJET- Wearable AI Device for Blind
Smart Tracking System
IRJET- Deep Learning Techniques for Object Detection
Modelling Framework of a Neural Object Recognition
33 102-1-pb
IRJET- Image to Text Conversion using Tesseract
IRJET- Navigation and Camera Reading System for Visually Impaired
Resume_Saifullah Abro
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Application on character recognition system on road sign for visually impaire...
Design of a Communication System using Sign Language aid for Differently Able...
Communication among blind, deaf and dumb People
Multi-modal Asian Conversation Mobile Video Dataset for Recognition Task
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
Precision Physiotherapy & Sports Training: Part 1
IRJET- Facial Expression Recognition using Deep Learning: A Review
Braille Technology
Ad

Similar to IRJET- Design and Development of Tesseract-OCR Based Assistive System to Convert Captured Text into Voice Output (20)

PDF
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
PDF
IRJET- Review on Portable Camera based Assistive Text and Label Reading f...
PDF
IRJET- Survey Paper: Image Reader for Blind Person
PDF
An Assistive System for Visually Impaired People
PDF
IRJET - A Review on Text Recognition for Visually Blind People
PDF
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
PDF
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
PDF
Smart Assistant for Blind Humans using Rashberry PI
PDF
IRJET - For(E)Sight :A Perceptive Device to Assist Blind People
PDF
Visual Product Identification For Blind Peoples
PDF
IRJET- Voice Assisted Text Reading and Google Home Smart Socket Control Syste...
PDF
A Survey on Portable Camera-Based Assistive Text and Product Label Reading Fr...
PDF
Visual, navigation and communication aid for visually impaired person
PDF
Speech Automated Examination for Visually Impaired Students
PDF
IRJET- Virtual Eye for Visually Blind People
PDF
A Robust Embedded Based String Recognition for Visually Impaired People
PDF
IRJET-Raspberry Pi based Reader for Blind People
PDF
An Approach for Object and Scene Detection for Blind Peoples Using Vocal Vision.
PDF
Device for text to speech production and to braille script
PDF
IRJET - Smart Blind Stick using Image Processing
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Review on Portable Camera based Assistive Text and Label Reading f...
IRJET- Survey Paper: Image Reader for Blind Person
An Assistive System for Visually Impaired People
IRJET - A Review on Text Recognition for Visually Blind People
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
Smart Assistant for Blind Humans using Rashberry PI
IRJET - For(E)Sight :A Perceptive Device to Assist Blind People
Visual Product Identification For Blind Peoples
IRJET- Voice Assisted Text Reading and Google Home Smart Socket Control Syste...
A Survey on Portable Camera-Based Assistive Text and Product Label Reading Fr...
Visual, navigation and communication aid for visually impaired person
Speech Automated Examination for Visually Impaired Students
IRJET- Virtual Eye for Visually Blind People
A Robust Embedded Based String Recognition for Visually Impaired People
IRJET-Raspberry Pi based Reader for Blind People
An Approach for Object and Scene Detection for Blind Peoples Using Vocal Vision.
Device for text to speech production and to braille script
IRJET - Smart Blind Stick using Image Processing
Ad

More from IRJET Journal (20)

PDF
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
PDF
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
PDF
Kiona – A Smart Society Automation Project
PDF
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
PDF
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
PDF
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
PDF
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
PDF
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
PDF
BRAIN TUMOUR DETECTION AND CLASSIFICATION
PDF
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
PDF
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
PDF
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
PDF
Breast Cancer Detection using Computer Vision
PDF
Auto-Charging E-Vehicle with its battery Management.
PDF
Analysis of high energy charge particle in the Heliosphere
PDF
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
PDF
Auto-Charging E-Vehicle with its battery Management.
PDF
Analysis of high energy charge particle in the Heliosphere
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
Kiona – A Smart Society Automation Project
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
BRAIN TUMOUR DETECTION AND CLASSIFICATION
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
Breast Cancer Detection using Computer Vision
Auto-Charging E-Vehicle with its battery Management.
Analysis of high energy charge particle in the Heliosphere
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
Auto-Charging E-Vehicle with its battery Management.
Analysis of high energy charge particle in the Heliosphere
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...

Recently uploaded (20)

PPTX
Construction Project Organization Group 2.pptx
PPTX
OOP with Java - Java Introduction (Basics)
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
PDF
PPT on Performance Review to get promotions
PPTX
Geodesy 1.pptx...............................................
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PPTX
Sustainable Sites - Green Building Construction
DOCX
573137875-Attendance-Management-System-original
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PPTX
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPTX
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PPTX
Lecture Notes Electrical Wiring System Components
PPTX
Foundation to blockchain - A guide to Blockchain Tech
PPTX
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PPT
Mechanical Engineering MATERIALS Selection
PDF
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
Construction Project Organization Group 2.pptx
OOP with Java - Java Introduction (Basics)
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
PPT on Performance Review to get promotions
Geodesy 1.pptx...............................................
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
Sustainable Sites - Green Building Construction
573137875-Attendance-Management-System-original
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
Embodied AI: Ushering in the Next Era of Intelligent Systems
bas. eng. economics group 4 presentation 1.pptx
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
Lecture Notes Electrical Wiring System Components
Foundation to blockchain - A guide to Blockchain Tech
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
Automation-in-Manufacturing-Chapter-Introduction.pdf
Mechanical Engineering MATERIALS Selection
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf

IRJET- Design and Development of Tesseract-OCR Based Assistive System to Convert Captured Text into Voice Output

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 509 DESIGN AND DEVELOPMENT OF TESSERACT-OCR BASED ASSISTIVE SYSTEM TO CONVERT CAPTURED TEXT INTO VOICE OUTPUT G.ELUMALAI1,J.SUNDAR RAJAN2, P.SURYA PRAKASAH3, V.L.SUSRUTH4,P.K.SUDHARSANAN5 1 Associate Professor, Dept. of Electronics and Communication Engineering, Panimalar Engineering College, Tamil Nadu, India 2,3,4,5 UG students, Dept. of Electronics and Communication Engineering, Panimalar Engineering College, Tamil Nadu, India ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract - The biggest challenge faced by the blind people, is their inability to view real life objects and to read. The only efficient system that exists so far, is the braille system, that enables the blind to read. This system is time consuming and the time taken to recognize the text is long. Our aim here is to reduce the time taken to read. In our work, using a Raspberry Pi, we have designed a smart reader, so that the blind people may read. The module that we have designed, either uses a webcam or a mobile camera that is linked with a RaspberryPi, to focus on a range of printed text. The OCR (OpticalCharacter Recognition) package installed in raspberry pi tests it into a digital article which is then subjected to skew modification, segmentation, before feature extraction to perform sorting. Our proposed project automatically focuses the regionsoftext in the object, after which the text characters are thenlocalized using a localization algorithm, that uses feature recognition and edge pixel distribution using artificialneuralnetwork.The text characters are then binarized into a machine readable form by using an OCR algorithm called as Tesseract. The recognized characters of the text are then converted into audio format, so that they may be recognized by the blind users. Key Words: Embedded devices, Image processing, OCR, Neural Networks, Raspberry Pi. 1. INTRODUCTION The importance of reading in our society, today, cannot be ignored. We could find printed text everywhere in the form of medical prescription, newspapers, books, handouts, medicine bottles and tablets and the like. While video magnifiers, optical aids, and screen readers can help with low vision and blind users to access documents, few devices do exist that provide optimum access to daily-life hand held objects such as fruit juice cartons and invention packages. The most important causes for visual degradation include diabetic retinopathy, aging factor, eye diseases such as ocular tumors and accidents, which lead to an increaseinthe number of visually disabled people, nowadays. Cataract is leading cause of blindness and visual impairment. Mobile applications are available these days, thataidinvisualization and they form an integral part of the blind peoples lives. Recent advances in mobile technology, digital camera, computer vision and camera-based application make it possible to support visually challenged persons by developing camera-based application that combine computer vision with other existing technology such as optical character recognition (OCR) system. To detect text information from image, practical difficulties exist, such as non-uniform backgrounds, due to the large variations in character font, size, texture, color, background, orientations, and many other reasons. Text detection from scene/text camera images is possible due to high resolution camera. 2. PROJECT OBJECTIVE The biggest challenge faced by the blind peopletodayistheir inability to read and recognize the real-world objects. The only efficient method that exists so far is the braille method, which enables them to read. It is somewhat time consuming to read the text using a braille and also it might be not be economically feasible. The objective of the project is to create a portable, light-weight module that enablestheblind to read. This module uses a webcam to capture the image of the focused text, extract separate text characters from the scene and then provide a speech output using the Tesseract OCR. 3. LITERATURE SURVEY In the field of Text To Speech conversion, researchisfocused on the identification of the input text and the ability to express it in a natural manner. Sentimental analysis procedure(positive/neutral/negative) is used in order to input features for expressive speech synthesis, where identification of the input text is a front end task. Different combinations of classifiers and textualfeaturesareevaluated for the determination of the most appropriate adaptation procedure. Semeval 2007 dataset and Twitter corpus are used for the evaluation of the effectivenessof thisscheme of Sentiment Analysis, which is appropriate for an expressive Text To Speech scenario. Conducted experiments validate the procedure proposed with respect to Sentiment Analysis [1]. MD-TTS TC not only considers not only the contents of the text but also its structure, in contrast to the topic of text classification tasks. The proposal is validated by the experiments that were conducted in terms of both subjective(synthetic speech quality) and objective(TC efficiency) evaluation criteria. [2] For the visually disabled people, travel aids have been developed in the recent years in the form of stereovision-capable devices. They are however, bulky and assist the blind person in avoiding obstacles only and they are head-mounted. We researched
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 510 about a wearable system that can not only assist in avoiding obstacles, but at the same time supports live streaming of videos over the 3G network. The module is comprised of an eye-glass and an embedded device that is power efficient. Two miniature camerasare used for imaging, on one end,by the eye-glass. For the combination of stereo images, FPGA and FIFO buffers are used. Also, live stereo matching on an embedded processor is achieved by means of parallel processing. Live streaming over 3G network, of the video that is captured by the system is also supported. A person with healthy eyesight can guide a visually impaired person, with the help of this livevideo streaming [3]. We researched about a system that helps the blind people to read from the labelsand product packages of common handhelddevicesin their daily lives, by means of a camera. The Region Of Interest(ROI) in the camera view is obtained by using an innovative motion-based approach, where the users were asked to shake the object. This is done in order to isolate the intended image from cluttered backgrounds and other objects in the camera view. A mixture of Gaussian based, background subtraction methodswere used by thismethod. Text information is obtained from the extracted ROI, by a combination of localization and recognition. Feature extraction and the distribution of edge pixels are studied, using Adaboost model, to automatically localize the text characters in the ROI. Localized text characters are then identified by using the OCR algorithms. Recognized text is conveyed to the blind users using audio output. The effectiveness of the system hardware is evaluated by using the dataset that is obtained from ten blind people[4]. In TTS technology, the ability is provided to a computer to speak. A synthesizer is used to convert the text into speech, thus making it eligible for further processing. By learningvarious textbooks related to languages, language mastery was obtained, traditionally. Due to the implied cost factor, moving along this method was very difficult. Various dictionaries were used by the existing system, to pronounce the identified text with correct pronunciation . Only for a set of words which are available in the dictionary, the operation is supported. [5] Studies in blind people have shown the activation of cortical areas that sub serve the vision, but it has long been controversial whether blind people have tactile acuity. We compared the passivetactileacuityofblind and sighted subjects on a fully automatedgratingorientation task and used multivariate Bayesian data analysis to determine predictors of acuity. Irrespective oftheamountof childhood vision, braille reading or light perception level, acuity was superior in blind people. Acuity is better in women than in men and it also decline with age. Acuity is chiefly dependent on the force of contact between the stimulus surface and the subject . The difference between blind people and healthy sighted people varied a lot in spite of the intragroup variation: the acuity of both sighted and blind people of the same gender remained the same, except that the acuity of a blind person was similar to that of a sighted person who was 23 years younger. The results suggest that cross modal plasticity may underlie tactile acuity enhancement in blindness. [6] 4. EXISTING SYSTEM A few systems already exist that help the blind people have access to daily-life hand held objects, but they are largely ineffective with respect to the focusarea of the labelling.For example, portable bar-code readers do exist, but it is very difficult for a blind person to locate a bar codeandthenpoint the laser beam on that bar code. Also, to enable the blind people to read, braille system exists. In spite of ignoring the manufacturing cost, the system can be said to be largely ineffective with regard to the processing speed, that is, the speed at which the blind subject can identify theinformation and then assimilate it is too low. 5. PROPOSED SYSTEM The proposed system uses a Raspberry Pi board, an ultrasonic sensor and a webcam to recognize the text in the scene. In this, the webcam is focussed on the scene. A video streaming is obtained, from which the images are captured frame by frame. The images are refined in order to eliminate any noise that is present in it. A feature called segmentation is used in order to separate each character from other in the text. Graphical details such as icons or logos, if any, are eliminated. Each obtained character are compared with the datasets that are created as a part of the Tesseract library. The Tesseract OCR is the most efficient algorithm available that checks for the obtained character in ten dimensions. Once, the character is recognized, it must be made available as an audio output. For this, we use a software calledfestival. The festival is used to provide the audio output for the recognized character. Apart from these features, an extra feature is added, that enables the blind to know the type of object that he/she interacts with.(a menu, newspaper and the like). An ultrasonic sensor is included as a part of the project, that makes the project obtain charactersonlywithin a particular distance.
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 511 Fig-1: Block diagram of the proposed system 6. SYSTEM DESIGN SPECIFICATION 6.1. Hardware specification 1. RASPBERRY PI:The RaspberryPi,usingtheARM1176JZF-S core, is a credit card sized single computer on chip or SoC. System on chip is a method by which all the electronic components that are necessary to run a computer systemare placed on a single chip. In order to start up, the Raspberry Pi needsan OS.On board non-volatilememory,whichstoresthe bootloaders are eliminated in the Raspberry Pi, with the aim of cost reduction. Traditional embedded systems use Linux kernels and file system. For this reason, an SD/MMCcardslot is provided. Raspberry Pi will be executed according to the application program and after the completion of successful boot loading. 2. ULTRASONIC SENSORS: Ultrasonic sensors provide cost effective, unique sensingmethods, which arenotprovidedby any other technology, thus serving the market at large. Application problems that are eithercostprohibitiveorthose that couldn’t be solved using other sensors can be solved by an ultrasonic sensor, by applying a wide variety of ultrasonic transducers and a wide variety of frequency ranges. Detection over distance are required by a variety of applications in industrial sensing. The detection range of ultrasonic sensors is quite high, as high as up to forty feet, which limit switches and inductive switching cannot do. Photo electric sensorscandetectlong distances,buttheylack a wide area coverage or a wide area coverage which can be obtained by using a numberof sensors. Migatron’sultrasonic sensor cover both narrow and wide areas. Proper ultrasonic transducer selection is what it all takes. Since there are wide range of target materials available, only ultrasonic sensors are independent of the target material composition . The target material may be solid, liquid, porous, soft or woody in nature and irrespective of its target composition, it can still be detected. Distance measurement is easy and it can be calculated by using the time from which theultrasonic signal leavesthe signaltothe time theultrasonicreturnsbacktothe sensor. This method is accurate to .05% of range which equates to +or- .002 of an inch at a distance of 4 inches. 6.2. Software specifications: 1. OPEN CV: OpenCV (Open Source Computer VisionLibrary) is a collectionof library functions, whose main objective is to enable computer vision. The main objective of OpenCV is to provide a framework forcomputer vision and to improvethe readabilityofmachinesover commercialproducts. Sinceitis BSD licensed, businesses can easily use and modify the code. Over2500 sophisticated algorithmsarepresentinthelibrary, which include both machine learning and computer vision algorithms. Using the algorithms, the computer can identify objects, identify faces, recognize moving objects, recognize human actions, extract 3D models of objects, combine different images to produce an entire image scene, red eye removal, scenery recognition and augmented reality etc. can be performed. Around 47 thousandmemberusercommunity for OpenCV exists and OpenCV is widely used by many commercial bodies, governmental institution and research organizations. 2. NEURAL-NETWORKS: Neural networksare a collectionof hardware and software modules that are patterned around the neural networks of the human brain, in technical terms of information technology. They are a variety of deep learning technology that are also termed as artificial neural networks. Complex signal processing and complex patter recognition tasks are some of the commercialapplicationsof the neural networks. Image processing and text recognition, handwriting recognition, weather prediction and facial recognition are some examples of the applications of neural networks since 2000. A neural network usually involves a large number of parallel operating processors arranged in tiers. Raw input information are received in the first tier – analogous to optic nerves in human visual processing. Instead of receiving the output from the raw input, each tier receives the output from the tier that precedes it. The last tier produces the output of the system. Every node that processes possess its own sphere of knowledge, including what it was originally programmed with, or what is has learnt or those set of rules which it has created for itself.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 512 3. OCR: Optical Character Recognition is the conversion of either printed or handwritten characters into a machine readable format(binary 0s and 1s). The text of concern may be obtained either from an imageof a document,orascanned copyof the documentor from areallife scene such asthetext on signs and billboards(Fig.6 and Fig.7) or from subtitle text underneath an image (such asthose thatoccur ina television broadcast). It is used to enter the information from different types of documents such as records, passports, cheques, invoice, bills or any other such suitable type of documentation. OCR is a widely used method for converting the printed data into digital format, so that they may be edited, searched , compressed, used in line display and in machine related activitiessuch asdata mining, TTS, machine learning and cognitive computing. 4. ESPEAK: A job is done by a computer in three distinct stages, wherein the first stage is called input (where information is often fed using a computer and a mouse), processing (where the input isrespondedtobythecomputer, say, by subtracting the numbers that you typed in or by enhancing the colors on a photo you scanned), and output (where you can see how the computer has processed the input, wherein it is obtained on a computer screen or as a printout on a paper).Espeak uses a technology called as speech synthesis, wherein a text material is loudly read out by the computer using a real or simulated voice via a loudspeaker. This technologyis referredtoasTextToSpeech. 5. PYTHON: Python is an object oriented programming language that is widely used in Pi development boards. It is the preferred language for data analytics and data science. It wasfound towards the end of the 1980sasa successor to the ABClanguage byGuido Van Rossum inthe NationalResearch Institute for Mathematicsand Computer Science.Python has a clear syntax and it is developed with the intention of making the code readable. All these haspaved the way for its popularity since its inception. Pythonisahigh-levellanguage. The code for python is largely written in simple English, that makes it understandable to the user. The commands provided to the Pi board are mostly in English.This isinstark contrast to an assembler that more or less operates in a way that is comfortable for the machine, thus making it cumbersome for a normal human to read.Pythonisvaluedby those whowant tolearnprogramming,on accountofitsclear syntax and the fact that it is a high level programming language. Raspberry Pi recommends Python for those who are willing to learn embedded programming from scratch. 7. IMPLEMENTATION OF THE PROPOSED SYSTEM Themainidea behinddeveloping the device is theportability and ease that it could offer to the end user. Hence, it is deliberately intended to be simple, so that it could be of light weight and hence, easy to carry. The module uses a Raspberry Pidevelopment board, whichis a creditcardsized SoC and is quite efficient. A webcam is interfaced with this board, in order to enable the transmission of the captured stream of video to the Pi board. An ultrasonic sensor is used as a range finder, that enables the module to detect text data only within a specified range. The intended simplicity of the module comes from the coding part, which is towards the software side. The most efficient OCR algorithm that is available is the Tesseract OCR,whose accuracyofdetectionis pretty high. Since portability is intended and the Pi board runs either using a PCpower supplyor an AC adapter,we are using a power banktopoweronthe Pi board. Theworkingof the device thusgoeson asfollows: The Raspberry Pi board is powered ON, on startup, the code for identifying and recognizingthe text gets executed(Fig.4). The camera is then focused towards the intended text. Ultrasonic sensor is used to enable thewebcam toread the textonlywithinastipulated area. The Tesseract algorithm then identifies each character of the text and feeds it to another software module called as Espeak. Espeak converts thedetected data into audiooutput. The audiooutput ismade available at the3mmearphonejack in the Pi board. The arrangement of the device(Fig.2 and Fig.3), the scanned text and the obtainedoutput are included below(Fig.5): Fig-2: Top view of the proposed system Fig-3: Side view of the proposed system
  • 5. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 513 Fig-4: Scanned text Fig-5: Obtained Output(since the audio output cannot be included, the output is included in visual form) Fig-6: Scanned object The object scanned by the web camera is orange which is identified by the name of citrus using the Imagga database (Fig.6).Thetesseract detects thescannedobjectanddetectsit as citrus and displays the output(as shown in Fig.7). Fig-7: Output 8. CONCLUSION AND FUTURE SCOPE A text detection and recognition with speech output system was successfully demonstrated on Android platform. This system is very handy and useful for the visually challenged persons. Comparedwith a PC platform,themobileplatformis portable and more convenient to use. This system will be helpful for visuallychallengedpersons to accessinformation in written form and in the surrounding. It is useful to understand the written text messages, warnings, and traffic direction in voice form by converting it from Text to voice. It is found that this system is capable of converting the sign boards and other text into speech. There is further scope in the development of this project. The project may further be developed into/embedded along with the walking stick, which blind peopleusuallyuse tonavigate.Apartfromthis,in the future, smaller development boards may come into existence or smaller camerasmay be developed,thusfurther enhancing the portability of this device. Flexible PCB boards is the latest trending technology in the field of printed circuit boards. If we implement all these into the project, the size of the project will be greatlyreduced, thusbeingahandysystem for the blind people. 9.REFERENCES [1]. Alexandre Trilla and Francesc Alías. (2013), “Sentence Based Sentiment Analysis for Expressive Text-to-Speech”, IEEE Transactions on Audio, Speech, and Language Processing, Vol. 21, Issue. 2. pp. 223-233. [2]. Alías F, Sevillano X, Socoró J. C, Gonzalvo X. (2008), “Towards high-quality next-generation text-to-speech synthesis”, IEEE Trans. Audio, Speech,LanguageProcess,Vol. 16, No. 7. pp. 1340-1354. [3]. Balakrishnan G, Sainarayanan G, Nagarajan R. and Yaacob S. (2007) “Wearable real-time stereo vision for the visually impaired”, Vol. 14, No. 2, pp. 6–14. [4]. Chucai Yi, Yingli Tian, Aries Arditi. (2014), “Portable Camera-based Assistive Text and Product Label Reading from Hand-held Objects for Blind Persons”,IEEE/ASME Transactions on Mechatronics, Vol. 3, No. 2, pp. 1-10. [5]. Deepa Jose V. and Sharan R. (2014), “A Novel Model for Speech to Text Conversion”, International Refereed Journal of Engineering and Science (IRJES) Vol.3, Issue.1, pp. 39-41. [6]. Goldreich D. and Kanics I. M. (2003), “Tactile Acuity is Enhanced in Blindness”, International Journal of Research and Science, Vol. 23, No. 8, pp. 3439–3445.