Вы находитесь на странице: 1из 40

“AUTOMATED READING PROTOTYPE”

(A.R.P.)
Cachola, Allyson Kae L. – 10 – Quantum Physicists
Dulay, Bianca Laine C. – 10 – Quantum Physicists
Ele, Emanuelle Nathalie M. – 10 – Biophysicists
Hojilla, Nathan Roche E. – 10 – Atomic Physicists
Yabut, Ivanne Cres V. – 10 – Quantum Physicists
University of Baguio Science High School
ABSTRACT

Reading is a way of processing and comprehending texts. Individuals who have reading
difficulties, such as visual impairment, learn through reading aloud as this help reinforce
the skills and may also help the individuals enjoy and learn from books they struggle to
read. Automated Reading Prototype (ARP) is a mock-up design which reads printed
manuscripts and converts text to audio. The prototype reads the manuscript by taking an
image of the full text, then, it converts the image to a reading manuscript. The prototype
reads the coded manuscript using Optical Character Recognition (OCR) and Text-to-
Speech (TTS) converters embedded in the prototype’s system to make an audio file. This
is a descriptive study that uses frequency to check the optimal distance in getting the full
text image, surface type and text styles of a manuscript. Results show from the tests made
that the ideal distance of the camera is five to seven inches away from the text. The
prototype can read text on both flat surface area and open books. The text styles that were
positively read vary from print, italicized, bold, and underlined texts. Thus, the prototype
made clear results in converting texts to audio file. The researchers recommend that the
prototype’s OCR can be widened and is not limited to specific fonts but also to handwritten
texts, surfaces that are not just flat and interpretation of images as well as flowcharts,
graphs and illustrative texts that makes an additional learning and creative way of
synthesizing ideas.

Keywords: Optical Character Recognition, Text-to-Speech, Python Programming


Language, Text, Audio
INTRODUCTION

1
Reading is a way of processing and comprehending texts (Virginia & De la Fuente,
2015). It may also be a skill or activity of getting information from books (Cambridge
Dictionary, 2018).
Reading is sometimes for pleasure or escape, other times it is for the disciplined acquisition
of information. Exposure to books, poetry, writing, and even film and videos can be
therapeutic and beneficial in helping us process our own life experiences. In other words,
literature can be used to help us figure out life, heal emotional traumas, and change
thoughts and behavior (Demott, 2018). A factor that is greatly affected by people
with visual disabilities is their ability to learn about new things. The manner in which most
people learn rely greatly on vision, and with visual disabilities, learning is greatly hindered.
This causes people with visual disabilities to have low motivation in learning, exploring
the environment, and interacting with people around them. Without experiencing these
basic skills, early motor development may be affected especially in children. They also
cannot relate to visual experiences with sighted people, having a negative impact on their
social skills and so, experience low self-esteem. These may limit their “control” over their
own lives (Texas Council for Developmental Disabilities, 2013). Lapkin (2018) says that
a way of helping a person with visual conditions is by reading aloud to the person as this
helps reinforce the skills and may also help the individuals enjoy and learn from books they
struggle to read. Audio books is another alternative solution as they can connect the sound
they hear to the letters they see.
Moreover, for avid readers like the researchers themselves, the idea of
bibliotherapy, which is the engagement with literature and written word has the potential
to change, calm, inform, inspire and heal, is not new at all according to Baker (2014). Many
people feel better after curling up with a good book. There is a feeling of goodness for the
heart and soul and it is not unusual to have a feeling of friendship within the page. Linked
to its impact on quality of life and well-being, this study reveals that reading plays an
important role in helping people in the society and partially sighted adults cope with
pressure, including significant moments such as bereavement; engaging in meaningful
activity that passes time, occupies the mind and represents a stimulating alternative to
activities that are no longer possible to undertake.

2
In this world, technology is extremely popular, and even with technology, comes
text and literature, outputs that are to be read in any form. Reading makes you aware, like
how reading signs makes you aware of what is there, and other necessary information that
can help and at the same time protect a person. Reading does not end there; it is also used
for entertainment and pleasure (Hebert, 2017). Reading embarks on a new emotional and
intellectual journey in life. It benefits people to open new horizons and understand more
about the world even if they cannot see it just like the researchers do. It engages them in
social issues and helps make decisions in life (Gail, 2015).
Love for reading is what the researchers have in common. Realizing how important
reading is in daily life and that it is taken for granted creates a profound appreciation for
this ability. Those who could not read, or have difficulty in reading, no matter what reason
is what inspired the creation of the study’s prototype. The prototype can give the individual
who needs it, the ability to read, in a world where information is important. Simply being
able to contribute to society by creating innovations that help people is enough motivation
for the researchers to develop such a device.
The aim of this study is to develop a prototype that uses Optical Character
Recognition (OCR) to translate images to digital text that will be “read” or converted into
audio using text-to-speech (TTS) technology. It is named as Automated Reading Prototype
as this is what the device does, reads automatically.
Automated, means to operate automatically, largely or wholly involuntary by using
prototypes or computers to reduce the time taken and work done by people through a self-
acting or self-regulating system (Merriam Webster & Cambridge Dictionary, 2018).
According to Bix (2018), automated systems can act faster in a certain function than what
most people can do, such as making decisions faster than what a standard person can and
are also used to do work that may be too repetitive or boring for humans, such as
assembling and packaging. A clear exemplar of an automated system would be a machine,
which is a piece of equipment that does a type of work faster and more efficiently than how
humans can (Cambridge Dictionary, 2018). As this device is not industrialized and is
subject to research and testing, it is merely a prototype.
The Automated Reading Prototype (A. R. P.) can help people with such conditions.
These people may listen to the prototype as it reads aloud the text it takes a picture of, while
the prototype reads, the users may also connect the words they hear to the letters and texts

3
that they see, just like an audiobook. It is also not a problem for these people to use the
camera to picture the documents as they have no trouble with seeing and only with reading.

Literature Review
Text-to-speech (TTS) converts text files to speech that is narrated or read using
phonemes and joined syllables using algorithms. (Kaushik, Mohith, Rangoonwala &
Samiappan, 2017). A TTS converter changes texts to speech and can be created by putting
together pieces of recorded speech that are stored in a database. Alternatively, a synthesizer
can integrate human vocal characteristics to create a synthetic audio output. (Jadhav, Patil
& Shetake, 2014). Optical Character Recognition is an algorithm to enable prototypes to
recognize characters from printed, typewritten documents or handwritten manuscripts. It is
a process that converts text documents into editable or searchable text through visual
recognition. (Kumar, 2014). OCR allows prototypes to automatically recognize text in
handwritten, printed and scanned images and convert it into an editable digital format for
deeper processing. (Hamad & Kaya, 2016). A reading machine is a device that helps those
affected by low vision, visual disabilities and blindness to continue enjoying the pleasure
of reading. These devices utilize OCR and TTS in order to accurately read out words from
a printed material, with these aids it is possible to read a wide array of printed text like
letters, newspapers, and more. (Vision Aid Technologies, 2017).
A proposed technique to extract text from typed documents, convert them into
machine encoded text, create the text files and then process them use Digital Image
Analysis (DIA0) to convert the text into audio output. Digital Image Analysis extracts
meaningful information from images; mainly from digital images by means of digital
image processing techniques (Labautopedia, 2013) It process a scene of data for an
autonomous machine perception which in this study’s case is the Automated Reading
Prototype.
As most reading machines have to be plugged to a socket and are often stationary,
this study aims to develop a prototype that is supplied by a power bank to be convenient
for the user to wear and bring around. In relation to concepts and researches found, there
is substantial foreground to develop this prototype.

4
The Raspberry Pi 3 Based Reader for Blind People study proposes a machine used
for the detection and reading of documented text in images to help the blind and visually
impaired people. It enhances the capabilities of the blind people by providing them a
solution so that the information can be fed to them in the form of a speech signal (Goel,
Sehrawat, Patil, Chougule, & Kahatavkar, 2017). This study would like to better their
research and make a prototype which is portable.

Paradigm of the Study

Input: Process: Output:


Raspberry Pi 3 B+ Gather all necessary materials. Automated
with Case Reading
Insert Micro SD card into the Prototype
Power bank motherboard.
USB cable Program OCR and TTS using
Python Programming
USB Camera Language.
Earphones Save program in Micro SD
Python Programming card.
Language Program button to take picture.
Programmable button Sew USB Camera unto Glove.
Micro SD card Sew motherboard case.
Schematic diagram Attach case to motherboard
Glove Attach USB camera, power
Custom Sewn Case bank, button and earphones to
the motherboard.
Schematic Diagram
Test prototype.

Significance of the Study


The prototype is designed to benefit the users by serving as an aid to successfully
convert printed text to audio. A few researchers have developed a Raspberry Pi Based
Reader aid for visually impaired people (Goel, Sehrawat, Patil, Chougule, & Kahatavkar,
2017). This prototype can be adopted by people who have trouble in reading large amounts
of texts due to dyslexia, literacy difficulties, other learning disabilities and the mild or

5
moderately visually impaired, and even for the people with who lack the visual capacity to
read.
This study can be used for future researchers that may want to explore the
capabilities of the Automated Reading Prototype, test its effectivity on potential people to
benefit like the visually disabled or those who have reading difficulties, and may serve as
a guide for researchers who wish to make a similar or improved prototype of this model.
The findings of the study also seek to explore the potential of the Automated Reading
Prototype by testing different surface areas of the text, different font types, and complex
figures. In doing so, the study will help the researchers uncover critical areas that previous
researches were not able to explore.
The design of this study’s Automated Reading Prototype (A. R. M.) is created to
be convenient and efficient for the user. The camera takes a picture of the document the
user wishes to “read” and translates it into audio almost immediately. It is portable and
light, making it easy to carry around. Earphones are available for the user to listen from
as to not cause disturbance to the people around him/her and at the same time is suitable
for reading documents the user may wish to remain confidential.

Objectives of the Study


This study intends to develop a prototype that can convert printed text to audio. To be able
to maximize the efficiency of the prototype, it aims:
· To determine the components of the prototype.
· To measure the distance between the camera and the text.
· To evaluate what kind of surface the text must be printed on.
· To compare what type of fonts the prototype is limited to reading.
· To assess the capabilities of the prototype to read diagrams and figures.
MATERIALS AND METHOD

Study Design
The study uses an experimental design to accomplish the objectives set forth by the
researchers prior to the testing of the prototype. In testing the limits of the prototype, the
experimental design will focus on four aspects, namely, the (1) distance between the

6
camera and text, (2) object’s surface area type, (3) fonts that can be recognized or read, and
(4) the capabilities of the prototype to interpret various figures and diagrams.

Materials
The Raspberry Pi 3 is a tiny credit card sized computer. Adding electronic
components such as mouse and SD card with installed Linux distribution will give you a
full-fledged computer that can run applications from word processors and spreadsheets to
games. As the Raspberry Pi supports HD video, it is possible to create a media center
within. It is an open source from get-go, expected to be the de facto embedded Linux board
in all the forums. This was the chosen computer used to serve as the motherboard and the
main brain for the whole prototype. It is protected by the given case it came with and
incorporates within it the Python 3 compiler as well as the Python programming language,
three Universal Serial Bus (USB) ports which the researchers used to plug the second
materials needed; the USB Camera; to capture the printed material. Python programing
language is a programming language used for general programming. Python is an
interpreted, object-oriented, high level programming language with dynamic semantics.
It’s high-level built-in data structures, combined with dynamic typing and dynamic
binding, make it very attractive for Rapid Application Development, as well as for use as
a scripting or glue language to connect existing components together. Python is simple,
easy to learn syntax emphasizes readability and therefore reduces the cost of program
maintenance. Python also increase productivity that it provides. A USB cable was used to
connect to a power bank, which served as the power supply of the prototype and for
portability. The last port will function for speakers and a given audio jack wherein the
earphones can be plugged whichever is preferred. A wrist mouthed glove and a custom
sewn case was used to attach the camera and the prototype to the user’s hand and arm
respectively. To test the prototype’s limitations, different distances between the camera
were tried using a ruler, different surface types like flat documents, opened books and texts
on bottles, fonts of texts, handwritten texts and even figures were used to be experimented
on. Refer to Appendix D for the pictures of the materials.

7
Raspberry Pi 3 Model B+ case Earphones

8
USB Camera Wrist Mouthed Glove

USB cable
Power bank

Raspberry Pi 3 B+ motherboard
Procedure
The Automated Reading Prototype was made by gathering all the materials needed
for the development of the invention.
In building the prototype, a Micro SD card was inserted to the motherboard, where the
programming of the button to capture the image, OCR and TTS was made and integrated.
With the use of a needle and thread, the USB camera was sewn into the wrist mouthed
glove, while the cable of the camera was attached to the USB port of the motherboard. The
prototype and power bank will then be attached to the custom sewed case to be worn by

9
the user. The power bank cable connected to the power bank was attached to one of the
ports of the motherboard as it served as the power supply. Speakers or earphones were then
attached to the motherboard via the third USB port or audio jack respectively. Bluetooth
devices may be used as well, depending on the preference of the user.
The number of trials for the different variables of the prototype’s limitations are
three each. In measuring the distance between the camera and the text, 6 distances will be
used, 1 inch, 3 inches, 5 inches, 7 inches, 9 inches and 11 inches. A ruler in the unit of
inches was used to determine the distances. To evaluate the object’s surface area
compatibility with the prototype, it will be tested to read texts on flat documents, opened
books and convex surfaces like water bottles. Comparison of recognizable font types will
be among printed texts, bold texts, italicized texts, underlined texts, handwritten print text,
handwritten cursive texts and digital texts or text displayed on computers. To further assess
the capabilities of the prototype, it will be tested on more complex images like figures of a
bar graph, flow chart, Venn diagram and tree diagram. The whole procedure took roughly
3 weeks to complete.

Gathering of all necessary materials Saving of program to the Micro USB Card

10
Inserting of Micro USD card into the Sewing of the USB Camera to the
mother board Wrist Mouthed glove

Programming of OCR and TTS using Sewing of Motherboard Case


Python Programming Language

Attaching of USB camera, power bank, Testing of Prototype

11
Button and earphones to the mother
Board

Treatment of Data
The data for this study is nominal or, nonparametric test was used by the researchers
for this study as no statistics were involved. Alongshore with its possibilities and
assumptions while determining whether there is any agreement between the results. Since
trials were used to test the device, a frequency test was also applied, determining how often
the prototype was successful in performing the given task or reading the given text.

Ethical Consideration
The researchers had executed and planned the entirety of the project with careful
consideration on the overall safety of each researcher in creating the prototype. Personal
protective equipment was used while assembling the prototype. An engineer was sought
out to help with proper wiring and building of the ergonomic and portable function of the
prototype.

RESULTS AND DISCUSSIONS

Components of the Prototype


Raspberry Pi is a low-cost computing platform. The model was programmed using
Python programming language. Python syntax is easy to read, very clean, with an emphasis
on readability and uses standard English keywords. The Raspberry Pi 3 B+ weighs 42
grams making it lightweight and easy to store and carry. It has excellent graphics with 400
MHz Dual Core Video Core IV GPU. It makes use of a microSD Card with a memory of
1 GB LPDDR2 SDRAM. The Raspberry Pi 3 B+ has HDMI ports that support
multichannel audio output and has an audio line out 3.5-mm headphone jack (analog). It

12
has Four USB 2.0 ports (up to 480 megabits per second) in which the researchers connected
a power bank with an input voltage of 5V DC. An A4TECH 16 megapixels USB camera
and a programmable USB button are connected to the Raspberry Pi 3 B+ model as well.

Distance Measurement Comparison of the Automated Reading Prototype


In the tests shown in table 1, the researchers compared how far the camera must be
from the desired text. The prototype was tested on six (6) distances, three (3) trials each.
The prototype’s response is successful if it is able to recognize the text given. However,
the response of the prototype is unsuccessful if the prototype cannot recognize the given
text. Results show that the prototype must be near the text to successfully read it and not
be too far as it would not recognize the characters that are too small.

Table 1

Distance Measurement Comparison


Distance (inches) Trial Prototype’s Response Result

1 Unrecognized text Unsuccessful

1 inch 2 Unrecognized text Unsuccessful

3 Unrecognized text Unsuccessful

1 Recognized incomplete text Unsuccessful

3 inches 2 Recognized incomplete text Unsuccessful

3 Unrecognized text Unsuccessful

13
1 Reads Text Successful

5 inches 2 Reads Text Successful

3 Reads Text Successful

1 Reads Text Successful

7 inches 2 Reads Text Successful

3 Reads Text Successful

1 Reads Text Successful

9 inches 2 Reads Text Successful

3 Unrecognized text Unsuccessful

1 Unrecognized text Unsuccessful

11 inches 2 Reads Text Successful

3 Unrecognized text Unsuccessful

The distance of 1 inch from the text shows that the prototype cannot recognize the
text completely, making it unsuccessful. 3 inches showed that the prototype recognized
some of the text. Words that were at the left and right corners of the document were not
read as they were not captured by the camera, while in the last trial the prototype was
unsuccessful in reading the text. Concluding this distance to be ineffective and not ideal.
At a distance 5 inches, the prototype completely read the texts. The distance of 7 inches
also remained favorable as the prototype read the text completely without failure. The
distances of 9 and 11 inches seemed too far from the text for the prototype to read, it did
not recognize the text one and two out of three trials respectively. These distances are not
advisable to use since the text may be too blurry for the camera to comprehend.
The ideal distance of the camera from the text would be 5 inches. Confirmed by
Engineer KC Mendoza, the researchers engineer, this enables the prototype to capture the

14
document completely, but not too close to lose recognition of some words and letters. The
appropriate distance of the camera from a document depends on the text itself. If the text
is small, the camera may need to be closer, if the text is small but the paragraphs take a lot
of space, the camera must be far enough to capture everything but not too far as to have
difficulty in recognizing characters. The distance of the camera from the text will depend
highly on what size and quantity of text is to be read. Determining this distance is easy as
one just needs to be practical.
The distance measurement was tested mainly to figure out the best distance of the camera
from the desired text, for a more convenient way to use the prototype. With the results of
the distance measurement comparison, the user of the prototype is given a guideline for
how far the camera should be so the text can be read easily. If the distance is inappropriate
for the size of text to be read, the prototype may fail in recognizing the characters from the
image and therefore will be unsuccessful in reading to the user.

Surface Area Compatibility of Automated Reading Prototype


In the tests shown in table 2, the researchers determined what kind of surface the
text must be printed on. The prototype was tested on three (3) variables, three trials each.
The prototype’s response is successful if it is able to read the given text on the object’s
surface area. However, the response of the prototype is unsuccessful if the prototype cannot
recognize the given text on the surface area it is on. Results show that the prototype
converted printed text to audio on even surfaces and opened books. Scanning the document
on convex surfaces resulted to the distortion of letters thus the capability of the camera to
effectively scan the documented was limited.

15
Table 2

Surface Area Compatibility of Automated Reading Prototype

Object’s Surface
Trial Prototype’s Response Result
Area

1 Reads Text Successful

2 Reads Text Successful


Flat Document

3 Reads Text Successful

16
1 Reads Text Successful

2 Reads Text Successful


Opened Book

3 Reads Text Successful

1 Unrecognized text Unsuccessful

2 Unrecognized Text Unsuccessful


Convex Surface
3 Reads Text Successful

The camera’s OCR distinguishes the characters in text clearly in a printed flat
document. These surfaces are ideal for the prototype to read since the picture and text are
easily recognized. Although the surface of an open book is not as flat, the prototype can
still recognize the characters in the book as long as the texts are not smudged or too
distorted, especially by the curve of the book’s pages near the book’s backbone. Usually,
texts on convex surfaces are too far apart that the image becomes too distorted for the OCR
to properly read. Some letters of the text, mainly the ones at the sides of the word may no
longer be captured by camera due to the curvature of the surface. This results in
unrecognizable text and if read by the prototype, the recognized word will only be read.
Text on curved surfaces may also vary, if the text is too far apart the prototype cannot
recognize the characters, but if the text is small yet still on a curved surface, the prototype
can read it provided that it is not too dilapidated.
The Engineer who helped the researchers with processing the said experimentation
explained that determining the object’s surface area helps the user know what kind of
printed manuscripts are preferred for the prototype to read. Flat documents are ideal but
since text is found not only on flat documents or books, testing the prototype’s ability to
recognize text on other surface objects that are common to have information, like bottles,
which may widen was concerned, but based on these results, text on convex areas are not
read by the prototype and limits the types of surfaces it reads. Further stated by Engineer
KC Mendoza, the results show that the prototype can easily read flat documents and opened

17
books, widen the array of printed information the user can access with the help of this
device. Although some texts cannot be recognized due to the surface object type, this gives
room for improvement on the prototype itself and knowledge for the user to abstain from
using the device on such surfaces since the recognition of text is ineffective.

Font Type Recognition of Automated Reading Prototype


In the tests shown in table 3, the researchers compared what type of font the
prototype is capable of reading. The prototype was tested on seven (7) variables, three trials
each. Results show that the prototype is capable of converting printed text to audio on bold,
italicized, underlined and handwritten font types.

18
Table 3

Font Type Recognition of Automated Reading Prototype


Font Type Trials Prototype’s Response Result

1 Reads Text Successful

Printed Text 2 Reads Text Successful

3 Reads Text Successful

1 Reads Text Successful

Bold 2 Unrecognized Text Unsuccessful

3 Unrecognized Text Unsuccessful

1 Reads Text Successful

Italicized 2 Reads Text Successful

19
3 Reads Text Successful

1 Reads Text Successful

Underlined 2 Reads Text Successful

3 Unrecognized Test Successful

1 Reads Text Successful

Handwritten (Print) 2 Unrecognized Text Unsuccessful

3 Unrecognized Text Unsuccessful

1 Unrecognized Text Unsuccessful

Handwritten 2 Unrecognized Text Unsuccessful


(Cursive)
3 Unrecognized Text Unsuccessful

1 Unrecognized Text Unsuccessful

Digital Text 2 Reads Text Successful

3 Unrecognized Text Unsuccessful

The prototype easily recognizes and reads texts that are printed. It has difficulty in
reading bold texts as some are too wide and distorted for the camera to recognize and
interpret, however if the boldness of the text is somewhat near to normal print, it can still
be read. Italicized texts are also read by the prototype. The slant of the italicized word does
not hinder the prototype’s recognition of characters. Underlined texts are recognizable so
as long as the underline does not cover the text itself; inasmuch as the line is not thick
covering the location of the text. Some underlines on a text compromises how the
characters of the text look, making it difficult for the device to comprehend such words.

20
Engineer Mendoza explained that the results show that the prototype can read texts that are
printed, not too bold, italicized and underlined. This is greatly beneficial as documents use
bold, italicized and underlined words to indicate emphasis, failure for the prototype to
comprehend these words would indeed be detrimental since the lack of these emphasized
words may affect the information given out in the printed document, by compromising the
paragraph or sentence these bold or italicized words were used in, the understanding of the
user of the given material may be insufficient. Handwritten print texts are difficult for the
prototype to recognize as handwritings vary. Notwithstanding, the prototype can still
perceive those that are written legibly, neatly and close to print fronts. This limits the
prototype in reading specific fonts and only certain handwritings. It cannot recognize
cursive handwriting as these characters may be too complex for the camera to recognize.
Reading digital text may vary, it can be read only if the text is not too small forasmuch as
the light from the gadgets screen does not reflect the camera, causing blurriness and
distortion of text.

Reading machines have already been proven and tested to be able to recognize text
and convert it to audio successfully. Based on the results of this experiment, the automated
reading prototype can also convert printed text like reading machines. However, this
experiment aims to push and test the limits of the prototype to know what other text fonts
it can read. Since text and information are not printed out and written all the same, each
font type and design has its own meaning and purpose. Common fonts like bold, underlined
and italicized are used in a great deal of literature, the prototype had to be tested on such
to determine if it can still comprehend text that are different from standard ones. Failure of
the prototype to read bold, underlined and italicized texts greatly limits the information the
user can access and may even sabotage the user’s understanding of the given text the
prototype may have failed to read. Since the prototype is successful in reading such fonts,
the user no longer needs to worry and look for manuscripts that don’t use bold, underlined
or italicized fonts. This gives the user access to more information in all kinds of literature
using the automated reading prototype. Text is not only limited to printed manuscripts, it
is also found in handwritings and digital text. Handwritten texts may be notes or added
information while digital text are the ones that are found on laptops and other gadgets. The
prototype can read handwritten text in print and some digital text but shows difficulty in

21
recognizing the characters. Since handwritings may vary and letters and notes are common,
while gadgets and digital texts are found everywhere, the text that the user can access with
the use of the device is limited, however if the handwriting is legible and the digital text
not too distorted for the camera, the user may still be able to access this information.
Cursive handwritten text is completely unrecognizable to the prototype, therefore, any
information handwritten in cursive is inaccessible to the user of the automated reading
prototype. Although, handwritten cursive texts are not too common for people to use that
it would cripple the significance of this device.

Figure Recognition Evaluation of the Automated Reading Prototype


In the tests shown in table 4, the researchers evaluated if the prototype can
recognize texts from bar graphs, flowcharts, tree and Venn diagrams. The prototype was
tested on three (3) trials each. Results show that the prototype is incapable of converting
printed text to audio on all figures as there are too much text and may be too confusing.

Table 4

Figure Recognition Capability of A. R. P


Figure Trials Prototype’s Response Result

22
1 Unrecognized Text Unsuccessful

2 Unrecognized Text Unsuccessful

3 Unrecognized Text Unsuccessful


Bar Graph

1 Unrecognized Text Unsuccessful

2 Unrecognized Text Unsuccessful

3 Unrecognized Text Unsuccessful


Flowchart

1 Unrecognized Text Unsuccessful

2 Unrecognized Text Unsuccessful

3 Unrecognized Text Unsuccessful


Tree Diagram

1 Unrecognized Text Unsuccessful

2 Unrecognized Text Unsuccessful

3 Unrecognized Text Unsuccessful


Venn Diagram

The Automated Reading Prototype can read text from a flowchart, but a flowchart
uses arrows and specific orders to make sense, the prototype may still be able to read the
text but it would not make sense to the user. It cannot read venn, and tree diagrams and
graphs as these contain too much words and illustrations that can confuse the prototype
and give it difficulty in recognizing the characters present. If the camera is close to enough
to a specific text and does not have other texts around or near it that may confuse the
prototype, it can still be read. The nature of figures are too complex for the prototype to
read effectively and in such a way that makes sense to the user. Engineer Mendoza

23
explained how the device is only coded to read from left to right so the text in figures may
not be translated properly, and correctly, in accordance to the way the information from
figures are intended to be comprehended.

The way text and information are presented is not limited to books and manuscripts.
Diagrams, Flow Charts and Graphs are also mediums to summarize information in an
easier to understand manner. The ability of the prototype to read such figures are tested to
fully be aware of the device’s potential. Unfortunately, such figures are too complex for
the automated reading prototype to recognize. Such figures have a specific way and order
to be read, so the user can understand the information depicted. The prototype does not
have the capability to read the texts in such a way that it makes complete sense. Being able
to comprehend the information on figures is of great help to the user since these are
presented to make the subject understood better. The accessible information through the
prototype may be limited but these results are references to guide the user of the prototype’s
limits and capabilities, as well as the information the device can understandably convey.

CONCLUSION AND RECOMMENDATION

With the results of the study, the researchers can proudly conclude that the
prototype created functions well and is convenient. The researchers tested the prototype
and made positive results in which the prototype read the given text with the use of the
camera and the earphones connected to the motherboard. The prototype the researchers
created is more effective in reading documents that are 5 - 7 inches away from the camera.
This distance is not too close that some texts are not captured therefore not read, and not
too far that the camera cannot recognize the text. To have a more effective reading, the

24
distance of the camera varies on the texts that are wished to be read. If the text is small and
compact, then putting the camera closer to the text is highly recommended, but if it is too
far and the font size is small, the camera will have difficulty is recognizing the text as the
image will end up being too blurry or distorted. To able the prototype to recognize the
characters of a document, it must be on a flat surface or an opened book. Text written on
convex surfaces serve too distorted for the camera to recognize and interpret correctly.
However, if the text is close enough together on a curved surface, the prototype may still
be able to read it, Fonts that the camera can easily recognize are printed, not too bold,
italicized, underlined and legible printed handwriting. These fonts are not usually too
dilapidated for the prototype to have difficulty in comprehending. The camera cannot
recognize messy handwriting, cursive handwriting, some digital texts and texts that are too
flashy, wide and baroque. Further testing the prototype on figures such as bar graphs,
flowcharts, tree and Venn diagrams, show that these may be too complex for the prototype
to properly read and interpret. Some texts may be read but cannot be understood as the
prototype can only read from left to right and cannot comprehend how the texts in such
figures are meant to be read so the user can understand it.

Nevertheless, to better this study, the researchers would like to suggest on


improving the Optical Character Recognition (O.C.R.) of the prototype so it’s recognition
of characters can be widened and is not limited to specific fonts and can read on surfaces
that are not just flat. Moreover, possible features of this invention that we would like to
recommend are to be able to interpret images as well as objects. Adding on to the
recommendations, a possible feature that could be integrated to the prototype is an addition
of a Basic Artificial Intelligence that could understand, interpret, and break down certain
figures such as, periodic tables, flowcharts, illustrative texts, charts, diagrams and graphs.
Based on The Institute of Electrical and Electronics Engineers Digital Library, it is possible
to pioneer artificial intelligence into the Raspberry Pi 3 motherboard with an affiliated OCR
and TTS codes. This is just one approach in improving the prototype; there may be other
and better approaches, but the researchers are immensely delighted and pleased that the
results are better than what was expected. This study is just a stepping stone on the path to
creating innovative technologies that are greatly beneficial to the people of our society.

25
Acknowledgements
This research would not have been possible without the inspiration and support of
a number of wonderful individuals – the researchers would like to thank and send their
appreciation to all of them being part of this journey and making this research possible.
The researchers owe their deepest gratitude to their Research Adviser; Ma’am Cynthia
Parrocha and Teacher; Ma’am Cynthia Parrocha. Without their enthusiasm,
encouragement, support, and continuous optimism this research would hardly have been
completed. The researchers would also also express their warmest gratitude to their
Research Engineer, KC Mendoza. His guidance into the world of building information
modelling has been valuable input for this research.

The researchers would want to express their gratitude to our batchmate, Jariel Que
for supporting the researchers and extending his help especially in coding the prototype
and Marianne Hojilla, a relative for helping, inspiring and continuously guiding them
throughout the study.

26
The researchers gratefully acknowledge the contributions of their friends; Jade
Pimentel, Vherannie Muelle, Franchesca Liceralde, and other close comrades for the
wonderful support and times the researchers have shared with them, especially during the
Regionals Research Competition. They are deeply thankful to their fellow students whose
challenges and productive critics as it provided new ideas to the work.

They also gratefully acknowledge the contributions of their teachers, Ma’am


Mary Ann Daculog, Sir Tim Sudcalen, Albert Bayawak and all research advisers who
were instrumental in developing the prototype and the manuscript itself. Their expertise
in building information modelling and commitment towards this project was a significant
influence in shaping many of the concepts presented in this research.

Finally, it is the researchers deep and sincere gratitude to their family for their
continuous and unparalleled love, help, and support. They are grateful for their mothers’
unending support; Mrs. Dulay, Mrs. Cachola, Mrs. Yabut, Mrs. Macawaris and Mrs.
Hojilla. The researchers are forever indebted to their parents for giving them the
opportunities and experiences that have made the researchers who they are. They
selflessly encouraged the researchers to explore new directions in life and seek their own
destiny and preferences. This journey would not have been possible if not for them and
the researchers dedicate this milestone for them.

27
REFERENCES

Journal Articles
Chougule, P., Goel, A., Khatavkar, S. Patil, A., & Sehrawat A. (2018). Raspberry Pi 3
Based Reader for Blind People. International Research Journal of
Engineering and Technology (IRJET), 05(06), pp. 1639-1642. Retrieved from
https://www.irjet.net/archives/V5/i6/IRJET-V516306.pdf
Hamad, A.K. & Kaya, M. (2016). A Detailed Analysis of Optical Character Recognition
Technology. International Journal of Applied Mathematics, Electronics and
Computers, 4 (Special Issue), pp. 244–249. Retrieved from
http://dergipark.gov.tr/download/article-file/236939
Kaushik V., Mohith, P., Rangoonwala, H., & Samiappan, D. (2017). Text to speech
conversion module. International Journal of Pure and Applied Mathematics, 115
(Special Issue), pp. 389-395. Retrieved from https://acadpubl.eu/jsi/2017-115-6-
7/articles/6/53.pdf

28
Shetake, P. S., Patil, S.A., & Jadhad, P.M. (2014). Review of text to speech conversion
methods. International Journal of Industrial Electronics and Electrical
Engineering, 2 (8), pp. 29-35. Retrieved from
http://pep.ijieee.org.in/journal_pdf/11-74-140689395029-35.pdf

Books
Berchmans, B. & Kumar, S. S. (2014). Optical character recognition: An overview and
an insight. Institute of Electrical and Electronics Engineers Digital Library, pp. 1-
3. doi: 10.1109/ICCICCT.2014.6993174

Electronic References
Bix, A. (2018). Automation. World Book Advanced. Retrieved from
https://www.worldbookonline.com/advanced/article?id=ar039000
Demott, K. (2017). February is low vision awareness month. Retrieved from
https://nei.nih.gov/news/briefs/five-innovations-harness new-technologies-people-
visual-impairment-blindness
Lapkin, E. (2018). Understanding your child’s trouble with reading. Retrieved on March
from https://www.understood.org/en/learning-attention-issues/children-reading-
issues/understanding-your-childs-trouble-with-reading
Texas Council for Developmental Disabilities. (2013). Visual Impairments. Retrieved
from http://www.projectdealonline.org/v/visual-impairments/

Online Dictionaries
Automated [Def.1]. (n.d.). Merriam-Webster Online. In Merriam-Webster. Retrieved
September 10,2018, from http://www.merriam-webster.com/dictionary/citation.
Automated [Def.1]. (n.d.). Cambridge Dictionary Online. In Cambridge Dictionary.
Retrieved September 10,2018, from
http://www.dictionary.cambridge.org/definition/english/automated.

Unpublished Materials:

29
Agbezree, L. (2013) Library Facilities and their use in Ghanaian Secondary School.
Case study of St. Peters. Unpublished MA dissertation, University of Ghana,
Legon.
Choi, C., Gat-eb J. & Macawaris, L. (2017). Water activated flashlight. Unpublished
research thesis, University of Baguio Science High School, Baguio City,
Philippines.

APPENDICES
Appendix A
SCIENCE HIGH SCHOOL
General Luna Road, Baguio City Philippines 2600

Telefax No.: (074) 442-3071 Website: www.ubaguio.edu E-mail Address:


ub@ubaguio.edu
Letter to the Adviser
August 29, 2018

MS. CYNTHIA PARROCHA


Adviser
University of Baguio Science High School

Dear Ma’am Cynthia,

Greetings!
We, the grade 10 students of the University of Baguio Science High School are
conducting a physical research in partial fulfilment of the requirements for Research 10
(S.Y 2018-2019) entitled Automated Reading Prototype (A.R.P.). We are writing to
humbly request your service and expertise to serve as an adviser for our thesis. We
believe that your knowledge and insights will be valued and greatly enrich our work. We
appreciate your assistance and support in this particular research endeavor.

Thank you for your consideration and we hope you will be able to fulfill our request.

30
Respectfully yours,

The Researchers:
ALLYSON KAE L. CACHOLA BIANCA LAINE C. DULAY

EMMANUEL NATHALIE M. ELE NATHAN ROCHE M. HOJILLA

IVANNE CRES V. YABUT


Noted by:
ROVERLYN SUMEG-ANG
Research Instructor

Conforme:
CYNTHIA PARROCHA
Adviser, University of Baguio Science High School

Appendix B

Program Code
import os
import pygame
import pygame.camera
if os.path.exists (“/home/pi//Desktop/pic.jnp”):
os.remove(“/home/pi/Desktop/pic.jpg”)
os.remove(“/home/pi/Desktop/ocr.txt”)
else:
pass
ocr = “cd /home/pi/Desktop/”
ocr2 = “tesseract pic.jpg ocr”
tts = “espeak -f ocr.txt
error = “espeak “Unable to recognize text, please try again.”
pygame.camera.init()
cam = pygame.camera.Camera(“/dev/video0”,(640,480))
cam.start()

31
img = cam.get_image()
pygame.image.save(img,”pic.jpg”)
os.system(ocr)
os.system(ocr2)
if os.stat(“ocr.txt”).st_size == 0:
os.system(error)
else:
os.system(tts)

Appendix C
Cost Analysis
Table 5 below shows the costs and amount of hours that have been put into the prototype
over the past year. Based on this, the researchers can evaluate the costs and benefits in the
way described by the methodology above.

Table 5

Aggregated Amount of Money

MATERIALS COST SOURCE

RASPBERYY PI 3 PHP 2,880.00 Shopee.com


MOTHERBOARD WITH
CASE

USB CAMERA PHP 438.00 Lazada.com

POWERBANK PHP 350.00 Lazada.com

32
WRIST MOUNTED PHP 139.00 Lazada.com
GLOVE

SHOULDER STRAP PHP 69.00 Lazada.com

SD CARD PHP 297.00 Shopee.com

Total PHP 4,173.00

Appendix D
Schematic Diagram

33
Appendix E

Timetable for Research

34
Prior to Study Conceptualizing the prototype

Beginning of Chapter 1

Identification and approaching of research analysis

Week 1 (August 1-7, 2018) Deadline of Chapter 1

Deadline of Letter to the Advisor

Beginning of Chapter 2

Deadline of Chapter 2

Week 2 (August 8-14, 2018) Beginning of Chapter 3

Beginning of construction of the prototype

Week 3 (August 15-21, 2018) Finalization of construction of the prototype

Beginning of testing of the prototype

Week 4 (August 22 -28, 2018) Deadline of Chapter 3

Beginning of Chapter 4

Week 5 (August 29- September 10, Deadline of Chapter 4


2018)

Editing of previous chapter

Permutation to IMRAD format

Week 6 (September 11-14,2018) Proofreading of article by research advisor

Week 7 (September 15 – 18, 2018) Final alterations before due date

September 19, 2018 Passing of the completed thesis with the write format

35
Appendix F
Curriculum Vitae

Allyson Kae Liwanag Cachola


#168 Sycamore Alley Asin Road Baguio City

Personal Information
Date of Birth April 21, 2002
Place of Birth Baguio City, Philippines
Citizenship Filipino
Civil Status Single
Provincial Address Baguio City, Philippines

Education
Secondary University of Baguio Science High School
Baguio City, Philippines
May 2019
Elementary
University of Baguio Laboratory Elementary school
Baguio City, Philippines
March 2015

36
Bianca Laine Castor Dulay
#35 Hamada Subdivision, Baguio City

Personal Information
Date of Birth September 11, 2002
Place of Birth Marietta, Georgia, United States of America
Citizenship Dual citizenship of Filipino and American
Civil Status Single
Provincial Address Baguio City, Philippines

Education
Secondary University of Baguio Science High School
Baguio City, Philippines
May 2019
Elementary
Torrens Primary School
Canberra, Australia
Northridge Academy Incorporation
Baguio City, Philippines
March 2015

Emmanuelle Nathalie Macawaris Ele

37
#338-E Happy Homes, Brgy. Ferdinand, Campo Sioco Baguio City

Personal Information
Date of Birth December 24, 2002
Place of Birth Iligan City, Northern Mindanao, Philippines
Citizenship Filipino
Civil Status Single
Provincial Address Baguio City, Philippines

Education
Secondary University of Baguio Science High School
Baguio City, Philippines
May 2019
Elementary
University of the Immaculate Conception
Davao City, Philippines
Berkeley School
Baguio City, Philippines
March 2015

38
Nathan Roche Macawaris Hojilla
#7 Spruce Richgate Montecillio Rd. Camp 7 Baguio City

Personal Information
Date of Birth December 27, 2002
Place of Birth Manila, Philippines
Citizenship Filipino
Civil Status Single
Provincial Address Baguio City, Philippines

Education
Secondary University of Baguio Science High School
Baguio City, Philippines
May 2019
Elementary
Berkeley School
Baguio City, Philippines
March 2015

Ivanne Cres Valena Yabut

39
BLK 4 L11 Bhutan Street, Shangrila Village Baguio City

Personal Information
Date of Birth May 24, 2003
Place of Birth Baguio City, Philippines
Citizenship Filipino
Civil Status Single
Provincial Address Baguio City, Philippines

Education
Secondary University of Baguio Science High School
Baguio City, Philippines
May 2019
Elementary Saint Louis School (Inc.) Center Elementary School
Baguio City, Philippines
March 2015

40

Вам также может понравиться