Distortion Analysis Of Tamil Language Characters Recognition

Similar documents
2. Problem formulation

Chord Classification of an Audio Signal using Artificial Neural Network

Contour Shapes and Gesture Recognition by Neural Network

CHAPTER-9 DEVELOPMENT OF MODEL USING ANFIS

Various Artificial Intelligence Techniques For Automated Melody Generation

Automatic LP Digitalization Spring Group 6: Michael Sibley, Alexander Su, Daphne Tsatsoulis {msibley, ahs1,

Smart Traffic Control System Using Image Processing

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Music Composition with RNN

Automatic Laughter Detection

Deep Neural Networks Scanning for patterns (aka convolutional networks) Bhiksha Raj

Predicting the immediate future with Recurrent Neural Networks: Pre-training and Applications

DISPLAY WEEK 2015 REVIEW AND METROLOGY ISSUE

Broken Wires Diagnosis Method Numerical Simulation Based on Smart Cable Structure

Avoiding False Pass or False Fail

Automatic Laughter Detection

A combination of approaches to solve Task How Many Ratings? of the KDD CUP 2007

Identifying Table Tennis Balls From Real Match Scenes Using Image Processing And Artificial Intelligence Techniques

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

LEARNING AUDIO SHEET MUSIC CORRESPONDENCES. Matthias Dorfer Department of Computational Perception

Planning Tool of Point to Poin Optical Communication Links

Reconfigurable Neural Net Chip with 32K Connections

Improving Performance in Neural Networks Using a Boosting Algorithm

Machine Vision System for Color Sorting Wood Edge-Glued Panel Parts

Detecting Musical Key with Supervised Learning

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

Doubletalk Detection

Scan. This is a sample of the first 15 pages of the Scan chapter.

WHAT'S HOT: LINEAR POPULARITY PREDICTION FROM TV AND SOCIAL USAGE DATA Jan Neumann, Xiaodong Yu, and Mohamad Ali Torkamani Comcast Labs

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

Objectives. Combinational logics Sequential logics Finite state machine Arithmetic circuits Datapath

A COMPUTER VISION SYSTEM TO READ METER DISPLAYS

An Improved Fuzzy Controlled Asynchronous Transfer Mode (ATM) Network

Robert Alexandru Dobre, Cristian Negrescu

A Bayesian Network for Real-Time Musical Accompaniment

Audio Compression Technology for Voice Transmission

Speech and Speaker Recognition for the Command of an Industrial Robot

Elasticity Imaging with Ultrasound JEE 4980 Final Report. George Michaels and Mary Watts

Music Emotion Recognition. Jaesung Lee. Chung-Ang University

Efficient Implementation of Neural Network Deinterlacing

IDENTIFYING TABLE TENNIS BALLS FROM REAL MATCH SCENES USING IMAGE PROCESSING AND ARTIFICIAL INTELLIGENCE TECHNIQUES

International Journal of Engineering Research-Online A Peer Reviewed International Journal

Neural Network Predicating Movie Box Office Performance

FPGA Implementation of Convolutional Encoder And Hard Decision Viterbi Decoder

Robust Transmission of H.264/AVC Video using 64-QAM and unequal error protection

VLSI System Testing. BIST Motivation

A Fast Alignment Scheme for Automatic OCR Evaluation of Books

Robust Transmission of H.264/AVC Video Using 64-QAM and Unequal Error Protection

RedEye Analog ConvNet Image Sensor Architecture for Continuous Mobile Vision

Optimization of Multi-Channel BCH Error Decoding for Common Cases. Russell Dill Master's Thesis Defense April 20, 2015

Testing Digital Systems II

Digital Lock-In Amplifiers SR850 DSP lock-in amplifier with graphical display

Halal Logo Detection and Recognition System

Auto classification and simulation of mask defects using SEM and CAD images

Attacking of Stream Cipher Systems Using a Genetic Algorithm

Digital Correction for Multibit D/A Converters

Optimizing Fuzzy Flip-Flop Based Neural Networks by Bacterial Memetic Algorithm

4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER. 6. AUTHOR(S) 5d. PROJECT NUMBER

Paulo V. K. Borges. Flat 1, 50A, Cephas Av. London, UK, E1 4AR (+44) PRESENTATION

Automatic Construction of Synthetic Musical Instruments and Performers

Jazz Melody Generation from Recurrent Network Learning of Several Human Melodies

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Defect detection and classification of printed circuit board using MATLAB

Hidden Markov Model based dance recognition

Design for Test. Design for test (DFT) refers to those design techniques that make test generation and test application cost-effective.

hit), and assume that longer incidental sounds (forest noise, water, wind noise) resemble a Gaussian noise distribution.

Reconstruction of Ca 2+ dynamics from low frame rate Ca 2+ imaging data CS229 final project. Submitted by: Limor Bursztyn

(Received September 30, 1997)

On the Characterization of Distributed Virtual Environment Systems

Jin-Fu Li Advanced Reliable Systems (ARES) Laboratory. National Central University

Fault Diagnosis of Mixed-Signal Analog Circuit using Artificial Neural Networks

MUSICAL INSTRUMENT RECOGNITION WITH WAVELET ENVELOPES

Synchronous Sequential Logic

Music Information Retrieval with Temporal Features and Timbre

MODULE 3. Combinational & Sequential logic

Off-line Handwriting Recognition by Recurrent Error Propagation Networks

WINTER 15 EXAMINATION Model Answer

Implementation of a turbo codes test bed in the Simulink environment

Hugo Technology. An introduction into Rob Watts' technology

SMART VEHICLE SCREENING SYSTEM USING ARTIFICIAL INTELLIGENCE METHODS

Constant Bit Rate for Video Streaming Over Packet Switching Networks

Data flow architecture for high-speed optical processors

Automatic Polyphonic Music Composition Using the EMILE and ABL Grammar Inductors *

Color Image Compression Using Colorization Based On Coding Technique

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

Classification of Musical Instruments sounds by Using MFCC and Timbral Audio Descriptors

Primitive segmentation in old handwritten music scores

Doctor of Philosophy

Introduction to Digital Signal Processing (DSP)

Design of a Low Power Four-Bit Binary Counter Using Enhancement Type Mosfet

Implementation of CRC and Viterbi algorithm on FPGA

Long and Fast Up/Down Counters Pushpinder Kaur CHOUHAN 6 th Jan, 2003

MAHARASHTRA STATE BOARD OF TECHNICAL EDUCATION (Autonomous) (ISO/IEC Certified)

Gender and Age Estimation from Synthetic Face Images with Hierarchical Slow Feature Analysis

Applying Bacterial Memetic Algorithm for Training Feedforward and Fuzzy Flip-Flop based Neural Networks

Designing for the Internet of Things with Cadence PSpice A/D Technology

Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1

Design and Implementation of High Speed 256-Bit Modified Square Root Carry Select Adder

Singer Traits Identification using Deep Neural Network

DISTRIBUTION STATEMENT A 7001Ö

Transcription:

www.ijcsi.org 390 Distortion Analysis Of Tamil Language Characters Recognition Gowri.N 1, R. Bhaskaran 2, 1. T.B.A.K. College for Women, Kilakarai, 2. School Of Mathematics, Madurai Kamaraj University, Madurai. ABSTRACT This research work demonstrates how character recognition can be done with a back propagation network and shows how to implement this using the MATLAB Neural Network toolbox. This is a slightly enhanced version of the character recognition application based on the MATLAB Neural Network toolbox. In this research article we are focusing on the distortion analysis of Tamil Language Characters in order to recognize them effectively using the neural network we have developed. We have used the commonly used representation method for recognizing digits and uppercase English letters as suggested by Guyon et.al [1] to start with and built a method for Tamil language letters over it. KEYWORDS: Neural Network, Natural Language Characters recognition, character shape representation, character classification, Pattern Recognition, Tamil Language Characters, Tamil Desktop Publishing, Feed forward network, Distortion 1. Introduction The word "recognition" plays an important role in our lives and understanding actually begins with proper recognition. It is a basic property of all human beings; when a person sees an object, he or she first attempts to gather all possible information about the object and compare its properties and behaviors with the existing knowledge stored in the mind. If one finds a proper match, one recognizes it. If no match is found, it is recognized as a new object. However, when pattern recognition has to be employed for identification of fixed and known set of patterns in a commercial or research environment, it is more appropriate to have an automated way. A machine that reads checks in a bank could do the job several times faster than human beings. In effect, machines that can read symbols could provide cost effective alternative to manual methods. This kind of application saves time and money, less error prone, and relieves human labor to work on other interesting jobs rather than doing a mechanical boring job. A neural network is an information processing system. It consists of massive simple processing units with a high degree of interconnection between each unit. The processing units work cooperatively with each other and achieve massive parallel distributed processing. The design and function of neural networks simulate some functionality of biological brains and neural systems. The advantages of neural networks are their adaptive-learning, selforganization and fault-tolerance capabilities. For these outstanding capabilities, neural networks are used for pattern recognition applications. Some of the best known neural models are back-propagation, high-order nets, time-delay neural networks and recurrent nets. 2. Design of the Neural Network The twenty-six by thirty five-element input vectors are defined in the MATLAB script file as a matrix of input vectors called alphabet. The target vectors are also defined in this file with variable called targets. Each target vector is a 26-element vector with a 1 in the position of the letter it represents, and 0 s in other positions. For example, the letter C is to be represented by a 1 in the third element (as C is the third letter of the alphabet), and 0 s in other positions. The network receives the 5 7 real values as a 35- element input vector. It is then required to identify the letter by responding with a 26-element output vector. The 26 elements of the output vector each represent a letter. To operate correctly, the network should respond with a 1 in the position of the letter being presented to the network. All other values in the output vector should be 0. In addition, the

www.ijcsi.org 391 network should be able to handle noise. In practice, the network does not receive a perfect letter as input. Specifically, the network should make as few mistakes as possible when classifying vectors with noise of mean 0 and standard deviation of 0.2 or less. We have used the commonly used representation method for recognizing digits and uppercase English letters as suggested by Guyon et.al [1] 3. Network Architecture Feed-forward networks often have one or more hidden layers of sigmoid neurons followed by an output layer of linear neurons. Multiple layers of neurons with nonlinear transfer functions allow the network to learn nonlinear and linear relationships between input and output vectors. The linear output layer lets the network produce values outside the range 1 to +1. The network is a two-layer network. The neural network needs 35 inputs and 26 neurons in its output layer to identify the letters. The log-sigmoid transfer function at the output layer was picked because its output range (0 to 1) is perfect for learning to output Boolean values. The hidden layer has 10 neurons. This number was picked by through several tests by changing the number of neurons and also the number of epochs and by changing the sum squared error to enhance learning. vector takes on a value of 1, and all others have a value of 0. The result of this post-processing is the output that is actually used. 3.1 Initialization The two-layer network is created with newff with logsig function and by creating a set of target and input vectors. 3.2 Network Training In order to create a network that can handle noisy input vectors it is imperative to train the network on both ideal and noisy vectors. To do this, the network is first trained on ideal vectors until it has a low sumsquared error. Then, the network is trained on 10 sets of ideal and noisy vectors each. The network is trained on two copies of the noise-free alphabet at the same time as it is trained on each noisy vector. The two copies of the noise-free alphabet are used to maintain the network s ability to classify ideal input vectors. However after, the training described above the network may have learned to classify some difficult noisy vectors at the expense of properly classifying a noise-free vector. Therefore, the network is again trained on just ideal vectors. This ensures that the network responds perfectly when presented with an ideal letter. All training is done using back propagation with both adaptive learning rate and momentum with the function traingdx. 3.3 Training without Noise The network is initially trained without noise for a maximum of 5000 epochs or until the network sumsquared error falls beneath 0.1. 3.4 Training with Noise Fig1: Architecture of the network The network is trained to output a 1 in the correct position of the output vector and to fill the rest of the output vector with 0 s. However, noisy input vectors may result in the network not creating perfect 1 s and 0 s. After the training, the output from the network is passed through the competitive transfer function compet. This makes sure that the output corresponding to the letter most like the noisy input To obtain a network not sensitive to noise, we trained with two ideal copies and two noisy copies of the vectors in alphabet. The target vectors consist of four copies of the vectors in target. The noisy vectors have noise of Standard Deviation 0.1 and 0.2 added to them. This forces the neuron to learn how to properly identify noisy letters, while requiring that it can still respond well to ideal vectors. To train with noise, the maximum number of epochs is reduced to 300 and the error goal is increased to 0.6, reflecting that higher error is expected because more vectors (including some with noise), are being presented.

www.ijcsi.org 392 3.5 Training Again Without Noise Once the network is trained with noise, it makes for good judgment to train it without noise once more to ensure that the system goes back to its original state and ideal input vectors are always classified correctly. Therefore, the network is again trained with code identical to the Training without Noise section. 3.6 Estimating the System Performance The reliability of the neural network pattern recognition system is measured by testing the network with hundreds of input vectors with varying quantities of noise. The network is tested with a noisy version of the Tamil Language Characters. The MATLAB script file tests the network at various noise levels, and then graphs the percentage of network errors versus noise. Noise with a mean of 0 and a standard deviation from 0 to 0.5 is added to input vectors. At each noise level, 100 presentations of different noisy versions of each letter are made and the network s output is calculated. The output is then passed through the competitive transfer function so that only one of the 26 outputs (representing the letters of the alphabet), has a value of 1. The number of erroneous classifications is then added and percentages are obtained. The solid line on the graph of Fig 6 and Fig 7 shows the reliability for the network trained with and without noise. The reliability of the same network when it had only been trained without noise is shown with a dashed line. Thus, training the network on noisy input vectors greatly reduces its errors when it has to classify noisy vectors. The network did not make any errors for vectors with noise of standard deviation 0.0 or 0.05. When noise of standard deviation 0.2 was added to the vectors both networks began making errors. For example we create a noisy version (Standard Deviation 0.2) of the Tamil Language Characters. The Network is trained with distortions and a testing of networks with noise level of 0.00 to 0.50 in steps of 0.05 is done for all the input characters of the input character given. The Fig 2 and Fig 4 are the result of distorted characters of Tamil language with a maximum sum squared error output of 0.5. Fig 3 and Fig 5 is the result with a maximum distortion of 0.5 given as a test input and the character recognized by the network. Fig 2: Tamil letter tha distorted Fig 3: Tamil letter tha recognized Fig 4.Tamil letter zha distorted

www.ijcsi.org 393 If a higher accuracy is needed, the network can be trained for a longer time or retrained with more neurons in its hidden layer. Also, the resolution of the input vectors can be increased to a 10-by-14 grid. Finally, the network could be trained on input vectors with greater amounts of noise if greater reliability were needed for higher levels of noise. Fig 5.Tamil letter zha recognized The graphs in Fig 7 and Fig 8 show the comparison of percentage of recognition errors of training the network with and without distortion respectively and for varying values of number of epochs. The diagrams and the graphs demonstrate how a simple pattern recognition system could be designed. Note that the training process did not consist of a single call to a training function. Instead, the network was trained several times on various input vectors. In this case, training a network on different sets of noisy vectors forced the network to learn how to deal with noise, a common problem in the real world. Normally, only feed-forward networks are used for pattern recognition. Feed-forward means that there is no feedback to the input. Similar to the way that human beings learn from mistakes, neural networks also could learn from their mistakes by giving feedback to the input patterns. This kind of feedback would be used to refine and reconstruct the input patterns and make them free from error; thus increasing the performance of the neural networks. Of course, it is very complex to construct such types of neural networks. These kinds of networks are called as auto associative neural networks. As the name implies, they use back-propagation algorithms. 4. Conclusion Fig 6: Percentage of recognition errors Pattern recognition could be achieved using both normal computing techniques and neural networks. In normal computing conventional arithmetic algorithms are used to detect whether the given pattern matches a known one. It is a straightforward method. It will say either yes or no. It does not tolerate noisy patterns. On the other hand, neural networks can tolerate noise and, if trained properly, will respond correctly for noisy patterns. Neural networks may not perform miracles, but if constructed with the proper architecture and trained correctly with good data, they give amazing results in pattern recognition. References : Fig 7: Percentage of recognition errors [1] I.Guyon.P.Albrecht,Y.Le.Cn, J. Denker, and W.Hubbard. Design of neural network character recognition for a touch terminal, Pattern Recognition, 1991.

www.ijcsi.org 394 [2] Gerald Wheatley- Applied Numerical Analysis [3] A. Mitiche and J. Aggarwal. Pattern category assignment by neural networks and the nearest neighbors rule. International Journal on Pattern Recognition and Artificial Intelligence,10 :393 408, 1996. [4] R. Plamondon. A model-based segmentation framework for computer processing of handwriting. Proceedings of the Eleventh IAPR International Conference on Pattern Recognition, pages 303 307, 1992 [5] K. Bhattacharyya and K. K. Sarma, "Innovative Segmentation of Handwritten Text in Assamese Using Neural Network", The 2009 International Conference on Genetic and Evolutionary methods (GEM'09), Las Vegas, July 2009. [6] Z. Chen, "Handwritten Digits Recognition", The 2009 International Conference on Genetic and Evolutionary Methods (GEM'09), Las Vegas, July 2009. [7] Tamil Standard Code for Information Interchange, http://www.tamil.net/tscii [8] S. Mori, H. Nishida, H. Yamada, "Optical Character Recognition", John Wiley & Sons, 1999. [9] R. Gonzalez, R. Woods, and S. Eddins, Digital Image Processing UsingMatlab, Prentice Hall, 2004. Gowri N. completed M.Sc. in Physics and obtained M.Phil. in Physics from Alagappa University, Karaikudi for her thesis on Simulation Studies in Crystal Growth. She is currently employed in TBAK College for Women in Kilakarai as Associate Professor in Computer Science Dept. and doing her Ph.D. in Pattern Recognition in School of Mathematics, Madurai Kamaraj University, Madurai. Dr. R. Bhaskaran is Chariperson and Head of Department, School of Mathematics, Madurai Kamaraj University, Madurai.