Introduction attern recognition could be formally defined as categorization of input data into identifiable classes via extraction of significant features or attributes of the data from the background of irrelevant detail. A pattern class is a category, determined by some given common attributes. When a set of patterns falling into disjoint classes are available, it is desired to categorize these patterns into their respective classes through the use of some automatic device. It is important to note that learning or training takes place only during the design (or updating) phase of a pattern recognition system. Once acceptable results have been obtained, the system is engaged in the task of actually performing recognition on samples drawn from the environment in which it is expected to operate. The main objective of this research is to investigate and develop a general approach formally from a new theory based on Structural or Syntactic Pattern Recognition. The problem of pattern recognition is divided into the following sub problems: 1. Image pre processing 2. Object Description 3. Classification # D The pre-processing also known as low level image processing is performed on the input image to improve the quality of the image and to simplify the image for further processing. The pre-processing steps involve noise removal, conversion of gray scale and colour images into binary images and the extraction of contour from the binary image. The object description module takes an input contour image and gives a vector of direction and length called a knowledge vector as output. This plays an important role in the whole process as the knowledge vector gives more information about the objects present in the image which is used to characterize the pattern. We use a syntactic approach for the description of objects. The knowledge vector obtained here gives information about the direction and the length of lines in the contour of objects which can then be given as input to a classifier module. The normalization of this vector plays an important role in the classification process. The vector has to be analyzed and normalized in such a way that it better suits for more variance of objects. The classification module takes the normalized vector as input and identifies them as a member of one of the predefined classes depending on the set of attributes that they hold. The design of a classifier involves the selection of the classifier and the estimate of parameters for the classifier. The feedforward artificial neural networks with back-propagation learning algorithm could be used for classification. Details of such neural networks could be obtained from standard literature. of noise in the image. The pixel point processing techniques such as binary contrast enhancement, stretching or shrinking and histogram equalization are used to improve the contrast of the image. The input image to the system may be a colour image, a gray scale image or a binary image. These images have to be converted to binary images for further processing. Segmentation is an important process in our object recognition system as it highlights the objects present in the image from its background. The function im2bw() in matlab is used to convert the images to binary. The graythresh() function computes a global threshold (level) that can be used to convert an intensity image to a binary image with im2bw. Level is a normalized intensity value that lies in the range [0, 1]. The contours are extracted as they give the outlines of shapes of the objects present in the image. The function bwtraceboundary() in matlab is used for finding the contours. The initial point of the object is identified, and the outline tracing algorithm is used. bwtraceboundary() traces the outline pixels of an object in binary image bw. Nonzero pixels belong to an object and zero pixels constitute the background. Figure 2 shows a sample binary image and its contour. The notion of 'Picture Description Languages' refers to the language of digital images defined over a regular array of vertices. Having extracted the contour of an image this process involves running through the contour pixel by pixel and performing some form of calculations using every pixel and its surrounding pixels. This involves finding the length of the contour in every possible direction. The eight possible directions are two horizontal directions, two vertical directions and four diagonal directions. As the lines other than straight lines have more variations in their directions, the length of the descriptor becomes large. It is then reduced to a single occurrence of eight directions and normalized to fixed number of pixels for making it invariant. Following is the knowledge vector of a hexagon. # GENERATION OF DIRECTION LENGTH CODE (DLC) The connectivity between the pixels is an important concept used in the field of object recognition. Two pixels are said to be connected, if they are adjacent in some sense and their intensity levels satisfy a specified criterion of similarity. If pixel p with coordinates (x,y) and pixel q with coordinates (s,t) are pixels of object (image subset) S, a path from p to q is a sequence of distinct pixels with coordinates (x0, y0), (x1, y1),???.., (xn, yn) where (x0, y0) = (x,y) and (xn, yn) = (s,t), (xi, yi) is adjacent to (xi-1, yi-1), 1 ? i ? n, and n is the length of the path. If there a path exists from p to q consisting of pixels in S, then p is said to be connected to q. The gap of one or two pixels marks the presence of the next component in the same object or the beginning of the next object which can be then found by analysing the relationship among the components. With this background, we shall see now as to how to apply Look Ahead Tracer (LAT) algorithm to various image types and extract knowledge from them. Let (x,y) be the initial point. The algorithm continues to find the immediate pixel connected to (x,y) in one of the 8 directions. In this technique, a previously recognized direction during a scan is given the first priority status when compared to the other preferential ones. If R is the identified direction, then the other preferential directions are DR, D, DL, L, UL, U, UR. The trace continues until there is no pixel connectivity to the current point (x,y), after removing the One can easily say that the above knowledge vector represents a square from the fact that the four directions R,D,L and U has the same length except the direction U that has 1 pixel less than other 3 sides. This is because of the immediate removal of the scanned pixels. The starting pixel at position <104,109> is removed after its scan. # III. # KNOWLEDGE VECTOR ANALYSIS The direction codes in the knowledge vector obtained by tracing the contour of an object are R, DR, D, DL, L, UL, U and UR. For better processing, let us categorize these directions into basic directions R,D,L and U and diagonal directions DR,DL,UL and UR. For example, consider the knowledge vector of a square in figure 5. R100*D100*L100*U99 Figure 5 : The knowledge vector of a square This shows the square consists of four basic directions with all lengths equal. Shapes with all the basic directions but with equal alternate length can be identified as a rectangle. Shapes with four sides consisting of only diagonal directions can be considered as rhombus. The task of identifying regular shaped objects is simple. Objects in a real time environment may not be regular always. For the effective identification of those objects, the shapes of the objects can be approximated to some regular polygons to which it is closer to. This is possible by predefining the classes of regular shapes with some pre conditions set and approximating the other shapes to one of these shapes by identifying them as a member of the predefined class. It is important to note that, the diagonal sides present in the shape of an object do not necessarily appear alone as the connection of only diagonal pixels. The occurrence of other pixels depends on the angle at which the diagonal lines appear in the image. Possibly D and R pixels appear more with DR, D and L pixels appear more with DL, U and L pixels appear with UL and U and R pixels appear with UR. For example consider a line with the direction code DR. As shown in fig: if the line bends more towards right the more R pixels appear, in the same way if it bends more towards down the more D pixels present with DR as shown in figure 6. In this respect, when the lineis rotated with a small change in the slope, they can be approximated to one of these shapes. Four basic sides or four diagonal sides with equal alternate lengths. In the same manner, shapes with more sides can be predefined. When the shape of an object appears it can be then classified by artificial neural networks as a member of one of these classes. # THE SHORT COMINGS OF TRADITIONAL PATTERN RECOGNITION SYSTEMS AND THE NEED FOR ADAPTIVE PROCESSING Objects in the real world from an image or image sequence of the world are detected with the advanced software tools the imaging applications like Automatic Object Recognition (AOR) that uses different object models which are known a-priori. Humans can perform this task of object recognition effortlessly and instantaneously. However, such a task of implementation based on machines is performed using algorithms and has been very difficult in networks for real time imaging problems which are non-linear in nature. In some applications, the patterns/objects to be recognized are so fuzzy that they cannot be modelled with conventional tools. In order to solve these neural network processors can be used as the best tool because of the fact that the processor can build a recognition engine from simple image annotations made by the programmer. It then extracts the characteristics or feature vectors from the annotated objects and sends them to the neural network. Neural networks having adaptive processing elements are capable of performing generalization and can consequently classify situations never seen before by associating them to similar learned situations. # a) Artificial Neural Networks Learning from a set of examples is an important attribute needed for most pattern recognition systems. Artificial neural network is an adaptive system being widely used in pattern recognition systems that changes its structure based on external or internal information that flows through the network. Artificial neural networks are adjusted, or trained, so that a particular input leads to a specific target output. The neural network design part consists of two processes, training and application. The training of the neural network continues until the mean squared error reduces to a certain threshold or until the maximum number of iterations is reached. Once training is completed, the network can be applied for the actual classification of the data. The classification technique used may be one of the following: 1. Supervised classification -in which the input pattern fall as a member of a predefined class. 2. Unsupervised classification -in which the pattern falls into an unknown class as there are no predefined classes. The learning or training takes place only during the design phase of a pattern recognition system. Once the results obtained are satisfactory, the system is ready to perform the task of recognition on samples drawn from the environment in which it is expected to operate. b) Feed-forward neural networks Feed-forward networks are commonly used for pattern recognition. A three-layer feed forward neural network is typically composed of one input layer, one output layer and one hidden layer. In the input layer, each neuron corresponds to a given input pattern while in the output layer each neuron corresponds to a predefined pattern. Once a certain sample is input into the network, he best situation is that the output will be a vector with all elements as zero only except the one corresponding to the pattern that the sample belongs to. Of course, it is very complex to construct such types of neural networks. The commonly used networks for minimizing the cost are multi-layer-feed-forward neural networks, which uses the back-propagation learning algorithm for training neural networks. and transmits it to all units in the hidden layer z j (j=1,?.,p). 5. Each hidden unit z j sums its weighted input signals and applies its activation function and sends this signal to all units in the output layer y k (k=1,?.,m). 1![Figure 1 : Description Object Recognition System a) Pre-Processing To improve the quality of the noisy images we get from the real time environment image enhancement and restoration techniques can be used. Pixel neighbour processing techniques such as low pass filtering, high pass filtering or median filtering are used for the removal](image-2.png "Figure 1 :") 2![Figure 2 : a. Binary Image & b. Contour b) Object Description](image-3.png "Figure 2 :") ![<69,98>/R64*D1*DR1*D2*DR1*D1*DR3*D1*DR1*D1 *DR1*D2*DR1*D1*DR1*D1*DR1*D1*DR1*D1*DR3*D 2*DR1*D1*DR1*D1*DR1*D1*DR1*D1*DR1*D1*DR1* D2*R1*DR1*D1*DR1*D1*DR1*D1*DR1*D1*DR1*D2* DR1*D1*DR3*D1*DR1*D1*DR1*DL1*D1*DL1*D2*DL 3*D1*DL1*D1*DL1*D1*DL1*D1*DL1*D2*DL1*D1*DL 3*D1*DL1*D1*DL1*D1*DL1*D2*DL1*D1*DL1*D1*DL 3*D1*DL1*D2*DL1*D1*DL1*D1*DL1*D1*DL1*D1*DL 1*D1*DL2*D1*DL1*D1*DL1*D1*DL2*L63*U1*UL1*U1 *UL1*U1*UL1*U2*UL1*U1*UL3*U1*UL1*U1*UL1*U1* UL1*U2*UL1*U1*UL1*U1*UL3*U1*UL1*U1*UL1*U2* UL1*U1*UL1*U1*UL1*U1*UL3*U1*UL1*U2*UL1*U1* UL1*U1*UL1*U1*UL1*U1*UL2*U1*UL1*U1*UL1*UR1 *U1*UR3*U1*UR1*U2*UR1*U1*UR1*U1*UR1*U1*UR 1*U1*UR3*U2*UR1*U1*UR1*U1*UR1*U1*UR1*U1*U R1*U2*UR3*U1*UR1*U1*UR1*U1*UR1*U1*UR1*U2* UR1*U1*UR3*U1*UR1*U1*UR1*U1*UR1*U2*UR1*/< 70,98># This can then be reduced to single occurrence of eight directions as <69,98>/R64+1*DR31*D58*DL33*L63*UL32* U57*UR32*/<70,98># and then normalized to ? R17*DR16*DL17*L17*UL16*UR17 So, the final vector is a normalised vector consisting of a single occurrence of eight directions with their length, which can then be fed into an artificial neural network for further process of classification. II.](image-4.png "") ![Year pixels which are already traced, thus producing a knowledge string component.By 'knowledge', we mean the direction code and length code of each contour component in the image.](image-5.png "") 34![Figure 3 : Direction Codes](image-6.png "Figure 3 :Figure 4 :") 6![Figure 6 : Slant lines with small changes in the slope © 2013 Global Journals Inc. (US)](image-7.png "Figure 6 :D") ![This section shows the approximation and the normalization of the Knowledge vector of different shapes to produce the input vector for aneural network.knowledge vector obtained is <71,167>/D109*L96*UR3*R1*UR2*U1*UR2*U1*UR1 0*U1*UR4*U1*UR6*R1*UR1*U2*UR6*U1*UR4*R1*U R2*U2*UR7*U1*UR3*R1*UR1*U1*UR2*U1*UR3*UR8 *U1*UR4*U1*UR5*R1*UR2*U2*UR6*U1*UR4*R1*UR 1*U1*UR1*U1*UR2*/<72,166># Vector limited to single occurrence of eight directions <71,167>/R6*D109*L96*U19*UR89*/<72,166># Vector approximated and normalized to 100 pixels ? D34*L30*UR36 <104,109>/R81*D81*L81*U80*/<105,109># <104,109>/R81*D81*L81*U80*/<105,109># ? R25*D25*L25*U25 <92,72>/R104*D63*L104*U62*/<93,72>#/<273,254 ># <92,72>/R104*D63*L104*U62*/<93,72>#/<273,254 ># ? R31*D19*L31*U19 © 2013 Global Journals Inc. (US) <86,102>/R25*D1*DR7*D1*DR1*R1*DR6*D1*DR2*D 24*DL1*L1*DL5*L2*D1*DL4*L1*DL3*D1*DL1*L2*DL1 Global Journal of Computer Science and Technology Volume XIII Issue II Version I UL2*U1*UL7*U1*UL3*L1*UL4*U28*R1*UR2*R1 *UR1*R2*UR5*R1*UR2*R1*UR1*R2*UR2*/<87,102> # Vector limited to single occurrence of eight directions <86,102>/R25+9*DR16*D24+5*DL15*L25+7*UL16* U28+2*UR13*/<87,102># Vector normalized as ? R14*DR13*D14*DL12*L14*UL11*U15*UR10 <81,138>/R11*DR1*R4*DR1*R2*DR1*R2*DR4*R1*D R2*R1*DR5*D1*DR3*D1*DR2*D1*DR1*D1*DR1*D2* DR1*D2*DR1*D3*DR2*D3*DR1*D16*DL1*D4*L1*DL1 *D3*DL1*D2*DL1*D2*DL1*D1*DL1*D1*DL1*D1*DL4* D1*DL5*L1*DL2*L1*DL4*L3*DL1*L1*DL1*L5*DL1*L1 0*UL1*L4*UL1*L2*UL1*L1*UL3*L1*UL1*L1*UL3*L1* UL4*U1*UL1*U1*UL1*U1*UL1*L1*UL1*U1*UL1*U2* UL1*U1*UL1*U3*UL1*U3*UL1*U3*UL1*U15*UR1*U4 *UR1*U3*UR1*U3*UR1*U1*UR1*U2*UR1*U1*UR1*R 1*UR1*U1*UR1*U1*UR1*U1*UR4*R1*UR3*R1*UR2* R2*UR3*R2*UR1*R4*/<82,137># Vector limited to single occurrence of eight directions <81,138>/R32*DR26*D45*DL25*L33*UL24*U48*UR2 3*/<82,137># Vector normalized as ? R4*DR20*D6*DL20*L4*UL21*U6*UR20 IV.](image-8.png "") ![i. Back-Propagation Algorithm Multi-layer feed-forward networks have been used as powerful classifiers. The training of backpropagation algorithm involves 4 stages © 2013 Global Journals Inc. (US) 1. Initialization of weights 2. Feed forward stage 3. Back propagation of errors 4. Updating of weights and bases During the feed-forward stage each layer in the network calculates its activation value and passes it to Global Journal of Computer Science and Technology Volume XIII Issue II Version I Year the layers in the next level. The neurons in the output layer produce the output of the network and compare it with the target output to determine the error. During the back-propagation stage the error is back propagated from the output layer to each layer in the previous level for the correction of the adjustable parameters. This process repeats until the error reaches a minimum threshold or until the maximum number of iterations performed. The Back propagation algorithm is discussed below. Parameters used x = (x 1 ,x 2 ??,x i ,?.,x n ) -Input training vector t = (t 1 ,t 2 ,??..,t k ,?.,t m ) -Output target vector ? k = error at output unit y k ? j = error at hidden unit z j ? = learning rate v ij = weights of input layer v oj = bios on hidden unit j z j = activation of hidden unit j w jk = weights of hidden layer w ok = bios on output unit k y k = activation of output unit k 1. Initialize the weights to small random values 2. While the stopping condition is false, do steps 3 to 10 3. For each training pair do steps 4 to 10 4. Each input receives the input signal x i (i=1,?..,n)](image-9.png "") 6![Each output unit y k (k=1,?,m) sums its weighted input signals and applies its activation function to calculate the output signals 7. Each output unit y k receives a target pattern t k corresponding to an input pattern and calculates the error term as 8. Each hidden unit z j sums its delta inputs from units in the layer above. The error information term is calculated as, 9. Each output unit y k updates its bias and weights. The change in weight is given by And the bias correction term is given by Therefore, and 10. The hidden unit z j updates its bias and weights. The weight correction term is given by and the bias correction term is Therefore, and 11. Test the stopping condition. V. RESULTS Neural network with 8 neurons in the input layer and 9 neurons in the output layer each representing 9 different classes of shapes is trained with a set of input patterns. The fastest back-propagation algorithm known as Levenberg-Marquardt back propagation (trainlm) present in matlab toolbox is used to train the neural network. trainlm is a network training function that updates weight and bias values according to Levenberg-Marquardt optimization. Fig. 5.7. shows the neural network created for our case of problem. Fig. 5.8 shows the performance of the neural network for the given set of training and test patterns.](image-10.png "6 .") 7![Figure 7 : Neural Network with 8 neurons in the input layer and 9 neurons in the output layer](image-11.png "DFigure 7 :") 1 1RDRDDLLULUURClass DNormalized Vector Codes for Object Recognition using Artificial Neural Networks in the Framework of Picture Description Languages DNormalized Vector Codes for Object Recognition using Artificial Neural Networks in the Framework of Picture Description Languages DNormalized Vector Codes for Object Recognition using Artificial Neural Networks in the Framework of Picture Description Languages In the same way the DLC of real time images also can be approximated and normalized to fixed number of pixels. Such code gives us the input vector that can be given to the neural networks for the classification of their shapes as the shape in one of the 9 classes. Later more analysis can be done for further classification of these shapes inside the classes to produce more detailed classes for the actual identification of objects. ## CONCLUSION In this Paper a syntactic approach is proposed for shape based object recognition. The knowledge vector is reduced to the input vector to a neural network by some vector approximation and normalization processes. The research efforts during the last decade have made significant progresses in both theoretical development and practical applications. The method presented here may offer a promising solution for object recognition problem. * <64,184>/R4*DR2*R2*D1*DR1*R2*DR4*R2*DR1*R1 * *R1*DR6*R1*DR2*R1*DR1*D1*DR1*R2*D1*DR1*D1 * *UL4*U1*UL1*L2*U1*UL1*L1*UL1*U3*UR5*R1*U R1*U1*UR1*R1*UR1*U1*UR2*R1*UR5*R1*UR1*R2* UR2*R1*UR3*R1*UR2*R1*UR1*R2*UR1*R2*UR2*R2* UR1*R2*UR2*R1*UR1*U1*UL1*U2*UL2*U3*UL1*U3* UL2*U5*UL1*U20*UR1*U8*UR1*R11*UR1*R26*DR1* R7*DR1*R5*DR1*R5*DR1*R3*DR1*R2*UR3*D1*U2* UR2*U1*UR3*U1*UR4*R1*UR1*U2*UR1*R1*UR2*U1 * *UR1*U1*UR2*R1*UR1*R2*UR1*/<65,183># <64,184>/R26+160*DR74*D19+91*DL82*L29+153* UL74*U20+94*UR77*/<65,183># References Références Referencias * Shape matching and object recognition using shape contexts SBelongie JMalik JPuzicha IEEE Transactions on Pattern Analysis and Intelligence 4 24 April 2002 * 3-D recognition and shape estimation from image contours FSCohen J.-YWang in Proc * IEEE Conf Computer Vision Pan. Recognition June 1992 * Cellular Logic Array Processing Techniques for High-Throughput Image Processing Systems EGRajan The Indian Academy of Sciences 18 2 June 1993 Invited paper, SADHANA, Special Issue on Computer Vision * Two-Dimensional Packing For Irregular Shaped Objects PingChen ZhaohuiFu AndrewLim BrianRodrigues 0-7695-1874-5/03 $17.00 Proceedings of the 36th Hawaii International Conference on System Sciences -2003 the 36th Hawaii International Conference on System Sciences -2003 IEEE 2003 * Efficient shape matching using shape contexts GregMori Member SergeIeee Belongie IeeeMember JitendraMalik Member Ieee IEEE Transactions on pattern analysis and machine intelligence November 2005 27 * Intelligent Object and Pattern Recognition using Ensembles in Back Propagation Neural Network THafiz MuhammadUHassan KashifKhalid Imran International Journal of Electrical & Computer Sciences IJECS-IJENS 10 6 * Neural network-based face detection HRowley SBaluja TKanade IEEE Transactions on Pattern Analysis and Machine intelligence 20 1998