Recognizing Ancient South Indian Language Using Articial Neural Network Associate With Opposition Based Grey Wolf Optimization Algorithm

: Recognizing signs and fonts of prehistoric language is a fairly difficult job that require special tools. This stipulation makes the dispensation period overriding, difficult, and tiresome to calculate. This paper presents a technique for recognizing ancient south Indian languages by applying Artificial Neural Network (ANN) associated with Opposition based Grey Wolf Optimization Algorithm (OGWA). It identifies the prehistoric language, signs and fonts. It is apparent from the ANN system that arbitrarily produced weights or neurons linking various layers plays a significant role in its performance. For adaptively determining these weights, this paper applies various optimization algorithms such as Opposition based Grey Wolf Optimization, Particle Swarm Optimization and Grey Wolf Optimization to the ANN system. Performance results have illustrated that the proposed ANN-OGWO technique achieves superior accuracy over the other techniques. In test case 1, the accuracy value of OGWO is 94.89% and in test case 2, the accuracy value of OGWO is 92.34%, on average, the accuracy of OGWO achieves 5.8% greater accuracy than ANN-GWO, 10.1% greater accuracy than ANN-PSO and 22.1% greater accuracy over conventional ANN technique.


Introduction
Several of the rock engravings that are found at various areas of the sphere expose the particulars of indulgence, regime, financial circumstance, principles, and managerial protocols trailed by numerous leaders and empires precise to those areas [1].Recently, there has been a fast developing concern in the study on more varied languages and writings [2]. Tamil is one of the noteworthy languages in Dravidian Language Family. It has an opulent script basis start from 300BC [3]. Hence, it is essential to develop a new scheme which can identify the prehistoric Tamil fonts accurately and can alter them into contemporary Tamil fonts [4]. The prehistoric Tamil fonts i.e. the initial and primitive Tamil can be normally found in rock engravings and palm leaves. Only epigraphists can examine those rock engravings. To spread the readability and to defend the old realistic abilities, we require a decent appreciation scheme that can translate the prehistoric writing to contemporary writing [5].
There are several schemes available for printed script identification. When the plans formation is over, the test image is input into the scheme. The system splits the image into separate letters and each slice of the image is related with the preloaded patterns. A related technique is revised for every fragmented letter in the image [6]. Handwritten letter identification is a computer facilitated process that enables inferring and reading readable handwritten inputs provided on ancient scripts like palm leaves, rock engravings, etc [7]. The development of imaging technology has swayed a few regions in the computer apparition community [8]. The process of prehistoric Tamil letter identification comprises removal of some categorized qualities named sorts to categorize an unidentified letter into one of the well-known classes. Pre-processing is fundamentally made use of to reduce ranges of handwritten letters. A feature extractor is rudimentary for creative data depiction and removing significant features for future dispensation. A classifier allocates the letters to one of the few classes [9]. A prehistoric Tamil letter identification procedure centred on fake resistant was suggested to advance the rate of letter identification [10]. ANN is an example of such an agenda that alters the masses of its networks from the training forms. These masses verifiably work as structures for cataloguing [11]. Sánchez et al (Sánchez, Melin, & Castillo, 2017) have suggested GWO for training segmental neural network with a granulated method for human acknowledgment [12]. Despite the fact that the investigator changes the prehistoric letters up to 90% precision, the translation verses cannot be agreed with the appropriate connotation in the sentence. [13] 1.1 Problem Definition Traditionally, recognizing the ancient language, symbols, and characters indeed required domain experts in that particular context. However, the availability of experts at the right place is not possible most of the time because of inadequate experts in this research field. In some complicated situations, the research experts themselves face complications and consume a long time for recognition. The research desperately required an efficient prediction model to resolve this issue. Hence, this research focuses on utilizing the ANN technique because of its ability to learn and model non-linear and complex relationships. The traditional ANN does not achieve remarkable performance since the randomly generated weight across the ANN layers does not yield appropriate results. Identifying appropriate weights through trial-and-error methods or manually varying all possibilities is monotonous and wastage of time to process.

Research Objectives
The significant research objective is to develop a machine learning model to recognize ancient language, symbols, and characters. The research intends to reconfigure the randomly generated weights to enhance the performance of the conventional ANN model.

Novelty of the work
The main novelty of the work is to recognize the ancient Language, symbols and characters through ANN. OGWO is applied with ANN to identify the appropriate weights. OGWO is chosen for optimization since it includes opposition strategy to increase the chance of predicting appropriate weights over conventional GWO and other optimization techniques.

Related Works
Kavitha B.R. and Srimathi C [14] have exploited the ultra-modern CNN for identifying handwritten Tamil letters in disconnected manner. CNNs differ from conservative procedure of Handwritten Tamil Character Recognition (HTCR) in removing the characteristics accordingly. That work was an effort to fix a yardstick for disconnected HTCR exploiting unfathomable learning methods. That work has generated a training accurateness of 95.16% which was very well compared with the orthodox methods.
Zheng-Ming Gao and Juan Zhao [15] have premeditated an enhanced grey wolf optimization procedure with varied weights (VW-GWO). Outcomes validate that the suggested VW-GWO procedure performs higher than the normal GWO, Ant Lion optimization (ALO), PSO and Bat algorithm (BA). The VW-GWO procedure was furthermore tested in high-dimensional problems.
In Manigandan T et al. [16] work, the engraving images collected from Tamil Nadu Archaeological Department are pre-processed and fragmented. Letters are categorized and established reliant on Vectors mined; exploiting Support Vector Machine (SVM) classifier and the designs of the letter are synchronized with recognized letters and anticipated by means of Trigram method. Hence the suggested agenda can pay attention to the severe problems in reading the engraved images.
Prashant M.Kakde and S.M.Gulhane [17] have proposed that identification of unrestrained (Handwritten) Devanagari letter was increasingly difficult because of the form of essential whacks. In that article, technique like PSO and SVM are applied and associated. An android phone was exploited for taking input letter for indicating the documented Devnagari letter. For linking the android device and MATLAB, they have used PHP language. The PSO technique provides exactness up to 90%.
N.Sridevi and P.Subashini [18] have provided a method to achieve grouping of handwritten prehistoric writings in Tamil. The approach utilises Extreme Learning Machine for grouping of handwritten prehistoric Tamil writings. The performance of Extreme Learning Machine is compared with Probabilistic Neural Networks. From the experiment outcomes it is collected that Extreme Learning Machine provides maximum precision rate of 95%.

Proposed Methodology
The significant intention of this research is to develop a system for recognizing the ancient Language, symbols and characters through Artificial Intelligence (AI) techniques. This research includes ANN to recognize ancient Language, symbols, and characters. This research intends to identify appropriate weights to enhance the performance. The techniques involved in this process are PSO, GWO, and OGWO. Figure 1 demonstrates the overall proposed process of ANN based recognition.

Fig.1 Block diagram of proposed framework 3.1 Training phase
In training phase, characteristics like area, centroid (X-coordinates and Y-coordinates), bounding-box (corners and X-Y width), major axis distance, minor axis distance, eccentricity, orientation, perimeter, equiv diameter are initially removed from the manually fragmented 77-letters (listed in table-3). These 13characteristics are considered as an input factor and their equivalent letter images are considered as output class. Generally, the training procedure performs to conclude applicable masses for effective identification.

Testing Phase
Pre-processing: The trained system receives given query image (RGB) and it pre-processes the image by converting it to a gray-scale, followed by applying a Gaussian filter to remove the noises.
Skew detection: A skew detection technique is used to straighten the image; which results in the image to binary form.
Segmentation based on Region growing: While reading the whole row of a pixel, if a white pixel is noticed and it is greater than 100, it is considered out of line and continues reading the succeeding pixel rows. If the white pixel not gets noticed up, it is considered as in line. A related procedure is made centred on columnwise for separating lyrics, and the above said 13-characteristics are removed for the fragmented word and those values are provided as input for the ANN prototype with organized masses. The presentation of every active method evaluates with exactness amount for two-query investigation case images.
Feature Extraction: Above mentioned 13-features are extracted for recognizing the given query character.

Artificial Neural Network (ANN)
ANN is a mathematical learning technique utilised in cognitive psychology and artificial intelligence. The ANN is an encoded computational prototype that aims to replicate the neural construction and working of the human brain. Figure 2 shows the ANN architecture.

Fig.2 ANN structure
In figure 2, ANN contains three types of layers: input layer, hidden layer, and the output layer. Each layer of the ANN consists of several neurons. In this, every single neuron in the input layer is associated with the concealed layer neuron, and every concealed layer neuron associated with the output layer with an arbitrary weight. In our ANN model, 5 hidden layers are considered with 10 neurons. These arbitrary weights are allocated to every interrelated layer.

Structure initialization
In the initialization process, 13 inputs based on the input layer weight wj and the hidden layer weights wij are initialized.

Input layer
The input layer considers 13-input attributes as mentioned in the previous section. The input layer is the first layer of the neural network in which each input is connected with hidden layer neurons. The inputs I1, I2,…..In are applied to the input neurons u1, u2,….un. Each neuron in the input layer is connected with the hidden layer neurons with random weight w11, w12,…..wij. The basic function of hidden neurons is evaluated based on the equation below, Where Xf is a basics function, wij is an input layer weight and i is a number of input with this basic function the active function is combined. The activation function of hidden neurons is evaluated in this sigmoidal function is exercised based on the equation shown below, Multilayer perception is a sigmoidal activation function in the form of a hyperbolic tangent. The next layer of the ANN is hidden layer.

Hidden layer
In this section, five numbers of hidden layers are revealed and in each hidden layer 10 neurons are performed. The number of neurons in the hidden layer is defined as Hu1, Hu2,….Hun, which are connected with the output layer neuron. Each neurons in the hidden layer is connected with the output layer neurons with random weight w11, w12,…..wij

Output layer
The hidden layer neurons are connected with the output layer and each connection has a weighted value such as w1, w2,…..wn. The recognised character is the output of ANN. For optimizing the weight, OGWO is employed.

Opposition based Grey Wolf Optimization Algorithm (OGWO)
When compared to traditional GWO, in OGWO, the opposition strategy utilizes to increase the probability of finding optimal weights to enhance the ANN predicting performance. The grey wolves sufficiently mount a Canidae's piece family and are respected as the top marauders presenting their place at the sustenance's food chain. They regularly display an disposition to make due as a set. The results produced by the alpha are distributed on to the set. The Beta says to the second rank in the striking sequence of the grey wolves. They are, fundamentally, supplementary wolves that sufficiently provide certain aid to the alpha in the choice making or equivalent set purposes. The omega, which is at the least bands of the grey wolf pack, by and great works as an ancillary proposing into the additional principal wolves very closely on every event and is allowed to have just the slight leftovers resembling a great doddle by the leader wolves. In our technique, the alpha (α) esteemed as the most suitable arrangement with a perspective to replicating the social pecking order of wolves logically while conceiving the OGWO. Thus, the second and the third-best arrangements are named as beta (β) and delta (δ) separately. The remaining applicant arrangements are regarded to be the omega (ω). In the OGWO method, the hunting (optimization) is guided by the α, β, δ, and ω.

.1 Initial and opposition based solution generation
The preliminary step in the optimization process; here, the solutions is generated randomly based on constraints. The randomly generated solutions range from [-100 to 100] and length of a solution is 280. The length of the solution is mathematically expressed as follows, Where NI is the number of inputs (features: 13), NH is the number of hidden neurons; considered 20 (userdefined). Along with the preliminary random solution generation, this research includes opposition based solution generation, which would increase the probability of identifying an optimal weight as output solution. The mathematical expression of opposition-based solution generation is as follows.
Xi ϵ X1, X2, X3 …. NP, are initial solution randomly generated suggest to generate opposition based solution where x and y are minimum and maximum values.

Fitness computation
This process is utilized to evaluate the fitness of randomly and opposition based solution generation with the aid of following mathematical function.
Where, Xi is known as input and α & β are called weights.

Updating the position
We assume that the alpha (best candidate solution), beta, and delta have improved knowledge about the potential location of the prey to reproduce the hunting behavior of the grey wolves mathematically. As a result, we hoard the first three best solutions attained so far and require the other search agents (including the omegas) to revise their positions according to the location of the best search agent. For repetition, the new solution c (t+1) estimated by using the formulae mentioned below.
To have hyper-spheres with different random radii, the arbitrary parameters A and C help the candidate solutions. Examination and usage are ensured by the adaptive estimations of A and a. With diminishing A, half of the iterations are committed to the investigation (/A/<1) and the other half are dedicated to the usage. Enclosing the conduct, the subsequent equations are utilized keeping in mind the end goal to give numerical model. The OGWO has only two main parameters (A and C) to be adjusted. However, we have kept the OGWO algorithm as simple as possible, with the fewest operators to be adjusted. The process will be continued until the termination is obtained.

Overall algorithm
The entire steps involved in the process of ANN-OWGO based character recognition is given in the following algorithm. _____________________________________________________________ ANN-OWGO Algorithm __________________________________________________________________ 1. Get character database for ANN-training process 2. Extract 13-Features from the characters 3. 13-features are utilized as an input for ANN training 4. OGWO associate with ANN for identifying appropriate weights 5. Here, opposition strategy associates with conventional GWO as OGWO to increase the chance of finding appropriate solutions (weights). 6. Finally, ANN will recognize the character for a given features value 7. The above mentioned steps are the fundamental of following testing process 8. The trained ANN model receives aw query image 9. For pre-processing. the trained system received a given query image (RGB) and pre-processed the image by converting it to a gray-scale, followed by applying a Gaussian filter to remove noises 10. A skew detection technique is used to straighten the image; results in the image to binary form. 11. Segmentation: line segmentation, while reading the entire row of a pixel, if white pixel detected and it is higher than 100, which is considered for line and keeps reading the subsequent pixel rows if the white pixel not gets detected up to that is regarded as the line. A similar process is done based on column-wise for segregating words, and the above said 13-features are extracted for the segmented word and those values given as input for the ANN model with configured weights. The performance of each employed technique evaluates with accuracy measure for two-query test case images. _____________________________________________________________________

Results and Discussion
This section discusses the different investigations with results from different techniques. Apart from OGWO, the other optimization techniques used in ANN weight determination are PSO and GWO. Accuracy and the convergence graphs are explained in the below sections.. In table-1, sample test case 1, table-2 sample test case  2 and table-3 input features extracted from concern characters are presented.    figure 5, carryout different optimization technique involves in designing the ANN structure namely OGWO, GWO, and PSO amid OGWO converge faster than other at the same time fitness error reveal by the proposed OGWO also good compare with contest technique. The straight competition between PSO and GWO up to 100 th iteration, then GWO varies low than PSO, and both saturate error value at 350 th to 500 th iteration. OGWO slightly saturate up to 400 th iteration to 500 th iteration, and the error value is lower than other two algorithms due to the optimization of ANN weights using OGWO. Employing optimization techniques to identify appropriate weights and integration of opposition strategy in GWO ensure the proficient performance of OGWO associate ANN

Conclusion
Resolving the research problem through redesigned ANN associated with optimization techniques accomplished effectively. Incorporation of opposition strategy in conventional GWO literally enhances the performance even better over other techniques for recognizing the ancient language, symbols, and characters. The results unveil that the proposed method attains accuracy as 93.6%, which is 5.8% greater than GWO, 10.1% greater than PSO, and 22.1% superior to conventional ANN model (Auto-configured weights). In the future, this research intends to reconfigure the structure of ANN to enhance predicting performance further.