Modified HOG Descriptor-Based Banknote Recognition System

This research presents a banknote recognition and counterfeit detection system that proposes a pre-processing approach before creating a new feature set by extracting Speeded-Up Robust Feature, creating image patches from its vertices and then computing the Histogram of Gradient descriptors from cells within the patch boundaries. A feature reduction approach is also designed based on a high correlation and Low variance filter to reduce the feature set by calculating the intra-feature value variability to ascertain the similarity between the elements to prune out highly correlated values and variances using a fixed threshold. For experimental evaluation, additional analysis was conducted using Principal Component Analysis, Bag of words, Low Variance filter, and High Correlation Filter method. Furthermore, the classification was performed using 10-fold cross-validation based One-Versus-One Support Vector Machine. Additionally, the developed technique was tested on counterfeit banknotes to authenticate genuine and fake currencies. Lastly, point matching method was applied to detect our suggested feature vectors to match images containing multiple notes to distinguish and determine the value of the bills.


Introduction
Technological advancement has made life more comfortable, but it also has its downside as criminals have taken advantage of this progression. Initially, only printing houses had access to the technology needed to produce banknotes but today with the aid of a personal computer and a good laser printer people can duplicate paper money easily in their residence. Banknote counterfeiting is one of the core crimes on earth. The Apex bank of every country is responsible for providing the security features and in Nigeria it is the responsibility of the Central Bank of Nigeria (CBN). Every Nigerian banknote contains several security features that can be used to identify and recognize different note denominations [1].
Over the years, differentiating between fake and real currency banknotes has become quite tricky due to the fast-scientific progress in color printing, cloning, and imaging and because fake notes are now reproduced via top-notch technology that uses security paper [2]. In the same view, counterfeit banknote detection has become an essential task in billing machines. Its implemented by using image analysis methods (transmittance analysis, reflectance analysis) with different light spectra (visible, infrared and ultraviolet). The drawback of this system is its high cost and its usage with only Automated Teller Machines (ATM), where an elevated level of reliability is required. Anti-counterfeiting banknotes security features can be classified into three categories: printing, ink and substrate features. Printing features like intaglio printing, micro letters, serial number, signatures, Guilloches, Ink features such as UV glowing ink, pearl ink, infrared ink, color-shifting inks, and substrate features like planchets, security thread, see-through registers, and watermark [3]. This research presents a paper money recognition system that includes multiple banknote identifications and Counterfeit note detection that could aid in reducing banknote counterfeiting offenses and positively influence a country's economy. The rest of the article is divided into the following sections: Section 2 discusses the research background and surveyed literature. In Section 3, the proposed system including its architecture, design, and stages of the approach is studied. Section 4 debates the analysis and discussion of the results of the system functionality, and the conclusion of the research and future work is given in Section 5.

Background & Literature Survey
Banknotes are utilized in every business venture in as much as electronic currencies is becoming identified as a substitute legal tender, the note is still dominant and central in the day to day dealings. Currency Recognition and detection is a critical problem in computer vision. Although, electronic and online currencies exist banknote is still the preferred option in daily transactions such ASTESJ ISSN: 2415-6698 as vending machine, the banking system, banks, shopping mall, railway ticket counter, ATM, vending machines, and currency exchange service. Furthermore, there are cases of money bundles withdrawn from ATMs contained a few counterfeit bills although banks are unwilling to accept these omissions. Even though this problem is majorly unavoidable, they are not also able to provide all branches with the needed counterfeit detection system [4].
One of the most straightforward methods that have been applied in paper money recognition system is the use of visible features found in notes like the size, texture, and color of banknotes. All banknotes have pictures and patterns impressed on them, such as portraits of historical persons or images of landmark buildings which are characterized as textures extracted as features using texture-modeling techniques. The drawback of this approach is that banknotes are mistreated so could get damaged and torn due to regular use leading to alteration of the notes thereby hurting the recognition rate [5]. On the other hand, [6] extracted the geometric features (length and width) and employed it in classifying bills with varying sizes such as Italy and Chinese banknotes. Although, this approach was simple and fast; it had significant limitations including its effectiveness for slant, stain, and tear. Furthermore, it was unsuitable for some banknote such as the US Dollars, which had several denominations of the same size.
Furthermore, many countries employ colors for differentiating banknotes although they could wear or tear it still retains its color [7]. [8] designed a system that identified the denomination of the Indian currency note based on its color. Likewise, another author also developed a technique that converted the image from RGB color space to HSV color space and then extracted the color feature. Their justification for employing the HSV color space was because of its nearness to the human conceptual understanding of color [9]. Additionally, a researcher recommended an approach based on characteristic point extraction of a unique part of the Euro paper currency that possesses the same color has also been conducted [10]. Similarly, [11] devised a US banknote counterfeit detection system that extracts its color feature and classifies the currency using fuzzy Hamming distance. Despite its high performance, if the system were tested on Euro banknotes. It would not be able to detect forgeries since its detection is based only on the color feature. Most countries have different colors distributed over the surface of its banknotes including the Nigerian Naira; all have only one dominant color. For instance, the dominant color of 10 Naira note is "Red." Although, some banknotes have the same dominant color due to wear and tear the color quality may degrade; hence leading to difficulties in using the color feature for automatic recognition. Also, using only the color as a discriminative feature may fail because some banknote denomination does have the same dominant color [12].
Likewise, [13] proposed a banknote classification technique based on discrete real-valued wavelet transform (DWT), they calculated the feature vector using the statistical properties of decomposed coefficients from the approximation and detail subbands. The technique was applied successfully in spatial and frequency domain, although it does not shift invariant and has no phase information. Another researcher also devised a new method called a hyper-complex wavelet transform (CWT) that is shift invariant and has a single phase. However, it also had its limitation like its ambiguity when used for 2-D signals [14]. Furthermore, an invariant shift extension of DWT and CWT called the quaternion wavelet transform (QWT) was also adopted. The technique had three phases and one magnitude, the first two stages were on the local image shifts, and the third focuses on image texture information [15]. [16] also suggested a new feature extraction system using the quaternion wavelet transform (QWT) that combined the partial and total Hilbert Transform (HT) using quaternion algebra to form one magnitude and three phases. It captures the statistical characteristics of QWT coefficients using the generalized Gaussian density (GGD) and classifies the banknote using the neural network. Furthermore, [17] presented a banknote recognition system that used a wavelet transform for extracting features and classified the features using Multi-layer perceptron (MLP) neural network; the method experimented on Persian paper currency. In another research work, a feature extraction technique was developed based on the Discrete Wavelet Transform (DWT). They employed Daubechies 1 (DB1), and the approximate coefficient matrix of the transformed image is extracted from the estimated efficient model and used for recognition, classification, and retrieval of currency notes [18].
Additionally, [19] formulated an approach that uses the edge and corners of an object in detection, the system was invariant to image rotation but was not able to handle issues associated with image scale. [20] then presented another technique that could automatically determine the scaling of the objects, it computed the Hessian matrix and used the Laplacian operator to locate the block like feature points in the image. Similarly, [21] also conducted further research on automatic scaling selection and implemented a Harris-Laplace detector in addition to the Hessian-Laplace to improve the automatic scale selection process. [22] also developed a SIFT method that finds the limits of the scale space in an object. The technique was invariant to scale and rotation and was very useful in extracting interest point descriptors which can be utilized in matching and classifying objects. [23] devised a method for the classification of Slovak banknotes based on SIFT detector and descriptors. Although, the approach posed its issues which were its computational complexity leading to high processing time limiting its acceptability in real-time applications.
In the same view, [24] also developed a colored SIFT approach that combines the color feature and the geometric data using the SIFT algorithm. The Mikolajczyk and Schmid framework was further applied to normalize introduced a nonlinear SVM method that juxtaposed images of different color space in addition to combining color space descriptors to evaluate the accuracy of the approach [25]. On the other hand, speeded up robust features (SURF) was created, the process was designed to improve on the SIFT method, the procedure retains the robustness of SIFT but with a decreased descriptor size and shorter processing rate and time.
SURF is also an interest point detector that is robust to scaling and rotation. It creates a unique, repetitive and vigorous process for capturing the interest points. The technique creates an integral image and uses it in place of the original image to diminish the processing time. It then uses the Hessian matrix to detect the highest and lowest second derivative value after this is achieved the interest point orientation is detected and rotated to align with the orientation of the predefined direction to complete the SURF descriptor creation [26]. A researcher applied SURF descriptor in Figure: 1. Proposed Modified Hog-LVHCF based Banknote Recognition System banknote recognition; the system was used to match reference regions with query images. It presented promising results and was robust to handle partial distortions as well as worn or wrinkled bills [27].

Proposed Methodology
In this research, we implemented a procedure for enhancing and cleaning up banknote images. It involves converting to grayscale then employing Gamma Correction and Histogram equalization to improve the image luminance and contrast of the intensity image. Furthermore, Weiner filter was applied to reduce the amount of noise after that the boundaries of the banknote are computed using Blob detection, cropped out and Gaussian pyramid utilized in standardizing the note. We also created a new feature set by extracting the Histogram of Gradient (HOG) Descriptors from image patches generated from the vertices of SURF interest points. After the HOG features have been obtained, the vector size may differ for each image in addition to being computationally demanding during processing. To minimize its effect on recognition/detection rate while maximizing the reduction of Feature dimensionality. A combination of dimensionality reduction techniques based on the descriptor correlation and variance is applied to reduce the Vector scale. A descriptor filter analysis using Pearson's Product Moment Coefficient ρ and variance σ 2 is employed to determine the best set of features for classification.
The approach answers two fundamental questions, determining the relevance of the descriptor to the feature set and its redundancy regarding other associated descriptors. Finally, we applied our technique to several datasets using OVO-SVM and Putative point matching to investigate the efficiency of the descriptor showing the effectiveness of Modified HOG/LVHCF descriptor for banknote identification, counterfeit detection, and multiple notes recognition. Figure 1 shows the phase of the proposed Modified Hog-LVHCF based paper money recognition system.

Image Pre-processing
During image acquisition, the irrelevant background image is also captured. This unwanted information must be removed to identify and recognize the notes. Furthermore, digitized banknote images sometimes have discontinuous particles, noise, illumination, positioning, and unanticipated boundaries. Also, when matching the image, if the banknotes are of varied sizes it could affect the recognition rate. In this research, image preprocessing plays a fundamental step as it involves enhancing and cleaning up the images for efficient feature extraction. Thus, in this research, a new combination of techniques for image preprocessing is presented, and a flowchart of the procedure is shown in Figure 2.
Step 1: The image was first converted to grayscale by calculating the weighted sum of the three channels of the RGB image and representing the result as a matrix whose fragments are assigned values based on how dark or bright the color of the pixel is at the corresponding position. The value allocated to the pixel during grayscale conversion ranges from 0 to 255 representing the brightness of the image.
Step 2: Gamma Correction was applied to improve the image luminance by enhancing the quality of the image in dark or blurry areas while also compressing the intensity of bright regions.
Step 3: Histogram equalization was then employed in enhancing the contrast of the intensity image, the approach redistributes the image histogram to ensure that the brightness value is constant across the image.
Step 4: Weiner filter was applied to reduce the amount of noise present in a signal comparing with an estimate of the desired noiseless signal.
Step 5: Background subtraction was also achieved by first finding all blobs in the image using Blob detection, select the boundary coordinates of the blob containing the banknote and cropping out the note using the detected blob edges.
Step 6: Banknote denominations sometimes are of varied sizes, and the note can also change due to the acquisition manner thus image reduction and standardization were achieved using the Gaussian pyramid.

Feature Extraction (Proposed Modified HOG Descriptor)
Feature extraction is a crucial step in image processing, the selection of features is vital in determining the recognition of an object, and an ideal feature set should be stable, invariant to rotation, scale, and illumination changes. Several feature Figure 2: Block diagram of the procedures undergone in the Image Pre-processing extraction methods have been proposed over the years. In this research, several useful descriptors and feature reduction approaches were investigated and compared to the proposed Modified HOG descriptor and Low Variance High Correlation Filter. In this study, six different descriptors were employed: Scale Invariant Feature Transform, speeded up robust features, gradient location and orientation histogram, Histogram of Gradient, shape context, and the proposed Modified HOG Descriptor. While to reduce the processing time several feature reduction methods were also implemented: Principal Component Analysis, Bag of Words, High Correlation filter, Low Variance Filter, and compared to the developed Low Variance and High Correlation Filter Algorithm. The proposed Modified HOG based descriptor technique is split into four phases: SURF point detection, SURF descriptor extraction, image patches creation from SURF points, and HOG feature computation.

SURF Point Detection
The SURF interest point detection is separated into four phases: integral image generation, calculation of the matrix, construction of the scale space and localization of the feature point.
Integral Image Generation: The Integral image generation was introduced to reduce the computational complexity of the square like convolution filters used in the future, it is mapped from the original image. The integral image is then computed by the sum of all pixels in the input image I within a given rectangle sector which can be deduced by Equation 1.
After the integral image has been calculated, three additions are carried out to compute the summation of the intensities over any four-sided vertical area consequently the computational time is not dependent on the image size using Equation 2.
Computation of the Hessian Matrix: The box filters are used as an approximation of the second-order Gaussian derivatives, this is computed on the integral image calculated earlier rather than the actual image hence reducing the processing time. A 9x9 squarelike filter was employed in the computation of the Gaussian denoted by with a scale of 1.2 which led to the creation of a blob response map over different scales in the image. A set of simple weights w were applied to the square-like sector to reduce the computational overhead.

Construction of the scale space:
The SURF approach is scale invariance, it applies the same filter at different sizes to the image at the same speed in parallel. The length of the dark lobe can only be increased by an even number of pixels to guarantee the presence of a central pixel (top). The scale space is categorized as a series of octaves; an octave is a set of filter response maps gotten by convolving each input image with a filter in different sizes. For an octave layer with a size N * N, the scale σ , the convolution of each layer in the image is computed on the response map using equation 3.
Localization of the interest point: The final step in the interest point detection is localizing the image over the different scales, a threshold is set for removing weak responses in the map. Several thresholds are checked until a middle ground is reached that takes into consideration the number of interest points and the strength of the detected interest point. Furthermore, a 3-dimensional scale space is applied to the non-maximum suppression, and only points bigger than all the neighbors is selected as the given interest points. Finally, scale space interpolation is applied to obtain the best interest points [28].

SURF Descriptor Extraction
SURF Descriptor computation is divided into three phases: Assignment of the interest point orientation, Computation of the descriptor using Haar Wavelet transform and Feature matching using Fast Indexing.

Assignment of the interest point orientation: Assignment of
Interest points orientation is needed to attain rotation invariance, so Haar Wavelet transform is used at each interest point (x, y) to determine the direction of a circular neighborhood in addition to its scale. After computing the wavelet transform the feature point is weighted using a Gaussian filter, the dominant orientation is decided by calculating the summation of all the horizontal and vertical responses within the sliding windows; the most extended vector is then selected as the local orientation vector for the feature point.

Computation of the descriptor using Haar Wavelet transforms:
The previous step used a box area centered around the point of interest in determining the orientation. In this phase, each sector is further divided into a 4x4 square like sector, to improve the reduced the geometric distortions, and localization inaccuracies that could occur the interest pointed are first weighted with a Gaussian filter then the Haar wavelet transform is calculated horizontally and vertically for each 5x5 paced interest points. Furthermore, the computed wavelet transform is then summed for each sector and use as the first level entry in forming the feature vector set. Moreover, to capture the change in intensity after each interest point the absolute value of the responses is also computed to create a 4-dimensional descriptor vector for each sub-sector. Finally, the sectors which are made up of a 4x4 sub-sector is concatenated resulting in a descriptor with a length of 64 dimensions.
Feature matching using Fast Indexing: This phase presents an additional functionality to the matching stage, it uses a fast indexing technique based on the previous computed Laplacian in the Haar wavelet transform stage. The Laplacian value can be used to differentiate lighter blobs on darker areas from darker blobs in brighter areas. Hence, used in removing features that do not share the same type of contrast thus reducing unstable interest points from the feature vector set thus reducing the processing time and as the Laplacian has already been calculated earlier it to does not add any extra computation overhead. Figure 3 presents SURF point descriptors of sample USD and CAD paper money.

Creation of Image Patches
An image patch is designed from polygons defined by the coordinates of its vertices; it consists of one or extra polygons that might or might not be connected. In this research, the 43 x 43 patch is formed by specifying the coordinates of each unique vertex of the SURF square-like descriptors and a matrix that defines how to connect these vertices to create the faces. This technique is preferred for SURF points as its patches multifaceted, so vertices shared by more than one face need be defined only once thereby requiring fewer data to design the patches.

HOG Descriptor Computation
Histogram of Gradient (HOG) feature is a viable feature extraction technique with the ability to select feature points irrespective of the banknote image view. It is based on the histogram of pixel gradients neighbors for image sections or blocks thus it is hardly affected by changes in lighting conditions or image geometry. The HOG feature extraction technique can be grouped into three main stages: Calculation of Gradient for each pixel, computing its Histogram and by combining the histogram of all cells in a block to normalizing the histogram [29]. The feature depicts the gradient dispersion on the different positioning of a banknote image and is conducted to obtain the dimension and texture feature of the banknote. The HOG algorithm partitions all note image into small evenly-dimensioned sections called cells comprise several pixels which individually have a weight computed by getting the gradient magnitude and histogram. The histogram shows the dispersion of the pixel taking into cognizance the orientation of the gradient. Cell histogram is offered as a vector that is merged to get the HOG feature of the overall banknote image [30]. Figure 4 shows the terminology of the HOG feature extraction. Finally, Table 1 shows the HOG method specification that provided the best result in this research. The Cell size of 16, four cells per block, 1 to 1 block overlap, nine bins, and 18x1 bin centers. The number of HOG feature is different for each Dataset because it is chosen based on the feature size that presented the best recognition and detection rate.

Feature Reduction (Proposed LVHCF)
After the HOG features have been extracted, the vector size is computationally demanding. Therefore, to unify and reduce the feature dimension; the techniques aim to minimize its effect on recognition/detection rate while maximizing the reduction of Feature dimensionality. The experiment conducts a descriptor filter analysis using Pearson's Product Moment Coefficient ρ and variance σ 2 to determine the best set of descriptors for classification. The technique answers two fundamental questions: determining which descriptor is relevant or not to the feature set and if the descriptor is redundant or not concerning other related descriptors.
The first question can be answered by just defining a threshold ρ value, but for the second question it is more complicated as it involves analysis of pairwise correlations between all descriptors. To answer both questions before filtering the descriptors we put forward an approach called a Low Variance and High Correlation Filter (LVHCF). In this experiment, it is assumed that if two descriptors are found to be redundant to each other one of them needs to be removed and removing the one that is less relevant to the descriptor-set concept keeps more information to predict the class while reducing redundancy in the data. As presented in Figure 5, given a dataset with M feature set and a descriptor set ds, the algorithm chooses a list of predominant descriptors D best for a given ds. It comprises two phases, the first phase (line 2-17) computes the Pearson correlation moment coefficient p for all pairs of descriptors in the ds using equation 1, extracts relevant descriptors into S1 i list based on the predefined threshold t1. The S1 i list is further processed to remove additional correlated pairs and only keeps predominantly uncorrelated Dq and Dp among all the earlier selected descriptors in the S1 i list (line 13). After one round of filtering all Dq pairs based on Dp, the algorithm will take the currently remaining feature right next to Dp as the new reference (line 18) to repeat the filtering process. The algorithm stops until there is no more feature to be removed from S1 i list . In the second phase (line 19 -38), the variance σ 2 of Append Di to S1 i list; 8 End; 9 Dp = getFirstDescriptor (S1 i list); 10 Dq = getNextDescriptor (S1 i list); 11 Compute ρ for all Dp and Dq 12 If (ρ >= t1) 13 remove Dq from S1 i list; 14 Dq = getNextDescriptor (S1 i list); 15 Else Dq = getNextDescriptor (S1 i list); 16 End until (Dq == NULL); 17 Dp = getNextDescriptor (S1 i list, Dp); 18 End until (Dp == NULL); 19 For End until (Dp == NULL); 39 Sbest = S2 i list; 40 End; Figure 5. Low Variance and High Correlation Filter (LVHCF) Algorithm each descriptor in the S1 i list is gotten using equation 2. Furthermore, the descriptors are extracted into a new list S2 i list based on the predefined threshold t2 and ordered in ascending order according to their σ 2 values. The list is further processed to remove redundant descriptors and only keeps predominant ones among all the selected relevant descriptors. A descriptor Dp that has already been determined to be dominant can always be used to filter out other features that are ranked higher than Dp and have Dp as one of its redundant σ 2 peers. The iteration starts from the first descriptor in S2 i list (line 26) and continues as follows. For all the remaining descriptors, if Dp happens to be a redundant peer to a feature Dq, Dq will be removed from S2 i list . After one round of filtering descriptors based on Dp, the algorithm will take the currently remaining descriptors right next to Dp as the new reference (line 37) to repeat the filtering process. The algorithm stops until there is no more feature to be removed from S2 i list . The descriptors matching the variance remaining in S2 i list is then used as the final dataset B est . Table 2 presents the feature reduction on the four datasets.

Classification
The classification stage is the decision segment of a recognition system, and it uses the features extracted in the previous phase. It involves identifying each banknote and assigning it the correct denomination class. Furthermore, the principal function of a pattern recognition system is to yield decisions concerning the class membership of the models with which it is confronted. Support Vector Machine (SVM) is a classification approach used for separating hyperplanes, it was designed in the 1990's for building nonlinear separating function, and real-valued estimation [31].
The Multi-Class SVM-based system uses a binary classification approach for the one versus one technique, it designs a classifier for each class and creates a hyperplane between the classes. Furthermore, the system assigns a class to a test sample based on a distance margin for the classes. Each data set is split into two groups, the training, and test groups. The system used a randomized approach for data selection, and the class of the test data was predicted based on the established class of the training mechanism. The Multi-Class SVM has two application method approaches; One Versus All (OVA) method and One Versus One (OVO) method. OVA differentiate one class from the remaining classes, and the OVO distinguishes each class from one class at a time while ignoring the remaining class [32].
For this research, the system used the OVO technique in combination with error correcting output codes. The error correcting output code divides a Multi-Class scheme into a double class system that constructs a classifier for each binary division. OVO technique compares all classes against each other and builds a binary classifier to distinguish each pair of class while ignoring the remaining classes. When classifying an unknown image to a class, the classifier conducts a vote, and the class with the highest majority is selected as the best choice to assign the banknote [33]. Further analysis was conducted using K-Nearest Neighbor (KNN), Weighted K-Nearest Neighbor (WKNN), Random Forest (RF) Decision Tree, and multilayer perceptron (MLP) Neural network. In conclusion, it is hard to decide which classification method is the best as research has only been able to prove that a classifier could have a higher recognition and detection rate than others based on a specific problem.

Result & Discussion
This system was implemented in MATLAB as its toolbox provides built-in support for SIFT, SURF, HOG, GLOH in addition to the several classifiers. The hardware utilized comprised of a Core i3, 2.4 GHz processor with 4 GB Ram, the operating system used was Windows 8.0 64Bit and image processing techniques, and other related MATLAB toolboxes employed in the development of the system. This research utilized 10-fold Cross-Validation for the analysis of the data. It is a statistical technique used in evaluating predictive models by dividing the data into training and validation sets. The approach divided the dataset into K-fold equal data sizes and crossed over each K-dataset partition in successive rounds such that each data point has a chance of being in the validation process. The Centre for Pattern Recognition and Machine Intelligence (CENPARMI) has developed a database for Banknote Identification (Experiment 1), Counterfeit Detection (Experiment 2), and Multiple Banknote recognition (Experiment 3). The database comprised of eight classes for the Naira Notes, US Dollars had six classes, five classes for the Canadian Dollars, and seven classes for the Euro Bills for both genuine and counterfeit bills [34,35].

Banknote Recognition (Experiment One)
The obverse and reverse of the banknotes were captured using a Canon T2i digital camera. Banknotes suffering from scratches, wrinkle, stains, wear and tear which represent the image quality of the paper monies in the real-world were captured for this project. The statistical distribution of the Dataset was Nigerian Naira, US Dollar, Canadian Dollar, and Euro data sets. It comprised 1790, 1120, 1000, and 1400 genuine banknotes and 1320, 823, 733, and 1020 fake banknotes respectively. Each dataset included images with multiple notes, clutters, overlapping and partial paper money. Furthermore, the datasets were made up of images captured in four directions for each banknote: the front (upward and downward) and back sides (upward and downward).  Figure 7 measures the performance accuracy using SIFT, SURF, HOG, GLOH, Shape Context, and Modified HOG using KNN, WKNN, RF, MLP, OVO-SVM classifiers. The xaxis shows the accuracy while the y-axis gives the extracted features and the z-axis has the classifiers for the Naira, USD, CAD, and Euro Banknotes. From the results, the best accuracy for the Features and classifiers were SURF/Modified HOG and OVO-SVM while the worst performance was achieved with SIFT and KNN Respectively although the proposed Modified HOG presented comparatively higher accuracy in most cases it was barely higher than the SURF techniques. Furthermore, RF and MLP did produce promising results, but their accuracy was dependent on the descriptor and currency as RF outperformed MLP in some cases and MLP outweighed RF in performance accuracy in other feature and classifiers combination. Although, WKNN outperformed KNN they both still had low rates. Lastly, the most inferior combination accuracy was SIFT and KNN with an average of 82%, and the highest accuracy combination was Modified HOG and OVO-SVM with an average rate of 97.3% with the proposed SURF closely following behind with an accuracy of 97.29%.
Further experiments were conducted with feature reduction techniques to remove features that hinder the accuracy and precision of the proposed approach, and the error rate was utilized in measuring the performance of which is presented in Table 3 above. The table gives the Error Rate of OVO-SVM of the different Descriptors and Feature Reduction on the Four Datasets.
The proposed approach Modified HOG/LVHCF presented the lowest error rate within the average rate of 0.0270 after feature reduction while shape context/PCA had the worse error rate with an average of 0.0585. The rankings of the descriptor and Feature reduction technique ranging from best to worse are Shape Context, SIFT, GLOH, HOG, SURF, Modified HOG and PCA, BOW, LVF, HCF, LVHCF respectively. Additionally, each feature reduction approach was applied to all the descriptors to determine the best pair with the highest recognition rate. From the result, PCA had the lowest acceptance rate across all data sets while Modified HOG/LVHCF presented the best results across all datasets with relatively higher performance than SURF. Regarding datasets, Naira notes offered the best result while CAD had the worse rate although there were cases where CAD outperformed Euro and USD data sets when SIFT, SURF and Shape Context descriptors were employed. Lastly, the proposed approach presented promising results in comparison with existing techniques and can be engaged in descriptor dimension reduction. Figure 8 depicts genuine and fake banknotes employed in this research, the first row consists of Naira notes, the second row has Euro paper money, the third row with the Canadian dollar, and the fourth/last row has US dollar. The dataset covers a wide variety of conditions, such as occlusion, rotation, changes in scaling, illumination, and viewpoints. The table presents the evaluation results of the proposed system on the data sets using OVO-SVM as it presented the best result during classification. For testing the system used 3110 samples for the Naira notes, 1943 instances for US Dollar, 1733 samples for Canadian Dollar and 2420 images of the Euro banknotes. Furthermore, the table also showed the Naira relatively had the highest correctly detected samples while the US dollar had the lowest successfully authenticated notes. However, all data sets had little incorrectly detected samples irrespective of the size of the data set; the system identified all images in the four datasets as paper money still, although some bills were wrongly classified. Finally, The Naira had the highest detection rate, the Euro came second, USA Dollar was third, and the Canadian dollar had the least authentication rate.

Multiple Banknote Recognition (Experiment Three)
For the multiple banknote recognition systems, an object detection method in a cluttered scene using point feature matching of extracted HOG features based on image patches computed from the vertices of SURF points normalized using low variance and high correlation filter. It reads the test image and preprocesses it by applying grayscale conversion, noise removal, background subtraction, and normalization. Furthermore, the HOG Feature points are detected and used in matching the test image containing multiple banknotes is paired with the template image using the putative matching technique. The approach computes the transformation matched points while eliminating outliers. This aids in localizing the object in the scene. Lastly, the system locates the bounding polygon of the detected image and draws a black box to indicate the denomination.   Figure 9 shows multiple banknote samples utilized in this research: a) comprises Naira notes, b) Canadain Dollar notes c) Euro bills, and d) US Dollar. The first row contains two bank notes; the second row contains three banknotes, the third row includes four bank notes, and the last/fourth row has five bank notes. The multiple notes dataset included clutters, distortions, rotations and overlapping captured in different illumination conditions. While Figure 10 presents the database employed for the numerous banknote detection phases of the system, it consisted of 460 Naira images, 372 US dollar samples, 440 Canadian dollar images, and 365 Euro Samples. Table 6 shows the dataset of Naira, Euro, CAD and US notes used in the multiple banknote recognition experiments ranging from images with one banknote to five bills. It shows the precision and recall of this analysis; this was computed to investigate the effectiveness of our systems in distinguishing many banknotes in one image and determine which of the several banknote classes can best predict the note denomination irrespective of distortions. Furthermore, in Figure 11 we have measured the Recognition Rate using One, two, three, four and five notes on the Naira, USD, CAD, and Euro Database. For simplicity, we have shown here only the recognition rate on the x-axis, while the number of the banknotes is fixed on the y-axis and the currency type on the zaxis. The reason for this experiment is to evaluate the effectiveness of the proposed feature extraction approach on multiple banknote recognition. The system recognized several currencies including same notes in the same image, separated, overlapped, cluttered and partially obstructed. After three bills the acceptance result reduced when the number of banknotes was increased. In all cases, the accuracy measure is more than 87%. Although one, two and three notes show more than 92% while four and five ranged between 87.50% and 92.45%. For this research, the best recognition rate across all datasets was three bills using Naira at 98.60%, and the worse was with five banknotes using USD at 87.50%. Table 5. Qualitative Analysis of feature descriptors In Table 5, we empirically compare the performance of all feature descriptors for multiple banknote identifications concerning its variants in five distinct scenarios: speed, scale, rotation, illumination, and affine invariance. The results of the experiment were analyzed qualitatively to give a general idea of the performance of each algorithm in distinct scenarios using four grades: Common, Good, Better, and Best. As shown in the table above, Shape context has significant advancements in speed, scale, affine, and rotation changes. But not illumination change. SIFT is slow; it is invariant to rotation and scale, maintains a stable degree of stability for affine transformation and a low threshold to illumination change. SURF is fast and has better performance than SIFT and GLOH, but it is not stable to the rotation and illumination changes. HOG is robust to rotation, illumination, positioning, and affine modifications but very slow and scale variant. GLOH has excellent performance in scale and rotation, presented reasonable speed and affine changes and had a deficient sense of illumination invariance. The Modified HOG descriptor maintains a proper stability illumination, rotation, and affine transformation. It can also handle reasonable scale invariance and compared with the other algorithms; it has distinct advantages regarding speed. Thus, can be deduced that the proposed technique presented a promising performance in banknote recognition. Lastly, the result of this qualitative analysis is not constant in all cases as changes to the dataset, and classifiers can lead to a different conclusion.

Training Sample Needed
One critical issue is determining the number of notes by denomination and authenticity is needed to ensure a reasonable performance from the proposed technique. To solve this problem, the various databases are evaluated concerning the sizes by varying the training set dimension. A remarkable increment on the recognition and error rate is seen when the training data set increases from 300 to 800 for banknote recognition, 100 to 250 for counterfeit detection, and 50 to 200 for multiple currency recognition. After which growing the data further only produced asymptotical performance improvement but had no effect on its accuracy and error rate, and this could be attributed to the inclusion of more distorted notes.

Conclusion and Future Work
This research was conducted to determine the efficiency of selected Modified HOG/LVHCF descriptor for automatic recognition of paper money. The proposed system was implemented using the Nigerian notes, and for experimental evaluation of the system, a further test was conducted on USD, CAD, and Euro notes. The proposed method extracted the HOG descriptor from image patches created from the vertices of SURF interest points feature of the banknotes, Low Variance, and High Correlation feature reduction filter was employed to reduce the feature set and classified using OVO-SVM in identifying and authenticating fake and genuine bills. Lastly, point matching technique was applied to detect feature vectors to match images containing multiple banknotes to determine the value of the currencies. Future research would include incorporating new feature descriptors and reduction approach to reduce processing time in addition to scaling down the application to work on mobile devices and reducing the computational complexity to enable embedding in micro-controllers and microchips.