Handwritten Font Classification Method Based on Ghost Imaging

Publication Name

Guangzi Xuebao/Acta Photonica Sinica


With the rapid development of the economy, financial services such as bills are also increasing day by day. Among them, the important information in the bill business, such as personal vouchers, checks, and other bills, requires manual reading and input of a large amount of digital information. In order to avoid the waste of human and financial resources, related researchers classify and recognize handwritten fonts based on the neural network classification idea of ​​deep learning. When the method based on deep learning is used for feature extraction of handwritten fonts, there is often a lack of detailed information such as edges and textures, which leads to the problem of low recognition accuracy. Aiming at the problem that the features of handwritten digits or letters are difficult to effectively extract, the recognition efficiency was not high and even caused recognition errors, a new automatic recognition method of handwritten digits or letters was proposed by combining the principle of ghost imaging and the classification network based on deep learning. This method utilizes the principle of ghost imaging. It can save the imaging process in the traditional image recognition method, and jumping out of the inherent thinking that identifying objects is identifying images, and can quickly classify the image of handwritten digits and letters only by the total light intensity value transmitted by handwritten digits or letters without extracting and identifying features of handwritten digits or letters. The automatic recognition of handwritten digits or letters based on ghost imaging solves the critical problem of needing to extract digits or letter images features in traditional handwritten font recognition methods, and can greatly improve the recognition efficiency of handwritten digits or letters. Firstly, a ghost imaging detection system is built using commonly used optical instruments such as lasers, digital micromirror arrays, and single-pixel detectors. The laser in the built detection system is used to generate a pseudothermal light source, and the digital micromirror array is used to obtain the Hadamard speckle sequence with a resolution of 32×32 irradiating the target object at different times. And realizing the irradiation of 17 239 handwritten images of handwritten digits and letters. Secondly, the single-pixel detector is used to collect data on the total light intensity value transmitted by the handwritten digits and letters. The data collection process is very fast and does not cause huge time costs. The value of the bucket detector after the collection is converted into a one-dimensional vector, and use the one-dimensional vector corresponds to the handwritten font as the input of network training. Finally, the network framework is built based on the advantages of the convolutional neural network in image classification and is used to solve the problems in the training process. The network degradation problem is added to the residual block structure, which can directly pass shallow information to deeper layers by skipping one or several layers through skip connections. In order to solve the problem of overfitting, the Dropout layer is added to it, and the robustness of the network to the loss of specific neuron connections is improved by reducing the weight. The experimental results show that: for handwritten digits, compared with the fully connected network, the precision, recall rate and F1 value of the convolutional neural network model are increased by 86.50%/97.25%, 86.40%/98.03%, 86.31%/97.60%; for handwritten letters, the precision, recall, and F1 value of the convolutional neural network under full sampling are 91.87%, 90%, and 90.23%, respectively. At the same time, in the case of undersampling and non-undersampling, the ten types of digits from 0 to 9 under the two models of convolutional neural network and fully connected neural network and randomly selected l, v, y, z, m, n, o, r, s, and h ten types of letters are compared and analyzed. The experimental results show that the accuracy rate of each type of digit and letter of the convolutional neural network is higher than that of the fully connected network under the same conditions. The accuracy of each type of digit and letter under the two models further verifies that as the sampling rate increases, the recognition accuracy also increases. By comparing the evaluation indicators of the convolutional neural network and the fully connected network architecture, the effectiveness and rationality of the proposed method are further illustrated. The classification and recognition results of handwritten letters verified by experiments further illustrate the versatility of the constructed convolutional neural network. It provides the possibility for the wide application of handwritten fonts in real life. The research on the classification and the recognition of handwritten fonts based on ghost imaging can effectively solve the bottleneck problem of low recognition efficiency of existing font recognition methods.

Open Access Status

This publication is not available as open access





Article Number


Funding Number


Funding Sponsor

National Natural Science Foundation of China



Link to publisher version (DOI)