Abstract
Deep learning techniques have been successfully applied in handwriting recognition. Oracle bone inscriptions (OBI) are the earliest hieroglyphs in China and valuable resources for studying the etymology of Chinese characters. OBI are of important historical and cultural value in China; thus, textual research surrounding the characters of OBI is a huge challenge for archaeologists. In this work, we built a dataset named OBI-100, which contains 100 classes of oracle bone inscriptions collected from two OBI dictionaries. The dataset includes more than 128,000 character samples related to the natural environment, humans, animals, plants, etc. In addition, we propose improved models based on three typical deep convolutional network structures to recognize the OBI-100 dataset. By modifying the parameters, adjusting the network structures, and adopting optimization strategies, we demonstrate experimentally that these models perform fairly well in OBI recognition. For the 100-category OBI classification task, the optimal model achieves an accuracy of 99.5%, which shows competitive performance compared with other state-of-the-art approaches. We hope that this work can provide a valuable tool for character recognition of OBI.
Highlights
Oracle bone inscriptions (OBI) were recorded from as early as the Shang Dynasty inChina [1]
We visualized the changes in the training loss value, the accuracy rates on the training set and the validation set of different models during the training process as the number of training epochs increases
OBI-100 can fill the gap of publicly available datasets in the applications of deep learning in OBI research
Summary
Oracle bone inscriptions (OBI) were recorded from as early as the Shang Dynasty inChina [1]. Convolutional neural network (CNN) [21] is a multi-layer feed-forward neural network that can extract features and properties from the input data. As shown, the basic CNN structure consists of an input layer, several convolutional layers, and pooling layers, as well as several fully connected layers and an output layer. The convolutional layer is designed to extract features from the input data, which contain many convolutional kernels. The parameters of the convolutional layer include the size of the kernel, the step size, and the padding method [23] These three factors jointly determine the size of the output feature map of the convolutional layer [24]. After feature extraction [26] by the convolutional layer, the output feature map is transferred into the pooling layer for feature selection and information filtering. The output layer uses a logistic function or a normalized exponential function to output the classification label or probability
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.