Fine-Tuning is a Surprisingly Effective Domain Adaptation Baseline in Handwriting Recognition.

arxiv(2023)

引用 1|浏览15
暂无评分
摘要
In many machine learning tasks, a large general dataset and a small specialized dataset are available. In such situations, various domain adaptation methods can be used to adapt a general model to the target dataset. We show that in the case of neural networks trained for handwriting recognition using CTC, simple fine-tuning with data augmentation works surprisingly well in such scenarios and that it is resistant to overfitting even for very small target domain datasets. We evaluated the behavior of fine-tuning with respect to augmentation, training data size, and quality of the pre-trained network, both in writer-dependent and writer-independent settings. On a large real-world dataset, fine-tuning on new writers provided an average relative CER improvement of 25% for 16 text lines and 50% for 256 text lines.
更多
查看译文
关键词
effective domain adaptation baseline,domain adaptation,recognition,fine-tuning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要