新疆大学学报(自然科学版)
新疆大學學報(自然科學版)
신강대학학보(자연과학판)
XINJIANG UNIVERSITY JOURNAL(NATURAL SCIENCE EDITION)
2015年
2期
215-220
,共6页
王剑%哈力木拉提·买买提%艾尔肯·赛甫丁%程园
王劍%哈力木拉提·買買提%艾爾肯·賽甫丁%程園
왕검%합력목랍제·매매제%애이긍·새보정%정완
维吾尔文倾斜校正基线拟合文本行
維吾爾文傾斜校正基線擬閤文本行
유오이문경사교정기선의합문본행
Uighur tilt correction baseline fitting a line of text
在维吾尔文扫描过程中,输入的文档不可避免地会发生倾斜现象,而现有的方法只进行了初步的倾斜校正。针对上述情况,同时为了方便维文图像的切分和识别工作,提出将基于凸多边形的最小面积外接矩形法和基线拟合法相结合的检测和校正图像方法。首先利用基于凸多边形的最小面积外接矩形法实现初步的倾斜校正,然后提取一行文本后采取基线拟合的方法实现文本行的单独校正,最后把校正过的所有文本行整合成一个文档。实验结果表明,该方法是行之有效的采用,该方法比现有方法在字母切分的准确率上平均提高约5%,最高提高约7%。
在維吾爾文掃描過程中,輸入的文檔不可避免地會髮生傾斜現象,而現有的方法隻進行瞭初步的傾斜校正。針對上述情況,同時為瞭方便維文圖像的切分和識彆工作,提齣將基于凸多邊形的最小麵積外接矩形法和基線擬閤法相結閤的檢測和校正圖像方法。首先利用基于凸多邊形的最小麵積外接矩形法實現初步的傾斜校正,然後提取一行文本後採取基線擬閤的方法實現文本行的單獨校正,最後把校正過的所有文本行整閤成一箇文檔。實驗結果錶明,該方法是行之有效的採用,該方法比現有方法在字母切分的準確率上平均提高約5%,最高提高約7%。
재유오이문소묘과정중,수입적문당불가피면지회발생경사현상,이현유적방법지진행료초보적경사교정。침대상술정황,동시위료방편유문도상적절분화식별공작,제출장기우철다변형적최소면적외접구형법화기선의합법상결합적검측화교정도상방법。수선이용기우철다변형적최소면적외접구형법실현초보적경사교정,연후제취일행문본후채취기선의합적방법실현문본행적단독교정,최후파교정과적소유문본행정합성일개문당。실험결과표명,해방법시행지유효적채용,해방법비현유방법재자모절분적준학솔상평균제고약5%,최고제고약7%。
In Uighur text scanning process, the input document inevitably occur tilt function, however, we have only carried on the preliminary tilt correction using the existing method. In view of the above situation, at the same time in order to facilitate the segmentation and recognition of Uighur character scanning this paper proposed one approach that combining minimum-area bounding rectangle method based on the convex polygon with baseline fitting method to detect and correct text. Firstly used the minimum-area bounding rectangle method based on the convex polygon to realize the initial correction. Then extracted a line of the text and used baseline fitting method to correct the tilted text individually, finally, integrated the corrected text lines into a document. The existing methods only carried out preliminary tilt correction. The experimental results show that this method is accurate and effective, using our scheme can obtain a general increase in the accuracy of character segmentation by about 5%compared to the existing methods, and get a highest accuracy up to 7%.