
		<paper>
			<loc>https://jjcit.org/paper/272</loc>
			<title>AN ENHANCED WORD LEVEL ARABIC OCR BASED ON DUAL ENCODER TRANSFORMER ARCHITECTURE</title>
			<doi>10.5455/jjcit.71-1746709575</doi>
			<authors>Khulood Gaashan,Maram Bani Younes</authors>
			<keywords>Arabic OCR,Multi-batch size,Transformer,Dual encoder transformer,Decoder,Feature extraction,Self-attention mechanism</keywords>
			<citation>2</citation>
			<views>2327</views>
			<downloads>918</downloads>
			<received_date>16-Jun.-2025</received_date>
			<revised_date>  12-Jul.-2025 and 12-Aug.-2025</revised_date>
			<accepted_date>  13-Sep.-2025</accepted_date>
			<abstract>Arabic script is one of the most sophisticated and difficult scripts. It uses different shapes of characters with 
complex diacritical marks that are difficult to distinguish from the dots of characters. This script’s distinctive 
features make the Optical Character Recognition (OCR) procedure more challenging and result in low-accuracy 
recognition. Different studies have aimed to introduce high-accuracy Arabic OCR in the literature.  However, 
enhancing the accuracy of reading the words has been an open issue that depends on the used dataset and the 
developed recognition model. Besides, considering diacritics has been limited and not sufficiently addressed. 
Experimental tests on words with diacritics in prior models have shown bad accuracy that does not exceed 60%. 
Consequently, this work aims to introduce a new, accurate deep-learning model for Arabic OCR that considers 
words with and without diacritical marks. It utilizes a dual encoder transformer (DTrOCR), a deep-learning 
architecture that has demonstrated superior performance in identification and classification tasks. The proposed 
DTrOCR creates multi-batch sizes. It has been trained using a comprehensive, generated Arabic word-based 
dataset named MFSRHRD and tested on unseen datasets. The accuracy of configuring Arabic words without 
diacritics reaches 98.5%. However, for words with diacritics, it achieved an accuracy of 89.9%.</abstract>
		</paper>


