Volume 21, Issue 4 (December 2023)                   Iranian Rehabilitation Journal 2023, 21(4): 639-654 | Back to browse issues page

Ethics code: IR.USWR.REC.1398.114


XML Print


Download citation:
BibTeX | RIS | EndNote | Medlars | ProCite | Reference Manager | RefWorks
Send citation to:

Zarifian T, Ashtari A, Nilipour R, Nematzadeh S, Bayat N. Developing the KHANA Test to Evaluate Reading Skills in Persian-speaking Students: A Preliminary Study. Iranian Rehabilitation Journal 2023; 21 (4) :639-654
URL: http://irj.uswr.ac.ir/article-1-1787-en.html
1- Department of Speech Therapy, School of Rehabilitation Sciences, University of Social Welfare and Rehabilitation Sciences, Tehran, Iran.
2- Department of Linguistics, Faculty of Literature, Al Zahra University, Tehran, Iran.
Full-Text [PDF 562 kb]   (212 Downloads)     |   Abstract (HTML)  (1009 Views)
Full-Text:   (111 Views)
Introduction
Being literate improves individuals’ quality of life since it allows them to pursue their studies and careers [1]. These days, the mainstream of human communication is via reading and writing. The technological progress also convinced individuals to use electronic devices more, which has become more crucial [2]. 
Reading is a complex cognitive ability and is necessary for language learning and communication [3]. This ability includes three skills, accuracy, rate, and reading comprehension. Accuracy of reading refers to the number of words that are correctly produced while reading aloud. The reading rate is related to the reading time, and reading comprehension is related to the number of questions that the readers answer correctly [4]. Any problems associated with reading will inevitably lead to difficult written communication and knowledge acquisition [5].
Reading disorder is the most prevalent type of learning disorder and is estimated to occur in approximately 5% to 12% of school-age children [6]. Students with reading impairments (SRI) have problems with accurate or fluent word recognition and spelling, as well as decoding, despite enough instruction, normal intelligence, and intact sensory abilities [7, 8]. Difficulties in educational skills, such as listening, reading, writing, mathematics, and problem-solving can affect children’s rate, accuracy, and comprehension skills, as well as academic performance and social communication [9]. SRIs face negative consequences, such as challenges in continuing their education and choosing a suitable job [1]. Therefore, it affects their social participation and quality of life [10]. 
Over the past years, a considerable number of studies have assessed children’s ability to learn reading and the challenges that they may face in learning this skill. These researches shed light on the nature of reading impairments as well as the assessment methods [3]. 
Tests are available to assess reading impairments, such as Woodcock reading mastery tests-revised (WRMTR) [11], the Gray oral reading test (GORT) [12], the qualitative reading inventory-5 (QRI-5) [13], the reading and dyslexia test (NEMA) [14], the diagnostic reading test [15]. These tests represent different aspects of the problem as well as the strengths and weaknesses that can be used to design and implement effective treatment plans [3]. WRMTR is one such test that evaluates reading ability in individuals from kindergarten to adulthood. It contains various subtests, such as letter and word identification, and word and passage comprehension [11]. GORT is another reading test that evaluates reading abilities (i.e. accuracy, rate, and reading comprehension) using silent and oral reading. GORT-5 is the fifth edition and a norm-referenced of this test. It has norms for people aged between 6 to 23 years. It is a valid and reliable test that contains 16 story passages followed by 5 comprehension questions [12]. Another test to assess reading skills is QRI-5, which provides information about conditions under which students can detect words and understand text successfully or unsuccessfully. This inventory contains graded word lists and numerous passages designed to assess the oral and silent reading and listening ability of students from the preprimary school through the high school levels [13].
Although several tests exist to assess reading skills in English, they cannot be used in Persian due to different characteristics, such as orthographic and script styles. Most Persian reading tests do not evaluate all aspects of reading ability and are not up-to-date because they are designed based on the previous version of the course books taught in schools. Also, a great number of Persian reading tests are left unpublished and are not available to use in clinics and research projects [4, 1617]. 
NEMA is the only Persian standardized test to diagnose dyslexia in students from first to fifth grades [14]. This test does not evaluate the rate and accuracy of reading in the passage and the guidelines for performing the tasks are not clear. The diagnostic reading test is another Persian tool to screen and diagnose reading impairments as well as assess spelling skills in second-grade students. No cut-off scores were calculated for this test and the passages designed for evaluation are very easy for students. Moreover, the test was designed based on the old version of the course books (significant changes were observed in Persian course books in recent years) [15].
As mentioned above, reading impairments hurt children’s lives because they deter students from succeeding at school. Early diagnosis of this impairment results in early treatment, which is more effective than treatment in later years. 
While various tests exist to assess reading in the world [12, 1819], Persian tests are limited and out of date. However, among these few tests, only one of them has been published (NEMA) [14], and the rest are merely available as final reports in the research and educational departments. None of these Persian reading tests are designed based on the current course books; therefore, this study aims to develop a valid and reliable KHANA test to evaluate reading skills in Persian-speaking students and study accuracy, rate, and reading comprehension in different grades and genders. This study reports the preliminary stage of developing a comprehensive test to evaluate reading ability in Persian students. 

Materials and Methods 
This study was a methodological study that had two phases. We developed the test and measured its psychometric properties in the first phase. In the second phase, 87 students were included and assessed using the designed test. 
First, we prepared two preliminary parallel forms of the student books (A and B), each book includes 12 reading passages that were organized based on increasing difficulty as the student moves from one passage to the next. Any passage contains five comprehension questions. These two parallel forms provide the opportunity to determine the reliability of reading skills assessment as well as to evaluate the student’s performance before and after treatment. This draft was based on course books from the second to seventh grades to assess reading skills (i.e. accuracy, rate, and reading comprehension). We were inspired by reading tests, such as WRMTR [11], GORT-5 [12], QRI-5 [13], NEMA [14], the diagnostic reading test [15], and Shafiei et al.’s test [4] in designing KHANA.
Then, we measured the content and face validity. Also, we determined the descriptive statistics, construct validity, and reliability, including test re-test, inter-rater, and the correlation between passages A and B.
In the current study, 75 typically developing students (TDS) (46 girls and 29 boys) who studied from the second to seventh grade were selected using a convenience sampling method. Twelve SRI (6 girls and 6 boys) also participated. The inclusion criteria for TDS included the second to seventh-grade student in Tehran City or Karaj City, Iran speaking Persian as a dominant language at home, having the normal hearing ability and a non-verbal intelligence quotient (IQ) of more than 85 based on the student health report available at school, having a good reading skill, no language deficit or speech sound disorders based on the examiner’s informal evaluation, and no reading or writing problems based on the teachers’ report. The exclusion criteria included having symptoms of speech sound disorders, visual problems, symptoms of sensory deficits (even wearing glasses or hearing aids), and symptoms of psychiatric disorders, such as autism spectrum disorders, psycho-physical delays, evident oro-motor deficits, and a history of recurrent middle ear infections, epilepsy, convulsion, syncope, and brain damages based on their medical health report, asking their parents and teachers. All the criteria were true for the SRI, except for having good reading skills. They had problems in reading based on their teachers’ reports and or a speech therapist’s diagnosis. Table 1 presents the demographic information of the participants. 



Procedure
In phase one, an expert panel, including 13 experts (8 speech therapists with experience in reading impairments, 4 linguistics with experience in psycholinguistics and clinical linguistics, and 1 methodologist) received the first draft of the 24 passages and evaluated the content validity based on being well-formed, coherent, having interesting text, having no bias in the text, being suitable for the students, as well as having relevant questions to the passage. To measure the face validity, 10 teachers with at least five years of teaching experience at primary schools received the passages and questions and were asked to classify passages for each grade, determine the difficulty of the sentences, and comment on whether the reading comprehension questions were appropriate. After reviewing the results, the final version of the test was prepared for the pilot test, in which 10 TDS (4 girls and 6 boys) participated.
A protocol was designed for the examiners that contained information about the test. This information was about how to record the participants’ voices while reading the passages, how to give feedback, or when to stop the test (if the participants had more than 10 mistakes in two consecutive passages, the test was stopped. This protocol was taught to the examiners who were final-year undergraduate students and speech therapist graduates. They were completely aware of the test and its aims, as well as its guidelines. The research team (the first and second authors) approved the examiners’ eligibility after piloting a group of students. In the second phase, the trained and eligible examiners took the test from the students living in Tehran and Karaj cities for seven months. 
The gathered data in the second phase were analyzed using SPSS software, version 24. Since the data did not follow a normal distribution (based on Kolmogorov-Smirnov and Shapiro-Wilk tests), non-parametrical tests were used. Descriptive statistics were used to determine the Mean±SD, minimum and maximum accuracy, rate, and reading comprehension in different grades. In addition to the content validity and face validity, construct validity to determine whether the test can differentiate rate, accuracy, and comprehension skills between the different grades, the TDS group and the SRI group as well as the two genders was measured. To evaluate the reliability, we measured the test re-test reliability (using Spearman’s correlation coefficient), the inter-rater reliability (using the intraclass correlation coefficient [ICC]), and the correlation between passages A and B. To evaluate the performance of the two genders and children with and without reading impairments, the Mann-Whitney U test was used. Also, the Kruskal-Wallis’s test was used to determine accuracy, rate, and reading comprehension between the different grades. 

Results
In the current study, 87 students participated, including 75 TDS and 12 SRI. The results of the statistical analysis are described in the following. 
The content validity of the A and B student books, each including 12 passages and five questions dedicated to them, was measured qualitatively based on the experts’ opinions. Most passages were scored 90 and more and a few of them were scored 70. Those passages which scored 70 were reviewed by the research team. Then, they were sent back to the experts, and they approved the passages. 
The face validity of the test was measured by asking 10 teachers to classify the passages and questions suitable for different grades and to determine the sentence difficulty percentage on a visual scale analysis. Based on the teachers’ opinions, the stories and questions were sorted according to their difficulty level. Finally, 24 passages (12 for each A and B) were sorted in terms of difficulty. 
Discriminant validity of the rate, accuracy, and comprehension in both passages were applied to determine the construct validity of the test. Table 2 and Table 3 present these results.




Based on the results presented in Table 2, for passage A, a significant difference was observed between all the grades in the rate of reading. For accuracy, only passages 3 to 8 and for reading comprehension, only passages 2 and 8 demonstrated significant differences. 
According to the validity results of discriminating the passages of group B, passages 1 to 9 were significantly different for the rate of reading, passages 4 to 7 and 10 for accuracy, and passages 3 and 11 for reading comprehension (Table 3). 
The test’s ability was measured to discriminate reading skills between TDS and SRI. According to the results, rate, accuracy, and comprehension skills were significantly better in TDS. Appendix 1 presents the results. Moreover, in detecting differences between boys and girls, no gender differences were found in the mean rate scores.




In accuracy, girls had significantly higher scores than boys in passages A1, A3, A4, A7, and B3. Girls also showed better performance in comprehension of passages A3 and B5. Appendix 2 presents the results. 






According to Table 4, the test re-test analysis indicated a significant correlation between the reading skills in most passages A and B.


The correlation coefficient results in Table 5 showed a significant correlation between the independent measurement of the accuracy of passages A and B by two evaluators.


Table 6 also indicates that all passages A and B are significantly correlated in terms of rate, accuracy, and comprehension, except for passage 8 (comprehension), passage 9 (comprehension), passage 11 (comprehension), and passage 12 (accuracy). 



Discussion
As a basic element for learning, reading is a crucial skill that opens doors to education, employment, and well-being [1]. Reading impairments hurt children; therefore, it is essential to identify SRI [20]. Since a limited number of published Persian tests exist (only one is available, NEMA), most of them are outdated and unavailable [1415], this study was conducted to develop a valid and reliable test for assessing reading skills in Persian-speaking students. KHANA contains two preliminary parallel forms of student books (A and B) that can be used to determine the reliability of reading skills assessment and evaluate the performance of the SRI before and after the treatment (however, the latter in the preliminary study was not performed). Also, accuracy, rate, and reading comprehension were measured in different grades and genders. 
Based on the content and face validity results, most passages and questions were valid. Those with the lower scores at this stage were reviewed and edited based on the experts’ and teachers’ opinions. 
In construct validity, reading skills were assessed between the students from the second to seventh grade, between TDS and SRI, as well as between the two genders using KHANA. According to the results, the rate of reading was significantly different between all the grades. Based on Ehri, students have slow reading speed in the first three years of school, and this rate increases in higher grades as their reading skills improve [21]. The results are also consistent with some Persian studies indicating that the rate of reading improves from the first to seventh grades [4, 1617]. 
Based on the accuracy results, this reading skill improved from the second to seventh grades in both passages A and B, which is consistent with Ehri’s results on reading development from alphabetical to orthographic reading [21]. It seems the student becomes more capable of using reading rules in higher grades. Therefore, we can expect to have higher accuracy scores in older students. It is also consistent with studies conducted by Shafiei et al., Aziziyan et al. and Jabbari et al. [4, 1617]. In the present study, some passages, such as passages A1 and A2, could not make a significant difference since they both evaluate the basic level of reading. These results confirm the construct validity of the test. 
Only a limited number of passages can make significant differences between the reading comprehension skills at different grades. Assessing reading comprehension is a challenge in most reading tests in foreign and Persian tests. This problem is evident in NEMA [14] and the diagnostic reading test [15]. Also, a great number of changes were made in the comprehension section of different versions of Gort [12]. It seems that the reason for this challenge is that the questions evaluate objective issues and the students did not have to deduct the whole passage for a correct answer. 
In comparing the female and male students’ reading rates, no significant differences were observed between the two genders; however, based on Shafiei et al., Aziziyan et al. and Shirazi, girls demonstrated better performance than boys [4, 16, 22]. Reading accuracy in a few numbers of the passages (such as A1, A3, and A7) was significantly better in girls. According to Shafiei et al. and Aziziyan et al., girls are better at accuracy than boys [4, 16]. Girls also had significantly higher scores in reading comprehension only in passages A3 and B3. The difference between the results may be due to the greater number of female students who participated in the current study.
We also compared the SRI and TDS reading skills. According to the results, TDS performed significantly better compared to SRI; therefore, KHANA can discriminate between these two groups. WRMTR [11] and GORT-5 [12] are other tests that have diagnostic capability. 
According to evaluation of the passages A and B, the results showed a highly significant correlation between the reading skills in most passages A and B. However, it is preferred to have parallel passages that are correlated. Four passages A and B were not highly correlated; therefore, it is necessary to revise these texts (passages 8, 9, 11, 12).
Moreover, according to the test re-test reliability results, an acceptable correlation was observed between the two evaluators in most administrations; therefore, KHANA is a reliable tool.

Conclusion
In conclusion, the development of the KHANA test facilitates assessing reading skills in Persian-speaking students. The study successfully created two preliminary parallel forms of student books (A and B) to evaluate the accuracy, rate, and comprehension in different grades and genders. The validity assessment, including content, face, and construct validity, demonstrates that the passages and questions are well-constructed and measure the intended skills effectively. Based on the test re-test analysis, the inter-rater reliability, and the correlation between the two passages, KHANA is a reliable tool and is capable of discriminating between SRI and TDS. In summary, the KHANA test emerges as a valuable and reliable instrument for evaluating reading skills in Persian-speaking students, with the potential to contribute significantly to educational and clinical settings.

Limitations and recommendations
This study had limitations. The COVID-19 pandemic affects the sample size, especially SRI. Also, the number of girls participating in this study was more than boys. The other limitation was about matching the TDS and SRI groups. They merely matched their grade levels. The preferred way is to have two groups matched in terms of the language testing results or literacy achievements; however, no standard Persian test is found to evaluate language or literacy in students. Therefore, the examiners used informal tests. It is suggested to use a random sampling method with a larger sample size and calculate the specificity and sensitivity, as well as standard scores for the reading skills in each grade in future studies. Furthermore, according to the low correlation between passages A and B and comprehension questions, it is recommended to revise them. 

Ethical Considerations
Compliance with ethical guidelines

The study was approved by the Ethics Committee of the University of Social Welfare and Rehabilitation Sciences (Code: IR.USWR.REC.1398.114). All participants and their parents were informed about the test and the procedure and signed the informed consent form. 

Funding
This work was financially supported by the Deputy of Research and Technology, University of Social Welfare and Rehabilitation Sciences (Grant No.: 2321).

Authors' contributions
Conceptualization, methodology and supervision: Talieh Zarifian, Atieh Ashtari, Reza Nilipour, Shahin Nematzadeh; Data collection: Talieh Zarifian, Atieh Ashtari, Narges Bayat; Data analysis: Talieh Zarifian and Atieh Ashtari; Investigation and writing: All authors.

Conflict of interest
The authors declared no conflict of interest. 

Acknowledgments
The authors appreciate the students, families, and teachers for participating in this study. 

References
  1. Snowling MJ, Hulme C. Annual research review: The nature and classification of reading disorders--a commentary on proposals for DSM-5. Journal of Child Psychology and Psychiatry, and Allied Disciplines. 2012; 53(5):593-607. [DOI:10.1111/j.1469-7610.2011.02495.x] [PMID]
  2. Poe MT. A History of Communications: Media and society from the evolution of speech to the internet. New York: Cambridge University Press; 2010. [Link]
  3. Fletcher JM, Foorman BR, Boudousquie A, Barnes MA, Schatschneider C, Francis DJ. Assessment of reading and learning disabilities a research-based intervention-oriented approach. Journal of School Psychology. 2002; 40(1):27-63. [DOI:10.1016/S0022-4405(01)00093-0]
  4. Shafiei B, Tavakol S, Alinia L, Maracy MR, Sedaghati L, Foroughi R. [Developing a screening inventory reading test (IRT) for the Isfahanian students of the first to fifth grade (Persian)]. Bimonthly Audiology. 2009; 17(2):53-60. [Link]
  5. MacCullagh L, Bosanquet A, Badcock NA. University students with dyslexia: A qualitative exploratory study of learning practices, challenges and strategies. Dyslexia. 2017; 23(1):3-23. [DOI:10.1002/dys.1544] [PMID]
  6. Sedaghati L, Foroughi R, Shafiei B, Maracy MR. [Prevalence of dyslexia in first to fifth grade elementary students Isfahan, Iran (Persian)]. Bimonthly Audiology. 2010; 19(1):94-101. [Link]
  7. Peterson RL, Pennington BF. Developmental dyslexia. Annual Review of Clinical Psychology. 2015; 11:283-307. [DOI:10.1146/annurev-clinpsy-032814-112842] [PMID]
  8. Colenbrander D, Ricketts J, Breadmore HL. Early Identification of dyslexia: Understanding the issues. Language, Speech, and Hearing Services in Schools. 2018; 49(4):817-28. [DOI:10.1044/2018_LSHSS-DYSLC-18-0007] [PMID]
  9. Alexander PA, Judy JE. The interaction of domain-specific and strategic knowledge in academic performance. Review of Educational Research. 1988; 58(4):375-404. [DOI:10.3102/00346543058004375]
  10. Riva S, Antonietti A. The application of the ICF CY model in specific learning difficulties: A case study. Psychology of Language and Communication. 2010; 14(2):37-58. [DOI:10.2478/v10057-010-0009-2]
  11. Eaves RC. Woodcock reading mastery tests-revised (WRMTR). Diagnostique. 1990; 15(1-4):277-97. [DOI:10.1177/15345084890151-425]
  12. Hall AH, Tannebaum RP. Test review: J. L. Wiederholt & B. R. Bryant. (2012). Gray oral reading tests-fifth edition (GORT-5). Austin, TX: Pro-Ed. Journal of Psychoeducational Assessment. 2013; 31(5):516-20. [DOI:10.1177/0734282912468578]
  13. Leslie L, Caldwell JS. Qualitative reading inventory. Indianapolis: Pearson Education; 2016. [Link]
  14. Moradi A, Hosaini M, Kormi Nouri R, Hassani J, Parhoon H. [Reliability and validity of reading and dyslexia test (NEMA) (Persian)]. Advances in Cognitive Sciences. 2016; 18(1):22-34. [Link]
  15. Sima-Shirazi T, Nili-pour R. [Development and standardization of reading diagnostic test (Persian)]. Archive of Rehabilitation. 1383; 5(1 and 2) :7-11. [Link]
  16. Aziziyan M, Abedi M. [Construction and standardization of the reading level diagnostic test for third grade primary school students (Persian)]. Iranian Journal of Psychiatry and Clinical Psychology. 2005; 11(4):379-87. [Link]
  17. Jabbari S, Khademi M. [Creating a reading and comprehension diagnostic test for elementary school students (Persian)]. Curriculum Studies Journal. 2014; 3(2):33-51. [DOI:10.22099/jcr.2014.2970]
  18. Neale MD. Neale analysis of reading ability. Adelaide: Australian Council for Educational Research; 1999. [Link]
  19. Cox MM. The relationship of speech and reading in an elementary school program. Communication Education. 1959; 8(3):211-8. [DOI:10.1080/03634525909377021]
  20. Nergård-Nilssen T, Eklund K. Evaluation of the psychometric properties of "the Norwegian screening test for dyslexia". Dyslexia. 2018; 24(3):250-62. [DOI:10.1002/dys.1577] [PMID]
  21. Ehri LC. Learning to read words: Theory, findings, and issues. Scientific Studies of reading. 2005; 9(2):167-88. [DOI:10.1207/s1532799xssr0902_4]
  22. Shirazi S. Phonological Awareness and its implications for Reading Acquisition. Iranian Rehabilitation Journal. 2006; 4(1):40-4. [Link]
Article type: Original Research Articles | Subject: Speech therapy
Received: 2022/10/12 | Accepted: 2023/03/14 | Published: 2023/12/1

Send email to the article author


Rights and permissions
Creative Commons License This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.

Designed & Developed by : Yektaweb