An Investigation into Online Content’s Readability in Light of a Brain-Based Model: A Case from the Sultanate of Oman

Document Type : Research Article

Authors

1 Professor, Gulf College, Muscat, Oman

2 PhD, Department of English, Faculty of Letters and Humanities, Ferdowsi University of Mashhad, Iran

3 Assistant Professor, Department of English, Faculty of Letters and Humanities, Ferdowsi University of Mashhad, Iran

Abstract

In the context of the rapid dissemination of information, scaffolding the readability of online content is critical. This study introduces the brainling model as a novel approach to measuring readability, which heeds readers’ cognition, senses, emotions, and cultural background in conjunction with language. To achieve the aims, two texts covering a general topic were selected, both possessing the same readability level according to the Flesch Reading Ease scale. However, one of the texts was modified in accordance with the brainling components (i.e., cogling, emoling, sensoling, and cultuling). Following each reading text, five multiple-choice comprehension questions, a 10-‎item Likert scale for ‎readability, and a scale for the text’s difficulty level scale were designed. The Google Form was used to collect responses from 209 individuals with intermediate language ‎proficiency at an educational institution in Oman. ‎ After verifying ‎the reliability and validity of the questions and scales, significantly higher mean scores ‎were ‎observed for the readability scale components of clarity and engagement in the brainling-based ‎‎modified text, compared to the unmodified text. Moreover, ‎the results obtained from the text difficulty scale, readers rated the brainling-based text as easier to comprehend. Furthermore, based on the reading comprehension test, participants achieved higher scores when reading the brainling-based text. These findings demonstrate that modifying a text based on the brainling model, which integrates both brain and linguistic structures, significantly enhances clarity, engagement, text difficulty, and reading comprehension scores. Adopting the brainling approach can be a solution for fostering mutual ‎understanding and collaboration ‎among culturally diverse members and alleviating challenges such as misunderstanding, that may ‎hinder ‎goal achievement.
 

Keywords

Main Subjects


Introduction

The Internet opens up infinite possibilities to everyone and allows people to access massive online information resources. Online content expedites the dissemination of pieces of information, leading to the availability of data anytime and anywhere. But how easy is it for people to understand them? To this end, the readability concept has emerged to determine how easily content can be read. Readability is an attempt to predict the difficulty level of texts or make them easy and comprehensible. Text comprehension (Kate et al., 2010) and text processing (Crossley et al., 2023) are two notions of readability. Text comprehension involves word and syntactic complexities, and discourse structures (Mesmer et al., 2012); however, text processing is generally associated with the speed of reading (Crossley et al., 2023). A reader-text mismatch in difficulty level may disrupt comprehension. Indeed, the success of readers is determined by the extent to which they can understand texts (Grabowski & Mathiebe, 2024). To check the readability level of texts with regard to the intended readers, several qualitative and quantitative studies have been conducted (e.g., Ahmad & Hasan, 2024; Crossley et al., 2023; Grabowski et al., 2010). As progress in the studies is made, the variety of factors becomes greater.

Historically, emphases on linguistic and syntactic features have made readability attractive in classic studies. Some researchers (e.g., Collins‐Thompson & Callan, 2005; Kincaid et al., 1975; Pitler & Nenkova, 2008) presuppose that the employment of formulae (e.g., the Flesch formula, the Fry graph, the Lexile, and Gunning Fog formula) would ensure objective and accurate results. Formulae-based investigations (e.g., Dale & Chall, 1948; Smith et al., 1989; Spache, 1953) measure the difficulty level by word frequency, length, and syllables. Word- and sentence-oriented formulae have hampered studies to encompass important textual factors (e.g., narrative aspect, semanticity, and text cohesion) (Crossley
et al., 2023
) and extra-textual factors (Schriver, 2000).

Thereafter, the classic readability models have been changed by the idea that the numerical scores may be misleading since the developers considered readability as a monolithic rather than a multifaceted phenomenon (Bailin & Grafstein, 2001). The presupposition underlying the modern formulae is that readability is not reducible to a single and simple measure and should not be limited to the surface properties of texts (Bailin & Grafstein, 2001). Accordingly, the presence of readers’ characteristics in reading comprehension is undeniable. Hence, readability may be the outcome of readers’ characteristics in association with text properties (Bailin & Grafstein, 2001; McNamara & Kintsch, 1996).

In reviewing multiple factors that affect readability, we intended to investigate that readability is not a matter of a factor or factors. If we consider the extent to which the interaction among the structures of the brain can create unique characteristics, then we can suggest that the demands that unreadable contents place on readers in terms of uniting the words and sentences may be in virtue of the brain-based features. Hence, this study was significant because it suggested a brain-based pattern (i.e., brainling model) that could be befitting to predict the readability of online texts. Introduced by Pishghadam and Ebrahimi (2020), the brainling pattern has four components: Cogling, emoling, sensoling, and cultuling. Cogling reflects the writer’s ability to use language in a way that matches readers’ background knowledge, thoughts, and intentions, as well as their expectations and needs. Based on the sensoling component, what is readable for a reader who has optimally experienced multiple senses (i.e., seen, heard, tasted, smelled, and touched) is perhaps different from what is easy to read for a person who is unfamiliar with the topic. According to the emoling, the more arousal a text is, the easier it is to be processed. However, this does not mean that positive texts are directly correlated with readability score, but the readers’ experiences and their aroused emotions make the text readable or unreadable to them. Finally, the general assumption in cultuling is that a text that may be easily read by readers with a specific cultural background will not be easy for another. The study focused on the implementation of the brainling components in modifying passages (Pishghadam et al., 2023).

To achieve the objective, we administered an online English text with college-level readability and a modified version of another text with a similar readability level that incorporated the brainling components to Omani EFL learners. The obtained readability results were compared with those acquired from contents whose readability was measured by the Flesch index. The rationale behind using the Flesch formula was its popularity in readability studies. The reason for conducting the study among Omani learners was that Oman is a multicultural society and successfully supports the state of coexistence with multifarious ethnicities, languages, nationalities, backgrounds, etc. (Al-Raisi et al., 2019). Moreover, regarding language as an important ramification of cultural diversity in multicultural societies, we specifically delved into examining English as a widely spoken language in Oman. In this respect, the significance of the study was remarking on the possible association between the brainling pattern and readability as it might enlighten content developers to tailor content with regard to readers’ various characteristics.

Following the results of the cognitive (Huckin, 1983), emotional (Pishghadam & Abbasnejad, 2016), sensory (Alarcon et al., 2020), and cultural (Bailin & Grafstein, 2001) approaches to readability, we hypothesized that the efficiency of the brainling components may be highlighted in the answers to the readability-based Likert-scale questions. Moreover, we assume that the inclusion of cogling, emoling, sensoling, and cultuling components into texts could be manifested in the responses to the text difficulty-based Likert-scale question. The third hypothesis is that incorporating the components may make the text more understandable and can manifest differences in the readers’ reading comprehension test scores.

 

Theoretical Framework

Readability

Rendering comprehensible texts well suited to readers’ abilities has constantly been a challenge for writers, reviewers, and educators. Researchers (e.g., Chall & Dale, 1995; Crossley et al., 2023; Fry, 1968, 1975; Grabowski & Mathiebe, 2024; Kincaid et al., 1975; Spache, 1953) have endeavored to measure reading comprehension ability and predict the difficulty level of texts. As a solution to these concerns, the readability concept and readability formulae
(e.g., Automated Readability Index, Dale-Chall readability formula, Fog Count, and Flesch Reading Ease Formula) have been designed. Readability means the degree to which a reader understands a text (McLaughlin, 1969) and reads it at an optimal speed (Dale & Chall, 1948). Since the 20th century, readability studies have been commenced by Thorndike’s (1921) book. The traditional readability measures (e.g., Dale-Chall readability formula, Chall & Dale, 1995; Reading Ease Readability Formula, Flesch, 1948; Automated Readability Index, Fog Count, and Flesch Reading Ease Formula, Kincaid et al., 1975) heed mostly to the surface-level linguistic features. The classic readability formulae evaluated the effects of word/sentence length, and word frequency as the core components of the analysis. For instance, word length was the core component in Flesch’s (1948) study, and word frequency, percentage of unfamiliar words, and average sentence length were noticed in Dale-Chall’s (1948) formula. The formulae adopted proxy measures and calculated the “number of characters per word”, besides the “number of words per sentence” for word complexity and syntactic sophistication analyses (Crossley et al., 2019).

However, the classic readability formulae have been criticized for their overdependence on quantitative factors leading to the negligence of qualitative variables (e.g., sociocultural and psychological). In particular, they failed to account for the fact that various sociocultural groups may shape copious vocabulary repositories over time (Lenzner, 2014). Actually, in the assessment of this multicomponent skill, readers’ background knowledge may finely adjust their reading comprehension (Davison & Kantor, 1982). As a result of the shift of focus from text-based to reader-based analysis, the psycholinguistic-based model of analysis reckons the nexus between readers’ interaction and the textual factors (e.g., cohesion and coherence) (Gernsbacher, 1997).

A synthesis of the advances in corpus and computational linguistics together with psycholinguistic and discourse-level analyses had been accomplished in the Coh-Matrix formula (Graesser et al., 2004). The formula is based on discourse processing and computational linguistics and takes notice of world knowledge, language, and discourse features together with cohesion relations (Graesser et al., 2004). To improve the readability concept, Pishghadam and Abbasnejad (2016) were critical of previous models for not accounting for readers’ senses and emotions. To fill the gap, they propose an emotioncy-based model [sense (emotion + frequency)] of readability. Based on the model, readers’ comprehension correlated with their emotioncy levels (i.e., sensory-induced emotions). In this respect, the reading difficulty level depends on human-oriented experiences rather than textual factors. Based on advances in computational linguistics, Crossley and colleagues (2019) develop new readability formulae to delineate readers’ judgment of texts, reading speed, and understanding. Crossley et al. (2023) stated the CommonLit Ease of Readability corpus can provide unique readability scores.

Generally, readability formulae have been widely used in educational centers, though they have often been criticized for a variety of weaknesses. Classic studies were mainly centered on text understanding analysis; however, focuses shifted to ease of text processing. In modern readability models, multiple qualitative factors together with readers’ internal factors in reading comprehension and making meaning out of texts come to power. The following part is devoted to the explanation of a brain-based model delineating its components as the framework of this study and a possible solution for modifying readability.

 

Brainling

It has long been established that the human brain consists of three (i.e., the reptilian brain, limbic brain, and neocortex) integrated structures (MacLean, 1978). To regulate functions
(e.g., emotions, cognition, language, memory, etc.), the structures are interconnected and do not operate independently. From this perspective, Pishghadam and Ebrahimi (2020) introduce the “brainling” pattern to clear up the significance of cognition, emotions, senses, and culture in connection with language. Based on the brainling pattern, language reflects brain macro functions and the interactions between them facilitate effective communication (Pishghadam & Ebrahimi, 2020). But the complexity of the brain, language, and communication does not stop here. The brainling pattern portrays this intricacy by delineating four components. The brainling components are as follows:

Cogling (cognition + language) demonstrates the manifestation of language in thoughts shaping individuals’ interpretations of the world. In this view, individuals’ visions are in connection with their word repertoires. On top of that, the selection and arrangement of words emerge a multitude of effects on interlocutors during communication. In this regard, language is a tool for better thinking, and via language, humans can communicate their thoughts.

In the emoling (emotion + language) component, the language that a person uses has positive or negative emotions. In this light, some concepts are positive or negative in essence, called automatic emotion. However, the positivity or negativity of some events is cross-culturally and/or interpersonally different, called deliberative emotions. As a result, emotions involved in language are not fixed, but dynamic. In bidirectional effects, the game of language is in fact the game of emotions paving the way for the game of thoughts (Ghadirzade Toosy & Jajarmi, 2023; Jahani & Aminzadeh, 2024).

Cultuling (culture + language) mirrors the interconnection between language and culture (Pishghadam, 2013). Cultuling analysis foregrounds unique language structures and phrases of a language entrenched in a nation. Cultuling awareness has profound positive effects on socialization.

Sensoling (sense + language) encompasses physical and linguistic senses. While the physical sense refers to the sensory inputs and channels, the linguistic sense frames issues related to communication. Sensoling highlights the impacts of sensory inputs on language. If mapped and employed correctly, sensory inputs illustrate a clear picture of the environment. Optimal sensory involvement creates thick-slice sensory relations (Pishghadam, 2018).

Generally, for effective communication, attention to the interlocutors is as important as attention to their thoughts, emotions, senses, and culture. Brainling pattern portrays a unified image of language and the brain leading to changes in behaviors. In this respect, our idea is that how the brain and language assist readers in comprehending texts should be studied with particular reference to their brain-based unique characteristics and experiences (e.g., cognitive, sensory, cultural, and emotional features).

 

Methodology

Participants

The participants of the current study were 209 individuals who had intermediate levels of language proficiency (i.e., ‎30 to 40 out of 60) based on the Oxford Quick Placement Test (Allan, 1992). They were recruited from an educational institution in Oman using a convenience sampling method. The researchers selected the participants who were available and willing to participate in the study, without applying any specific criteria or randomization. The majority of the participants (94.7%) were Omani nationals, while the rest were from Yemen (2.4%), Tunisia (1.9%), and Jordan (1%). The gender distribution was fairly balanced, with 52.6% females and 47.4% males. The participants were from 17 to 37 years old (M = 20.07, SD = 2.69). The educational level of the participants varied from foundation to master, with 50.7% having a bachelor’s degree, 1.9% having a master’s degree, and 47.4% being in the foundation program.

 

Instrument

To conduct the study, two online English passages with general topics were utilized. The passages consisted of 500 words and to measure the readability level of the passages, the Flesch Reading Ease score formula (Flesch, 1948) was used. Their readability level was analyzed by https://readability formulas.com/. Accordingly, their Flesch Reading Ease scores were between 30-50 indicating the college student reading level. One of the texts remained unchanged, but the other was modified based on the concepts of cogling, emoling, sensoling, and cultuling. Adjusting a text based on the cogling concept involves simplifying complex ideas and structures to ensure that they align with readers’ cognitive processes. The emoling concept engages the use of words with positive or negative connotations to connect with the readers’ emotional levels. Furthermore, creating more vivid sensory experiences through the five senses enabled the application of sensoling. Finally, adapting the text to the readers’ cultural background and the incorporation of culturally relevant examples modified the text based on the cultuling concept. Notably, the readability level did not change after the modifications.
An example is provided to clarify the modification; the second sentence is an adjustment of the first one:

1) Our beautiful ecosystem is being systematically damaged by industrial pollution channeled into rivers like the Ganga and the Yamuna, nuclear wastes from atomic plants routed into the world’s oceans, and poisonous gases such as carbon dioxide mixed with the ozonosphere.

2) Our beautiful ecosystem is like a precious jewel that shines with different colors and shapes; however, it is being damaged by industrial pollution that goes into rivers like the Wadi Dayqah and the Wadi Bani Khalid, nuclear wastes that are dumped into the Arabian Sea, and poisonous gases that harm the ozone layer.

Following each reading text, five comprehension multiple-choice questions were designed. Afterward, as the second task, a 10-item Likert-scale readability (1= Strongly disagree; 2= Disagree; 3= Neutral; 4= Agree; and 5= Strongly agree) was created to evaluate the texts’ clarity and engagement. As clarity and engagement are two components of readability (Ghafourian et al., 2023; Hartley, 1994), each accounted for half of the questions. The clarity questions pertained to the clarity and understandability of the passages, their organization, ‎and coherency, the appropriateness of vocabulary and grammar, the clear presentation of main points and ‎concepts, and the avoidance of unnecessary jargon and technical terms. The engagement questions ‎focused on whether the passages were interesting and engaging, suitable for the intended purpose and ‎audience, provided examples and illustrations to support arguments, stimulated readers’ curiosity, and ‎motivated them to learn more or take action. ‎Finally, task 3 involved a question about the text’s difficulty level, asking participants to rate it on a scale from 1= Extremely easy to 5= Extremely difficult (See Appendix).

 

Procedure

The two passages with general topics and a similar difficulty level were inserted into Google Forms. As the first step, the participants were asked to complete the demographic section It was followed by the first passage, which participants were asked to read carefully before selecting the best answer from the provided options based on the information presented in the passage. The aim of Task 1 was to ensure the extent to which the participants truly comprehended the passage. In Task 2, they were asked to indicate the degree to which they agree or disagree with each of the statements about the text they have read, using a 5-point Likert scale (from 1= strongly disagree to 5= strongly agree). Then, in Task 3, they were asked to determine the difficulty level of the text from 1= Extremely easy to 5= Extremely difficult. The same format of the tasks was repeated for the second text which was modified by the brainling model.

 

Data Collection

The recorded data were imported into IBM SPSS Statistics software (version 25) to check the reliability of the Readability Scales and the underlying subconstructs through Cronbach’s alpha and test-retest model. Then, AMOS was run to verify the construct validity of the Readability Scale. In the third step, correlational analysis for readability, reading score, and text difficulty was examined through the Pearson product-moment correlation. Finally, a paired samples
t-test was conducted to compare the readability, reading scores, and text difficulty between text 1 and text 2.

 

Results

Table 1 shows descriptive statistics for the readability of text 1 and text 2.

 

Table 1. Descriptive Statistics for the Readability of Text 1 and Text 2

 

Min

Max

Mean

SD

Readability of Text 1

10.00

49.00

27.77

8.36

Clarity

5.00

25.00

14.01

4.20

Engagement

5.00

25.00

13.76

4.63

Reading Score

0

5

2.10

1.18

Text Difficulty

1

5

3.44

.96

Readability of Text 2

12.00

50.00

33.17

9.72

Clarity

5.00

25.00

16.52

5.30

Engagement

5.00

25.00

16.65

5.03

Reading Score

0

5

2.51

1.36

Text Difficulty

1

5

2.84

1.17

 

Reliability Estimates

Table 2 shows the reliability estimates for the Readability Scales, in addition to its underlying subconstructs. As can be seen, Cronbach’s Alpha estimates are all above .70, which is considered acceptable. Moreover, the reliability of the text difficulty item was verified through the test-retest method. The correlation coefficient of the measured values, calculated at two separate time intervals, was estimated as r = .94.

 

Table 2. Reliability Estimates for the Readability Scale

 

N of Items

Cronbach’s Alpha

Readability

10

.91

Clarity

5

.88

Engagement

5

.84

 

Validation of the Readability Scale

In order to verify the construct validity of the Readability Scale, Confirmatory Factor Analysis (CFA) was used. Harman’s single-factor test was calculated before running the CFA. Based on the result of the test, the first factor accounted for only 47.05% of the variance and confirmed the construct’s multi-dimensionality. The scale has two subconstructs of Clarity (5 items) and Engagement (5 items). To improve model fit, no items were eliminated from the scale. See Table 3 for the Goodness-of-fit indices.

 

 

Figure 1. Measurement Model for the Readability Scale

 

To check if the model fits the data, goodness of fit indices were measured in Amos. The criterion for accepting the model is different according to different researchers. In the present study, values for χ²/df (i.e., chi-square index divided by the degrees of freedom (χ²/df) should be less than 3 (Ullman, 2001), Tucker-Lewis Index (TLI) and Comparative Fit Index (CFI) were over .90, and Root Mean Square Error of Approximation (RMSEA) and Standardized Root Mean Squared Error (SRMR) were equal to or less than .08 (Browne & Cudeck, 1993).

 

 

 

Table 3. Goodness of Fit Indices for the Model

Models

χ²/df

df

CFI

TLI

RMSEA

SRMR

The Readability Scale

2.04

32

.97

.96

.07

.04

 

Correlational Analysis

The Pearson product-moment correlation was used to examine the possible relationships between readability, reading score, and text difficulty for text 1 (Table 4) and text 2 (Table 5). The results, presented in Tables 4 and 5, show that some of the variables have statistically significant correlations with each other.

According to Table 4, the readability of text 1 has a positive relationship with the participants’ reading score (r = .21, p < .01) and a significant negative relationship with text difficulty (r = -.48, p < .01). There exists no significant relationship between text difficulty and the participants’ reading score.

 

Table 4. Correlational Analysis for Readability, Reading Score, and Text Difficulty (Text 1)

 

Readability

Reading Score

Text Difficulty

Readability

1

 

 

Reading Score

.21**

1

 

Text Difficulty

-.48**

-.03

1

**. Correlation is significant at the 0.01 level (2-tailed).

 

According to Table 5, readability of text 2 has a positive relationship with the participants’ reading score (r = .44, p < .01) and a significant negative relationship with text difficulty (r = -.49, p < .01). There is also a significant negative relationship between text difficulty and the participants’ reading score (r = -.45, p < .01).

 

Table 5. Correlational Analysis for Readability, Reading Score, and Text Difficulty (Text 2)

 

Readability

Reading Score

Text Difficulty

Readability

1

 

 

Reading Score

.44**

1

 

Text Difficulty

-.49**

-.45**

1

**. Correlation is significant at the 0.01 level (2-tailed).

 

 

 

 

Mean Differences

In order to examine if there are any significant differences in the readability, reading score, and text difficulty between text 1 and text 2, a series of paired samples t-tests were run.

As Table 6 shows, there are significant differences in readability (t (208) = -6.49, p = 0.000), reading score (t (208) = -3.31, p = 0.001), and text difficulty (t (208) = 5.92, p = 0.000) between text 1 and text 2. Text 2 appears to be easier than text 1 and, therefore, has higher readability and reading scores.

 

Table 6. Paired Samples t-Test for Readability, Reading Score, and Text Difficulty of Text 1 and Text 2

 

Variable

N

Mean

SD

df

t

Sig. (2-tailed)

Readability

Text 1

209

27.77

8.36

208

-6.49

.000

Text 2

209

33.17

9.72

Reading Score

Text 1

209

2.10

1.18

208

-3.31

.001

Text 2

209

2.51

1.36

Text Difficulty

Text 1

209

3.44

.96

208

5.92

.000

Text 2

209

2.84

1.17

Discussion

Text readability which pertains to the comprehensibility of words and sentences is an attribute of multiple factors like clarity, engagement, and text difficulty (Hargis et al., 2004).
To examine this crucial factor in language learning, this study analyzed two texts that had similar readability scores according to the Flesch Reading Ease score readability formula, targeting Arab English language learners. Notably, one of the texts incorporated brainling theory. This part aims to provide a comprehensive discussion of the findings obtained through a five-point Likert readability scale, a text difficulty scale, and reading comprehension scores.

The results confirmed our first hypothesis, which acknowledged that constructing the text based on brainling components (i.e., cogling, emoling, sensoling, and cultuling) resulted in higher levels of readability compared to the text created using the Flesch Reading Ease score readability formula. Specifically, the significant differences in the participants’ responses to the five-point Likert readability scale between the two texts evidenced that the quadripartite model of the brainling theory significantly enhanced clarity and engagement. Several studies have also integrated readability into information quality assessment, highlighting its presence in clarity (e.g., Hartley, 1994; Temnikova et al., 2015; Velez & Ashworth, 2007) and engagement (e.g., Ghafourian et al., 2023; Leonhardt & Makienko, 2018). In particular, well-organized and coherent texts (Meyer, 2003) with appropriate vocabulary and grammar (Chall & Dale, 1995; Zamanian & Heydari, 2012; Worrall et al., 2020), as well as clear explanations of main points and concepts (Heydari & Riazi, 2012) are considered more readable. These results are in line with the findings of Sanatipour et al. (2024), which employed the brainling model to systematically analyze commonly used English textbooks in both private and state educational settings. Their study entailed a specific analysis of the reading comprehension difficulty (cogling), the cultural appropriateness of content and visuals (cultuling), the presence of positive and negative language and the fostering of motivation (emoling), and the engagement of sensory experiences (sensoling). Significant differences were identified among the English textbooks used in different educational settings. According to teachers’ and students’ responses, the lowest component used in state schools was emoling, while the lowest component in textbooks used in private schools was cultuling. Sanatipouret al.’s (2024) analysis revealed the absence of certain components derived from the brainling model in these textbooks.

Building upon this understanding, the current study compared the comprehensibility of two texts with the same Formula Readability score, while incorporating brainling components into one of the texts. The findings revealed that the modified text exhibited a higher readability score compared to the other, indicating the positive effect of incorporating brainling elements on improving comprehensibility. Therefore, it can be assumed that adjusting a text based on brainling principles can be helpful for language learners. Such an account provides the implication that a text amended according to the brainling principles would be more readable, thereby enhancing clarity and fostering engagement. Language learners would benefit from reading texts that incorporate cogling, emoling, sensoling, and cultuling.

Furthermore, considering the Internet as an important source of information, studies (Gallagher et al., 2017; Worrall et al., 2020) have analyzed the readability of online texts. When examining online texts, Gallagher et al. (2017) reported a lack of correlation among the readability formulae and measures. Based on their findings, they mentioned that “online texts were the most disparate with respect to text difficulty” (p. 1). Worrall et al. (2020) also found significant differences in readability scores between various web pages. They emphasized the importance of ensuring universal readability in online texts and urged webpage producers to be aware of readability levels to enhance understanding. Although these authors compared and analyzed the comprehensibility of online texts, they failed to provide any practical solutions. In light of our findings on clarity and engagement, the brainling approach can be suggested as a practical solution to this problem.

The third task results indicated that the brainling-modified text received higher mean scores for ease of understanding compared to the unchanged text. In other words, the participants found the brainling-modified text as easier despite both texts having similar Flesch Reading Ease scores. Considering text difficulty as a key indicator of readability (Cunningham & Anne Mesmer, 2014; Jian et al., 2022; Zhang, 2022), we deduced that modifying texts based on the brainling components could enhance comprehension. To be specific, the brainling model goes beyond treating readability as a singular or multiple factor and considers the interaction among brain structures. Consequently, we found that brainling-based modifications could lower the text difficulty level. These conclusions were further reinforced by the findings on the readers’ reading comprehension scores.

The increase in mean scores in comprehending the brainling-based text supports the third hypothesis, which posits that incorporating cogling, emoling, sensoling, and cultuling components may make the text more understandable and can manifest differences in the readers’ reading comprehension test scores, indicating that the scores are significantly improved through these modifications. Considering reading scores as a means of assessing readers’ comprehension (Scott, 2008), we found that incorporating readers’ cognition, emotion, sense, and culture in conjunction with language can enhance their comprehension. We also deduced that attending to the brain’s structure is required to provide a readable text. In particular, reading comprehension involves mental processes that can be promoted by aligning with the brain’s structure and function, rather than relying solely on formulae. Actually, material developers have used formulae results as “explanations of difficulty” rather than “predictors of difficulty” (Crossley et al., 2008). Such scores can emanate the erroneous impression that embedding short and uncomplicated words in short sentences can make texts readable. We imply that the logic behind the formulae cannot adequately represent readability since they may overlook readers’ fundamental characteristics.

 

Conclusion

Overall, our findings demonstrate that modifying a text based on the brainling model, which combines brain and language structure, is a significant factor in text modification. However, this consideration has been overlooked in the creation of readable texts. This study does not negate the role of classical factors (e.g., word, sentence, grammatical complexities, frequency, etc.) in readers’ cognitive demands. We also do not argue that readability formulae’s results are untrustworthy; rather, they fail to capture some essential reader features. Our idea is that exploring how the brain and language assist readers in comprehending texts should involve studying their brain-based unique characteristics and experiences (e.g., cognitive, sensory, cultural, and emotional features). Thus, we believe that studies should shift their focus from text- and reader-based analyses to considering readers’ brain structures.

Our findings suggest that implementing brainling components will provide content developers with the opportunity to tailor online content to readers’ basic characteristics.
In particular, considering Oman is a multicultural society with multifarious ethnicities, languages, nationalities, backgrounds, etc. (Al-Raisi et al., 2019), targeting text comprehensibility can support coexistence among culturally diverse members and alleviate challenges such as misunderstanding that may hinder goal achievement. In this vein, adopting the brainling approach can be a solution for fostering mutual understanding and collaboration. In general, this study has fruitful implications for educational communities. With regard to the increasing prominence of standardized tests and texts, the results can help teachers and test designers in developing readable texts. This, in turn, would enable teachers to identify readable texts that incorporate an appropriate match to readers’ cognition, emotions, senses, and culture, facilitating reading comprehension and conceptual processing.

While the research has achieved its objectives and has been meticulously prepared, there are several limitations to consider. The first limitation is directed towards the method of study. To obtain a more detailed insight into the readability level, complementary studies could be conducted with more than two texts and extended passages involving readers with various proficiency levels. Furthermore, correlating the results of the brainling-based text with unmodified texts, whose readability is determined using multiple readability formulas, could further strengthen the study’s findings. Additionally, future research could replicate the investigation with a larger sample size and more diverse groups.

 

Funding

The research leading to these results has received funding from the Ministry of Higher Education, Research and Innovation (MoHERI) of the Sultanate of Oman under the Block Funding Program (Block Funding Agreement No.: MoHERI/BFP/GULF/2023, Project Code: (BFP/RGP/ICT/23/135).

 

Ahmad, Al H., H., & Hasan, Y. N. (2024). Applied Arabic textbook readability level for university students at Al-Balqa Applied University. Theory and Practice in Language Studies, 14(1), 186-191. https://doi.org/10.17507/tpls.1401.22
Alarcon, R. A., Moreno, L., & Martínez, P. (2020). Word-sense disambiguation system for text readability. In A. Pereira, M. Ribera, & C. Yang (Eds.), Proceedings of the 9th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-Exclusion (pp. 147-152). Association for Computing Machinery. https://doi.org/ 10.1145/3439231.3439257
Allan, D. (1992). The Oxford quick placement test. Oxford University Press.
Al-Raisi, M. Y., Rawahi, B. S. A., Omrani, N. H. A., Hooti, N. K. A., & Porkodi, S. (2019). The effect of cultural diversity on employees’ performance and productivity in Shell Oman marketing company SAOG. International Journal of Multidisciplinary Research and Publications, 1(9), 14-17.
Bailin, A., & Grafstein, A. (2001). The linguistic assumptions underlying readability formulae: A critique. Language and Communication, 21(3), 285-301. https://doi.org/ 10.1016/S0271-5309 (01)00005-2
Browne, M. W., & Cudeck, R. (1993). Alternative ways of assessing model fit. In K. A. Bollen & J. S. Long (Eds.), Testing structural equation models (pp. 136-162). Sage.
Chall, J. S., & Dale, E. (1995). Readability revisited: The new Dale-Chall readability formula. Brookline Books.
Collins‐Thompson, K., & Callan, J. (2005). Predicting reading difficulty with statistical language models. Journal of the American Society for Information Science and Technology, 56(13), 1448-1462. https://doi.org/10.1002/asi.20243
Crossley, S. A., Greenfield, J., & McNamara, D. S. (2008). Assessing text readability using cognitively based indices. TESOL Quarterly, 42(3), 475-493. https://doi.org/10.1002/j.1545-7249.2008. tb00142.x
Crossley, S. A., Heintz, A., Choi, J. S., Batchelor, J., Karimi, M., & Malatinszky, A. (2023).
A large-scaled corpus for assessing text readability. Behavior Research Methods, 55, 491–507. https://doi.org/10.3758/s13428-022-01802-x
Crossley, S. A., Skalicky, S., & Dascalu, M. (2019). Moving beyond classic readability formulas: New methods and new models. Journal of Research in Reading, 42(3-4),
541-561. https://doi.org/ 10.1111/1467-9817.12283
Cunningham, J. W., & Anne Mesmer, H. (2014). Quantitative measurement of text difficulty: What’s the use?. The Elementary School Journal, 115(2), 255–269. https://doi.org/10.1086/678292
Dale, E., & Chall, J. S. (1948). A formula for predicting readability: Instructions. Educational Research Bulletin, 27(1), 37-54.
Davison, A., & Kantor, R. N. (1982). On the failure of readability formulas to define readable texts: A case study from adaptations. Reading Research Quarterly, 17(2), 187–209. https://doi.org/ 10.2307/747483
Flesch, R. (1948). A new readability yardstick. Journal of Applied Psychology, 32(3), 221-233. https://doi.org/10.1037/h0057532
Fry, E. (1968). A readability formula that saves time. Journal of Reading, 11(7), 513-578.
Fry, E. (1975). The readability principle. Language Arts, 52(6), 847-851.
Gallagher, T., Fazio, X. E., & Ciampa, K. (2017). A comparison of readability in science-based texts: Implications for elementary teachers. Canadian Journal of Education/Revue Canadienne de L’éducation, 40(1), 1-29.
Gernsbacher, M. A. (1997). Coherence cues mapping during comprehension. In J. Costermans & M. Fayol (Eds.), Processing interclausal relationships. Studies in the production and comprehension of text (pp. 3-22). Erlbaum.
Ghadirzade Toosy, S. & Jajarmi, H. (2023). ESQ in L2 willingness to communicate and communicative ability. Business, Communication, and Technology, 2(2), 15-27. https://doi.org/10.56632/bct. 2023.2202
Ghafourian, Y., Hanbury, A., Knoth, P. (2023). Readability measures as predictors of understandability and engagement in searching to learn. In O. Alonso, H. Cousijn, G. Silvello, G. M. Marrero, C. Teixeira Lopes, & S. Marchesin (Eds.), Linking theory and practice of digital libraries (Vol. 14241, pp. 173-181). Springer. https://doi.org/10.1007/978-3-031-43849-3_15
Grabowski, J., & Mathiebe, M. (2024). A direct functional measure of text quality: Did the reader understand?. Written Communication, 41(2), 203-229. https://doi.org/10.1177/ 07410883231222952
Grabowski, J., Weinzierl, C., & Schmitt, M. (2010). Second and fourth graders’ copying ability: From graphical to linguistic processing. Journal of Research in Reading, 33(1), 39-53. https://doi.org/10.1111/j.1467-9817.2009.01431.x
Graesser, A. C., McNamara, D. S., Louwerse, M. M., & Cai, Z. (2004). Coh-Metrix: Analysis of text on cohesion and language. Behavior Research Methods, Instruments, and Computers, 36(2), 193-202. https://doi.org/10.3758/BF03195564
Hargis, G., Carey, M., Hernandez, A. K., Hughes, P., Longo, D., Rouiller, S., & Wilde, E. (2004). Developing quality technical information: A handbook for writers and editors (2nd ed.). Prentice Hall Professional Technical Reference.
Hartley, J. (1994). Three ways to improve the clarity of journal abstracts. British Journal of Educational Psychology, 64(2), 331-343. https://doi.org/10.1111/j.2044-8279.1994.tb01106.x
Heydari, P., & Riazi, A. M. (2012). Readability of texts: Human evaluation versus computer index. Mediterranean Journal of Social Sciences, 3(1), 177-190.
Huckin, T. N. (1983). A cognitive approach to readability. In P. Anderson, J. Brockman, & C. Miller (Eds.), New essays in technical and scientific communication: Research, theory, practice (1st ed., pp. 90-108). Routledge. http://doi.org/10.4324/9781315224060-8
Jahani, Z., & Aminzadeh, A. (2024). Emo-sensory intelligence in high school teachers: Examining gender and age dynamics in Iranian educational settings. Cognition, Emotion, and Education, 2(1), 55-65. https://doi.org/10.22034/cee.2024.431909.1015
Jian, L., Xiang, H., & Le, G. (2022). English text readability measurement based on convolutional neural network: A hybrid network model. Computational Intelligence and Neuroscience, 2022, Article 6984586. https://doi.org/10.1155/2022/6984586
Kate, R., Luo, X., Patwardhan, S., Franz, M., Florian, R., Mooney, R., Roukos, S., & Welty, C. (2010). Learning to predict readability using diverse linguistic features. In C.-R. Huang, & D. Jurafsky (Eds.), Proceedings of the 23rd International Conference on Computational Linguistics (pp. 546-554). Coling 2010 Organizing Committee.
Kincaid, J. P., Fishburne, R. P., Rogers, R. L., & Chissom, B. S. (1975). Derivation of new readability formulas (automated readability index, Fog Count and Flesch Reading Ease Formula) for navy enlisted personnel. Institute for Simulation and Training, 56, 1-39.
Lenzner, T. (2014). Are readability formulas valid tools for assessing survey question difficulty?. Sociological Methods and Research, 43(4), 677-698. https://doi.org/10.1177/ 0049124113513436
Leonhardt, J. M., & Makienko, I. (2018). Keep it simple, readability increases engagement on Twitter: An abstract. In N. Krey & P. Rossi (Eds.), Back to the future: Using marketing basics to provide customer value (pp. 333-334). Springer. https://doi.org/10.1007/978-3-319-66023-3_116
MacLean, P. D. A. (1978). Mind of three minds: Educating the triune brain. National Society for the Study of Education.
McLaughlin, G. H. (1969). SMOG grading: A new readability formula. Journal of Reading, 12(8), 639-646.
McNamara, D. S., & Kintsch, W. (1996). Learning from texts: Effects of prior knowledge and text coherence. Discourse Processes, 22(3), 247-288. https://doi.org/10.1080/ 01638539609544975
Mesmer, H. A., Cunningham, J. W., & Hiebert, E. H. (2012). Toward a theoretical model of text complexity for the early grades: Learning from the past, anticipating the future. Reading Research Quarterly47(3), 235-258. https://doi.org/10.1002/rrq.019
Meyer, B. J. F. (2003). Text coherence and readability. Topics in Language Disorders, 23(3), 204–224. https://doi.org/10.1097/00011363-200307000-00007
Pishghadam, R. (2013). Introducing cultuling as a dynamic tool in culturology of language. Language and Translation Studies, 45(4), 47-62.
Pishghadam, R. (2018). An introduction to thin-slice sensory education: Less is more [Conference presentation]. International Academic Conference on Economics, Business and Social Sciences, Tbilisi, Georgia.
Pishghadam, R., & Abbasnejad, H. (2016). Emotioncy: A potential measure of readability. International Electronic Journal of Elementary Education, 9(1), 109-123.
Pishghadam, R., Al Abdwani, T., Jajarmi, H., & Shayesteh, S. (2023). Enhancing general and language aptitude tests by incorporating cultural and emo-sensory constructs. International Journal of Society, Culture and Language, 11(3), 1-12. https://doi.org/10.22034/ijscl.2023.704891
Pishghadam, R., & Ebrahimi, S. (2020). Introducing the “brainling” model and examining its role in effective communication: A moving beyond communicative competence. Language and Translation Studies, 53(3), 1-32. https://doi.org/10.22067/lts.v53i3.87911
Pitler, E., & Nenkova, A. (2008). Revisiting readability: A unified framework for predicting text quality. In M. Lapata & H. T. Ng (Eds.), Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing (pp. 186-195). Association for Computational Linguistics.
Sanatipour, S., Shayesteh, S., Pishghadam, R., & Boustani, N. (2024). Designing and validating a brainling model: A case of high school and language school textbooks. Iranian Journal of Applied Language Studies, 16(1), 1-20.
Schriver, K. A. (2000). Readability formulas in the new millennium: What’s the use?. ACM Journal of Computer Documentation, 24(3), 138-140. https://doi.org/10.1145/ 344599.344638
Scott, D. B. (2008). Assessing text processing: A comparison of four methods. Journal of Literacy Research, 40(3), 290-316. https://doi.org/10.1080/10862960802502162
Smith, D., Stenner, A. J., Horabin, I., & Smith, M. (1989). The Lexile scale in theory and practice: Final report. MetaMetrics.
Spache, G. (1953). A new readability formula for primary-grade reading materials. The Elementary School Journal, 53(7), 410-413.
Temnikova, I., Vieweg, S., & Castillo, C. (2015). The case for readability of crisis communications in social media. In A. Gangemi, S. Leonardi, & A. Panconesi (Eds.), Proceedings of the 24th International Conference on World Wide Web (pp. 1245-1250). Association for Computing Machinery. https://doi.org/10.1145/2740908.2741718
Thorndike, E. L. (1921). The teacher’s word book. Teachers College Press.
Ullman, J. B. (2001). Structural equation modeling. In B. G. Tabachnick & L. S. Fidell (Eds.), Using multivariate statistics (4th ed., pp. 653-771). Pearson Education.
Velez, P., & Ashworth, S. D. (2007). The impact of item readability on the endorsement of the midpoint response in surveys. Survey Research Methods, 1(2), 69-74. https://doi.org/ 10.18148/srm/ 2007.v1i2.76
Worrall, A. P., Connolly, M. J., O’Neill, A., O’Doherty, M., Thornton, K. P., McNally, C., McConkey, S. J., & De Barra, E. (2020). Readability of online COVID-19 health information: A comparison between four English speaking countries. BMC Public Health, 20(1), Article 1635. https://doi.org/10.1186/s12889-020-09710-5
Zamanian, M., & Heydari, P. (2012). Readability of texts: State of the art. Theory and Practice in Language Studies, 2(1), 43-53. https://doi.org/10.4304/tpls.2.1.43-53
Zhang, B. (2022). Readability analysis of texts in college English textbooks and reading passages in CET-6. Open Access Library Journal, 9(11), Article e9445. https://doi.org/10.4236/oalib. 1109445