Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
970
result(s) for
"Criterion referenced"
Sort by:
Cognitive and Motivational Challenges in Writing: Studying the Relation With Writing Performance Across Students' Gender and Achievement Level
by
De Smedt, Fien
,
De Naeghel, Jessie
,
Van Keer, Hilde
in
Academic Achievement
,
Achievement
,
Achievement gap
2018
In the past, several assessment reports on writing repeatedly showed that elementary school students do not develop the essential writing skills to be successful in school. In this respect, prior research has pointed to the fact that cognitive and motivational challenges are at the root of the rather basic level of elementary students' writing performance. Additionally, previous research has revealed gender and achievement-level differences in elementary students' writing. In view of providing effective writing instruction for all students to overcome writing difficulties, the present study provides more indepth insight into (a) how cognitive and motivational challenges mediate and correlate with students' writing performance and (b) whether and how these relations vary for boys and girls and for writers of different achievement levels. In the present study, 1,577 fifth- and sixth-grade students completed questionnaires regarding their writing self-efficacy, writing motivation, and writing strategies. In addition, half of the students completed two writing tests, respectively focusing on the informational or narrative text genre. Based on multiple group structural equation modeling (MG-SEM), we put forward two models: a MG-SEM model for boys and girls and a MG-SEM model for low, average, and high achievers. The results underline the importance of studying writing models for different groups of students in order to gain more refined insight into the complex interplay between motivational and cognitive challenges related to students' writing performance.
Journal Article
Technology functions for personalized learning in learner-centered schools
by
Huh, Yeol
,
Reigeluth, Charles M.
,
Lee, Dabae
in
Active Learning
,
Competency Based Education
,
Criterion Referenced Tests
2018
Personalized Learning (PL) has been widely promoted. Despite the increasing interest in PL, it is difficult to be implemented, because it can be complicated, costly, and even impossible without the help of powerful and advanced technology. This national survey study aimed at systematically investigating technology usage and needs of teachers in learner-centered schools in the U.S based on the conceptual framework of the Personalized Integrated Education System (PIES). PIES specifies four major functions: recordkeeping, planning, instruction, and assessment. A total of 308 learner-centered schools were identified that met at least three of the five criteria of PL: (1) personalized learning plans, (2) competency-based student progress, (3) criterion-referenced assessment, (4) problem- or project-based learning, and (5) multi-year mentoring. Survey responses of 245 teachers from 41 schools were analyzed. Results indicate that only 12% of teachers responded that they had a technology system that integrated the four major functions. Among the rest, 21% reported that they had no such systems. Technology was most widely used for planning and instruction but not for recordkeeping and assessment.
Journal Article
Physical Literacy as A Framework of Assessment and Intervention for Children and Youth with Developmental Coordination Disorder: A Narrative Critical Review of Conventional Practice and Proposal for Future Directions
2020
A framework of literacy may have roles to play in the assessment and treatment of children and youth with developmental disorders. This review aims to evaluate the conventional practice of assessment and treatment for children and youth with a developmental disorder in the physical domain, called developmental coordination disorder (DCD), and explore how the framework of physical literacy could contribute to the advancement of the current practice. This study adopts a method of narrative critical review based on a non-systematic search for its broad coverage to provide insights into the trend and future alternative directions. Over recent decades, children and youth with DCD have been typically assessed with standardized norm-referenced tests, before and after task-oriented intervention, for aiding diagnosis and evaluating the treatment effect. However, a recent high-quality systematic review showed limited evidence for the treatment effect assessed by the tests. Here, a framework of physical literacy is proposed to be used as an alternative to the conventional practice by recalibrating treatment goals and modifying the assessment and intervention approaches; criterion-referenced real-life authentic assessment and activities are encouraged with an emphasis on the enjoyment of movement and value of physical activity towards the attainment of physically active and healthy lifestyle goals from a lifespan perspective. The application of the physical literacy framework to the assessment and treatment of DCD needs to be further examined conceptually and empirically, while exploring a potential contribution of the literacy framework to transform the conventional assessment and treatment of children and youth with other developmental disorders.
Journal Article
Student evaluations of teaching: teaching quantitative courses can be hazardous to one’s career
2017
Anonymous student evaluations of teaching (SETs) are used by colleges and universities to measure teaching effectiveness and to make decisions about faculty hiring, firing, re-appointment, promotion, tenure, and merit pay. Although numerous studies have found that SETs correlate with various teaching effectiveness irrelevant factors (TEIFs) such as subject, class size, and grading standards, it has been argued that such correlations are small and do not undermine the validity of SETs as measures of professors’ teaching effectiveness. However, previous research has generally used inappropriate parametric statistics and effect sizes to examine and to evaluate the significance of TEIFs on personnel decisions. Accordingly, we examined the influence of quantitative vs. non-quantitative courses on SET ratings and SET based personnel decisions using 14,872 publicly posted class evaluations where each evaluation represents a summary of SET ratings provided by individual students responding in each class. In total, 325,538 individual student evaluations from a US mid-size university contributed to theses class evaluations. The results demonstrate that class subject (math vs. English) is strongly associated with SET ratings, has a substantial impact on professors being labeled satisfactory vs. unsatisfactory and excellent vs. non-excellent, and the impact varies substantially depending on the criteria used to classify professors as satisfactory vs. unsatisfactory. Professors teaching quantitative courses are far more likely not to receive tenure, promotion, and/or merit pay when their performance is evaluated against common standards.
Journal Article
The Parenting to Reduce Adolescent Depression and Anxiety Scale: Assessing parental concordance with parenting guidelines for the prevention of adolescent depression and anxiety disorders
by
Mackinnon, Andrew J.
,
Cardamone-Breen, Mairead C.
,
Yap, Marie B.H.
in
Adolescence
,
Adolescent depression
,
Adolescents
2017
Despite substantial evidence demonstrating numerous parental risk and protective factors for the development of adolescent depression and anxiety disorders, there is currently no single measure that assesses these parenting factors. To address this gap, we developed the
as a criterion-referenced measure of parental concordance with a set of evidence-based parenting guidelines for the prevention of adolescent depression and anxiety disorders. In this paper, we used a sample of Australian parents of adolescents to: (1) validate the PRADAS as a criterion-referenced measure; (2) examine parental concordance with the guidelines in the sample; and (3) examine correlates of parental concordance with the guidelines.
Seven hundred eleven parents completed the PRADAS, as well as two established parenting measures, and parent-report measures of adolescent depression and anxiety symptoms. Six hundred sixty adolescent participants (aged 12-15) also completed the symptom measures. Concordance with the guidelines was assessed via nine subscale scores and a total score. Reliability of the scores was assessed with an estimate of the agreement coefficient, as well as 1-month test-retest reliability. Convergent validity was examined via correlations between the scale and two established parenting measures.
One proposed subscale was removed from the final version of the scale, resulting in a total of eight subscales. Reliability was high for the total score, and acceptable to high for seven of the eight subscales. One-month test-retest reliability was acceptable to high for the total score. Convergent validity was supported by moderate to high correlations with two established measures of parenting. Overall, rates of parental concordance with the guidelines were low in our sample. Higher scores were associated with being female and higher levels of parental education. Greater parental concordance with the guidelines was associated with fewer symptoms of depression and anxiety in adolescent participants.
This initial validation study provides preliminary support for the reliability and validity of the PRADAS. The scale has potential for use in both clinical and research settings. It may be used to identify parents' strengths and potential targets for intervention, and as an outcome measure in studies of preventive parenting interventions.
Journal Article
A University–Lab School Writing Partnership Project: Benefits of Curriculum‐Based Measures and Intervention for Students With Learning Differences
by
Lindo, Endia
,
Keaney, Lauren
,
Faggella‐Luby, Michael
in
1-Early childhood
,
2-Childhood
,
3-Early adolescence
2020
This department highlights literacy leaders across the globe who are working in the field of literacy education. It features a variety of invited authors whose work is timely and relevant to the theme of access.
Journal Article
Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability
by
Scielzo, Shannon A.
,
Ryder, Hilary F.
,
Abdelfattah, Kareem
in
Accreditation
,
Criterion-referenced tests
,
Internal medicine
2023
Background: Little research to date has examined the quality of data obtained from resident performance evaluations. This study sought to address this need and compared inter-rater reliability obtained from norm-referenced and criterion-referenced evaluation scaling approaches for faculty completing resident performance evaluations.Methods: Resident performance evaluation data were examined from 2 institutions (3 programs, 2 internal medicine and 1 surgery; 426 residents in total), with 4 evaluation forms: 2 criterion-referenced (1 with an additional norm-referenced item) and 2 norm-referenced. Faculty inter-rater reliability was calculated with intraclass correlation coefficients (ICCs) (1,10) for each competency area within the form. ICCs were transformed to z-scores, and 95% CIs were computed. Reliabilities for each evaluation form and competency, averages within competency, and averages within scaling type were examined.Results: Inter-rater reliability averages were higher for all competencies that used criterion-referenced scaling relative to those that used norm-referenced scaling. Aggregate scores of all independent categories (competencies and the items assessing overall competence) for criterion-referenced scaling demonstrated higher reliability (z=1.37, CI 1.26-1.48) than norm-referenced scaling (z=0.88, CI 0.77-0.99). Moreover, examination of the distributions of composite scores (average of all competencies and raters for each individual being rated) suggested that the criterion-referenced evaluations better represented the performance continuum.Conclusion: Criterion-referenced evaluation approaches appear to provide superior inter-rater reliability relative to norm-referenced evaluation scaling approaches. Although more research is needed to identify resident evaluation best practices, using criterion-referenced scaling may provide more valid data than norm-referenced scaling.
Journal Article
Criterion-Referenced Assessment of Intelligence as Adaptation to the Environment: Is It Possible, Plausible, or Practical?
by
Landy, Jenna
,
Chowkase, Aakash
,
Parra-Martinez, Fabio Andres
in
achievement
,
Achievement tests
,
Adaptation
2022
Criterion-referenced testing is usually applied to the assessment of achievement. In this article, we suggest how it can also be applied to the assessment of adaptive intelligence, that is, intelligence as adaptation to the environment. In the era of the Anthropocene, we argue that adaptive intelligence is what is most important not only for individual success, but also for success in terms of preservation of the world as we know it. We define criterion-referenced testing and compare it to norm-referenced testing. We then discuss two kinds of scoring of criterion-referenced testing, namely, with respect to external criteria and with respect to internal (theory-based) criteria. We then discuss past research on intelligence that could be viewed as criterion-referenced. Finally, we suggest how criterion-referencing could be applied to the assessment of adaptive intelligence.
Journal Article
Diagnostic Decisions of Language Complexity Using Informal Language Assessment Measures
2022
Purpose: The current study aimed to understand the clinical decision-making skills of school-based speech-language pathologists (SLPs) using narrative and expository discourse information from three sources: perception of language through listening to language alone, standardized criterion-referenced narrative assessment data, and word- and morpheme-level language sample analysis data. Method: Twenty-eight current school-based SLPs participated in the study. During this study, participants rated language quality and made decisions regarding the provision of language services after being provided information from informal assessment measures. Results: SLPs' ratings of language quality and complexity varied across the story retell, wordless picture book generation, and expository samples. There was a lack of consistency in ratings within each context across areas of clarity, sample complexity, language complexity, and vocabulary across all SLPs. Self-reported factors that influenced SLPs' ratings included components of structure, syntax, and semantics. SLPs did not indicate a need for services after listening alone. When provided with criterion-referenced narrative assessment scores and word- and morpheme-level language sample data, more SLPs made a recommendation for services. Conclusions: The study demonstrates the need for objective language measure data during diagnostic decision making. Additionally, SLPs may not realize the information obtained from real-time perception, and analysis of language samples may be an unreliable and inconsistent picture of a child's language abilities. The results of this study highlight the need to continue to rely on multiple sources of assessment data. SLPs should continue to incorporate systematic methods to minimize variability of perceptions in the process of making diagnostic decisions.
Journal Article
Validating Parallel-Forms Tests for Assessing Anesthesia Resident Knowledge
by
Chen, Yi
,
Goodman, Stephanie R.
,
Chatterji, Madhabi
in
Anesthesia
,
Anesthesiology
,
Cesarean section
2024
We created a serious game to teach first year anesthesiology (CA-1) residents to perform general anesthesia for cesarean delivery. We aimed to investigate resident knowledge gains after playing the game and having received one of 2 modalities of debriefing. We report on the development and validation of scores from parallel test forms for criterion-referenced interpretations of resident knowledge. The test forms were intended for use as pre- and posttests for the experiment. Validation of instruments measuring the study's primary outcome was considered essential for adding rigor to the planned experiment, to be able to trust the study's results. Parallel, multiple-choice test forms development steps included: (1) assessment purpose and population specification; (2) content domain specification and writing/selection of items; (3) content validation by experts of paired items by topic and cognitive level; and (4) empirical validation of scores from the parallel test forms using Classical Test Theory (CTT) techniques. Field testing involved online administration of 52 shuffled items from both test forms to 24 CA-1's, 21 second-year anesthesiology (CA-2) residents, 2 fellows, 1 attending anesthesiologist, and 1 of unknown rank at 3 US institutions. Items from each form yielded near-normal score distributions, with similar medians, ranges, and standard deviations. Evaluations of CTT item difficulty (item p values) and discrimination (D) indices indicated that most items met assumptions of criterion-referenced test design, separating experienced from novice residents. Experienced residents performed better on overall domain scores than novices (P < .05). Kuder-Richardson Formula 20 (KR-20) reliability estimates of both test forms were above the acceptability cut of .70, and parallel forms reliability estimate was high at .86, indicating results were consistent with theoretical expectations. Total scores of parallel test forms demonstrated item-level validity, strong internal consistency and parallel forms reliability, suggesting sufficient robustness for knowledge outcomes assessments of CA-1 residents.
Journal Article