2  Assessment Operations

2.1 Test Design and Development

The test specifications document that describes our approach to assessment and test design for the ORExt is published in the OR Extended Assessment-Item Development Info. The document includes our approach to reducing the depth, breadth, and complexity (RDBC) of grade level content standards, an overview of the essentialization process and EAF documents, the planned test design for the ORExt, test development considerations, sample test items, item specifications, and universal tools/designated supports/accommodations. Only Grade 7 Math field test items were developed in 2017-18 which were in accordance with the 2014-15 test specifications, and are the most current available. A brief historical review of the ORExt is provided at the following link: Historical Review ORExt.

2.1A ORExt Purpose

The stated purpose of the ORExt is to provide the state technically adequate student performance data to ascertain proficiency on grade level state content standards for students with significant cognitive disabilities. A long-term goal of the program is to also provide information regarding annual student growth related to these content standards over Grades 3-8, as measured by vertically scaled assessments in ELA and Mathematics. The results of the assessment are currently reported in comparison to four performance levels: Level 1, Level 2, Level 3, and Level 4. Levels 3 and 4 denote a proficient level of performance, while Levels 1 and 2 denote performance that is not proficient. BRT and ODE developed a scaled score interpretation guide to assist stakeholders in interpreting the meaning of the scaled scores generated by the ORExt, supported by the state’s achievement level descriptors. This guidance is published in the Decision Making Related To Scaled Scores.

2.1B ORExt Test Blueprint

The ORExt Test Blueprint includes the entire test blueprint for the ORExt, as conveyed by the balance of representation across content areas and domains. Field-testing is conducted each year in order to support the continuous improvement of test functioning. However, items are selected to maintain this balance of representation. Oregon teachers validated the content of the assessment, agreeing with the standards that were and were not selected to develop the Essentialized Standards to which the ORExt test items are aligned.

2.1C Test Development Processes

The test development process implemented for the ORExt is conveyed in the ORExt Item Development Process. including standard selection and validation, item development, item review, review of all Oregon teacher feedback and updating of items, and scaling and item selection. The “ORExt Item Development Process” articulates the process used to generate the test administrator Scoring Protocol with comma separated value files used to create item templates that feed into Adobe InDesign© through a data merge. Student Materials are generated through the electronic test application, and then merged to pdf for a 1:1 correspondence for electronic and paper/pencil Student Materials. Final test packages are reviewed for accuracy and content and then disseminated to Oregon Qualified Assessors and Qualified Trainers through the secure password protected electronic test application platform, and paper/pencil materials are available for download on the secure password protected Training and Proficiency site.

2.1D Computer-Adaptive Considerations

The ORExt is not a computer-adaptive instrument, so these concerns do not apply.

2.2 Item Development

Item writers were recruited by ODE staff using an existing Qualified Assessor/Qualified Trainer listserv.

2.2A Project Description:

Behavioral Research and Teaching at the University of Oregon recruited Oregon teachers to participate in item development for a new alternate assessment prior to the 2014-2015 school year. Selected teachers were asked to develop 360 items in English Language Arts, Mathematics, or Science over the course of the summer, from mid-June through end of August. The Project Director worked with lead item developers to provide training, ongoing review and feedback, and quality assurance. All participants were expected to provide documentation of their qualifications and sign test security agreements. In addition, all item developers were expected to participate in a half-day item development training based upon the following schedule: ELA - Tuesday, from 8 AM to 12 PM; Math - Wednesday, from 8 AM to 12 PM; Science - Thursday, from 8 AM to 12 PM.

2.2B Minimum Qualifications:

All licensed Oregon public school teachers with at least three years of teaching in a life skills/severe needs program (SPED) or a general education classroom (GEN-ED), respectively, were encouraged to apply. Preference was given for item writing experience, additional years of teaching experience, and higher education degree status.

2.2C Compensation:

Teachers who participated in this process were compensated at a rate of $20/hr via professional service contracts. It was anticipated that teachers would produce 4 ELA items/hr, 6 Science items/hr, and 8 Math items/hr. As such, the maximum contract amount for ELA was $1,800, for Science $1,440, and for Math $900. Item development focused primarily on writing the stem and 3 options, with no need to produce graphics (rather use labels for a BRT graphic designer to produce).

2.2D Contact:

Because the timeline required work over the summer, Oregon teacher recruitment was challenging. BRT researchers thus performed an additional on-campus recruitment within the College of Education using the same information. The final pool of item writers included 18 item writers: seven Oregon teachers (all with MA degrees), five PhD candidates within the COE, and six BRT researchers (four PhD candidates, one PhD, and one with an MA). Item writers averaged 11.5 years of teaching experience. The teachers recruited all had prior experience developing items for the ORExt, as did all of the BRT researchers. The five PhD candidates within the COE had no prior item development experience. All item development was reviewed by BRT researchers and the Project Manager.

The item development process followed is elaborated in the Item Writer Training PowerPoint used in training all Oregon item writers. The item development process was structured with the following steps. Item writers were first oriented to the student population, as the pool of item writers included both content and special education experts. The Essentialization Process used to RDBC grade level standards was then modeled so writers would understand how the item alignment targets, the Essentialized Standards, were generated. Lecture, guided practice, and independent practice activities and follow-up discussion ensured comprehension of the process. BRT staff developed exemplar items for every Essentialized Standard, varying the complexity from Low (L) to Medium (M) to High (H) levels of complexity to convey the different performance expectations at each level. The balanced vertical scaling design provided an overall form-to-form and grade-to-grade level framework for the test formation process once items were developed (see ORExt Assement Vertical Scaling Project). Sample items are provided in the ORExt Electronic Practice Tests for stakeholder reference, demonstrating the format and style of typical items on the ORExt.

2.3 Test Administration

The ORExt assessments are administered according to the administration, scoring, analysis, and reporting criteria established in the Summative Testing Administration Manual. Important updates to the testing process are distributed via the Assessment and Accountability Updates listserve, as well. ODE uses this system to communicate information that is relevant for the statewide assessment system, including the ORExt. Announcements are sent to the listserv by email and are also posted to the ODE website. The standardization of test administration is supported by a comprehensive training process described below in Section 2.3B.

2.3A Administration and Accommodations

The state has ensured that appropriate universal tools, designated supports, and accommodations are available to students with disabilities and students covered by Section 504 by providing guidance and technical support on accommodations in the Oregon Accessibility Manual, and the Oregon Accommodations Manual. Guidelines regarding use of the accommodations for instructional purposes are included in the document, as all students are expected to receive test accommodations that are consistent with instructional accommodations.

Accommodations are built into the flexibility provided by the ORExt test though they have not yet been researched for the ORExt. However, annual training and proficiency testing efforts related to becoming a qualified assessor and/or qualified trainer for the ORExt support standardized use of available accommodations that are not already part of the test design. Based on annual analyses, results demonstrate that student performance varies according to their abilities and not construct-irrelevant factors, such as sex, race, or ethnicity (See Section 4.2).

The state has ensured that appropriate accommodations are available to students with limited English proficiency by providing guidance and technical support on accommodations in the Oregon Accessibility Manual. Communication systems for this student population are limited; exposure to multiple languages can make a student’s communication system more complex. The ORExt uses universal design principles and simplified language approaches in order to increase language access to test content for all students. In addition, directions and prompts may be translated/interpreted for students in their native language.

An analysis of accommodated versus non-accommodated administrations is needed in order to demonstrate that the provision of language accommodations is not providing any advantage to students with limited English proficiency, nor any disadvantage to other participants. Accommodations information was collected this year as an option for data entry. Entering accommodations information will be required each year. Analyses of the impact of accommodation provision on the ORExt is feasible after each years administration.

The Oregon Extended assessments can be administered using both Large Print and Braille (contracted and non-contracted) versions, as well. Oregon has ensured that the Oregon Extended assessments provide an appropriate variety of accommodations for students with disabilities. The state has provided guidance on accommodations in presentation, response, setting, and timing in the Accommodations Manual 2021-22: How to Select, Administer, and Evaluate Accommodations for Oregon’s Statewide Assessments in the Oregon Accommodations Manual. The Oregon Extended assessments are also designed according to universal design principles and utilize a simplified language approach (see Reducing the Depth, Breadth, and Complexity of Items).

In the 2021-2022 school year, the state redesigned a training and proficiency program for sign language interpretation of its assessments and has significantly updated the site during this time. The Sign Language Training process included videos of interpreters administering items to students, materials that support appropriate administration (i.e., transcripts, closed captioning and PowerPoint slides that supplement the video administrations and the current ODE accessibility manual), and proficiency testing to support standardized interpretation for Oregon’s assessments, including the ORExt. An 11-item proficiency test was administered, with an 80% required for passing (9/11 items correct). In 2021-22, the site was used to train 66 participants. All participants passed the assessment within 2 attempts. The overall average scores on the proficiency test were between 94%-96%.

The ORExt assessments provide an appropriate variety of linguistic accommodations for students with limited English proficiency. They also use a simplified language approach in test development in order to reduce language load of all items systematically (see Reducing the Depth, Breadth, and Complexity of Items). Any given student’s communication system may include home signs, school signs, English words, and Spanish words, for example. With the exception of items that require independent reading, the ORExt assessment can be translated or interpreted by a Qualified Assessor (QA) working with an interpreter in the student’s native language, including American Sign Language. QAs are allowed to translate/interpret the test directions. QAs can adapt the assessment to meet the needs of the student, while still maintaining standardization due to systematic prompts and well-defined answers.

2.3B Comprehensive Training System

Comprehensive information for ongoing training for all qualified assessors (QAs) and Qualified Trainers (QTs) is provided in the following QT Training Video. Through an online distribution and assessment system, QA/QT Training and Proficiency is determined annually. This website hosts all resources and information needed to administer, score, report, and interpret the results from the ORExt. The website also includes proficiency assessments that are required for all QAs and QTs who may administer the ORExt. QTs are directly trained by ODE and BRT staff as part of a train-the-trainers model. QTs then provide direct trainings for new QAs in their respective regions. The Oregon Department of Education (ODE) provided four direct statewide trainings for new Qualified Trainers (QTs) and returning QTs in Zoom regional trainings. The regional trainings were provided at two separate time slots on November 10th and November 17th for a total of four QT training opportunities (see QT Training Video link above). Only trained Qualified Assessors (QAs) can administer the Oregon Extended assessment. Qualified Assessors who also receive direct instruction from ODE and BRT may become Qualified Trainers (QTs) who are certified to train local staff using the train-the-trainers model. Training for new assessors must be completed on an annual basis. Assessors who do not maintain their respective certifications for any given year must re-train if they choose to enter the system again.

The tables below contain data from the Oregon Extended Assessment Training and Proficiency Website. All assessors need to complete training each year to retain their status for administering the Extended Assessments.

New assessors and returning assessors needed further training in 2021-22 and were required to pass proficiencies with a score of 80% or higher. These proficiencies covered areas in Administration, English Language Arts (ELA), Mathematics, and Science. Returning QAs or QTs for the 2021-22 school year only needed to pass a Refresher Proficiency, again with a score of 80% or higher. The tables below contain data on the number of assessors (participants) in both New Assessor and Refresher Proficiency tests. Included in the data is the number of attempts needed to attain a passing score as well as the average passing score of the participants.

The table below outlines registered users from the Oregon Extended Assessment Training and Proficiency Website.

Passing Scores (as percentage correct) by User Type, Assessment, and Number of Attempts.
Count Number of Attempt Mean (SD) Range of Passing Scores
Qualified Assessor
new assessor 230 1 92% (7%) [80%, 100%]
new assessor 28 2 89% (7%) [80%, 100%]
new assessor 5 3 94% (9%) [80%, 100%]
refresher 479 1 92% (7%) [80%, 100%]
refresher 53 2 87% (5%) [80%, 100%]
refresher 9 3 93% (7%) [85%, 100%]
refresher 2 4 90% (14%) [80%, 100%]
Qualified Trainer
new assessor 46 1 90% (6%) [80%, 100%]
new assessor 3 2 89% (7%) [83%, 97%]
refresher 48 1 94% (6%) [80%, 100%]
refresher 2 2 85% (7%) [80%, 90%]
Note:
The New Assessor Training Exam has 30 items; the Refresher has 20 items

A higher number of assessors completed the Refresher Proficiency test than the New Assessor Proficiency tests reflecting a greater number of returning assessors compared to new assessors. Data showed the majority of those who took either the Refresher or New Assessor Proficiency tests passed the first attempt. A limited number of participants had to take a second attempt, and very few to none had to take a third and fourth attempt.

Trainer Confidence Scale Percentages

All technical assistance questions that we received from the field as part of our HelpDesk are tagged and reviewed through HelpScout. The most common inquiries for the 2021-2022 test administration window involved status upgrades, missing students,adding additional schools and districts to user accounts, credential verification, and rostering. Some other common inquiries included student registration, access to monitoring for DTC’s, and technical issues with individual tablets.

Help Desk Email Report

A HelpScout FAQ website was also created to assist in answering questions from the field. Some of the frequently viewed pages included Logging into the ORExt Electronic app, Rostering, Supporting Electronic Test Administration, Curricular and Instructional Materials, Data Entry, and Credential Verification. All help desk inquiries will be taken into consideration for the 2022-2023 QA/QT training, and as the training site and FAQ site are updated prior to opening the 2022-2023 testing window.

A summary of visits to the FAQ page on ORExt HelpScout is below.

FAQ Conversations Created

Oregon monitors the quality of its system in several ways in order to support continuous improvement. In terms of the assessment quality, item statistics are reviewed each year and items that are not functioning as intended are removed and replaced by better functioning field-test items.

In 2014-15, items were reviewed in two phases, first using classical test theory (CTT) and second using Rasch analyses. All items flagged as a result of the statistical reviews were analyzed, item-by-item, by a team of measurement and content experts at BRT. Not all flagged items were removed, as several did not have apparent design flaws. Considerations regarding domain representation as well as item difficulty range also were considered during the review process. Different decision rules for unique items versus horizontally- or vertically-scaled anchor items were imployed. It was important in many cases to maintain anchor items. Items with clear design flaws were removed from subsequent analyses and reporting. The following flagging criteria were employed:

  • CTT: A unique item was flagged if it had a p-value of .10 or lower, .90 or higher, or a point biserial < .15. Anchor items were flagged if they had a p-value of .10 or lower or .95 and higher on all forms or a point biserial < .45 on any form.
  • Rasch: Unique items were flagged if their outfit mean square values were between 0 and .25 or > 1.5. Anchor items were flagged if their outfit mean square values were < .5, > 1.8 for horizontal items, or > 2.0 for vertical anchor items.

Out of a total of 5,929 items developed in 2014-15, 166 were removed (2.8%).

A consequential validity study was implemented in 2018-19 that surveyed QAs and QTs regarding the academic and social consequences of the ORExt, both intended and unintended. The Consequential Validity report is published in the Consequential Validity Survey Results. ODE and BRT staff reviewed the results of the survey to determine what program improvements were needed. A summary of the results is provided below.

ODE implemented a research survey program to address the need to document the consequences, both intended and unintended, of the ORExt Assessments. The research questions were framed based upon current consequential validity approaches for alternate assessments in the literature, as well as issues that were of specific value in Oregon. The survey included 121 respondents. This was 11% of the solicited respondents, who were all Qualified Assessors (QAs) and Qualified Trainers (QTs) in the or.k12test.com database. The sample was 83% female and represented all regions of the state, as well as age ranges. The survey included a range of quantitative and qualitative components. The quantitative results demonstrated that QAs and QTs continued to feel that the ORExt test items were easy to administer and score (64.2% Strongly Agree) and felt confident in their ability to interpret scaled scores and Achievement Level Descriptors for the ORExt (69.8% Strongly Agree and Agree). They also felt that the items were accessible for students who participated (78% Strongly Agree and Agree) and that the ORExt reflected the academic content that SWSCD should be learning (68.4% Strongly Agree and Agree). QAs and QTs felt marginally positive about the educational impacts of the ORExt and marginally negative about its social impacts. The results again demonstrated that the ORExt content area assessments generally required up to one hour to administer.

The qualitative results revealed two areas in which educators appreciated the ORExt and four areas of needed improvement. QAs and QTs said that they appreciated: 1) the assessment’s efficiency (i.e., more streamlined administration, ease of administration, easier to give and score online, online materials distribution); and, 2) overall item and test design (i.e., one item per page, visual supports, scoring protocol and student materials design, accessibility of test questions). Teachers recommended the following areas of improvement, not all of which are actionable: 1) Option to administer the assessment electronically was beneficial, 2) A functional skills assessment should be added, 3) New items for very low functioning students should be developed, and 4) request for a math assessment composed of more practical/life skills problems involving time and money. Complete results, including anticipated responses, from the survey can be found in the Consequential Validity Survey Results.

2.3C Technology-based Assessments

The ORExt was implemented using a technology-based platform as Phase 3 of the ORExt Tablet Administration. The 2017-18 testing window was the first year all grade level and subject area assessments were available on a tablet application/web-based platform (see ORExt Pilot Tablet Study Report). Administration of the tablet application mirrors paper/pencil administration with each item read aloud to the student, and the student asked to select one of three answer choices. Tablet functionality includes optional discontinuation if the student misses 10 out of the first 15 items, directing the assessor to administer the ORora. To support understanding of the system by both teachers and students, a separate practice test tablet application is available. Helpdesk inquiries and feedback from the field indicated much preference of the tablet administration versus paper/pencil. Qualified Trainers and Qualified Assessors reported their students’ were more focused during tablet administration, and because the tablet application scores automatically it was much more efficient for assessors. Improvements will be made to the electronic test based on technology improvements and feedback from the field. Data entry for all platforms is now maintained and monitored by secure BRT servers.

2.4 Monitoring Test Administration

ODE maintains a rigorous training system to support standardized test administration for the secure Oregon K12 website see screenshot below for an example of training content.

Training and Proficiency Home Page

The or.k12test.com website includes a training section that addresses any systems updates, the process for becoming a Qualified Assessor or Qualified Trainer, student eligibility expectations, student confidentiality and test security, test administration and scoring expectations, examples of appropriate and inappropriate administration (with video examples), supporting student access to items without violating the test construct, content area trainings that demonstrate how to administer items in ELA, Math, and Science (with video examples and supporting test materials), and how to access secure tests and complete data entry. Information for QAs, QTs, and parents regarding the ORExt is also provided, as are all necessary support materials. For QAs, these materials include practice tests to prepare both themselves and students for the annual assessment and all of the training materials used on the website. In addition to these materials, QTs have access to all training materials necessary to provide annual training to QAs in their purview (see screenshot below):

Materials Page

In addition, monitoring and unofficial reports related to test administration for the ORExt can be found in the Student Details tab of the training and proficiency website. Official reports are addressed via general ODE reporting systems. Information regarding this process can be located in the general assessment system Peer Review evidence submission.

2.5 Test Security

2.5A Prevention of Assessment Irregularities

Test security policies and consequences for violation are addressed in the Test Administration Manual on an annual basis, see Test Administration Manual. These policies include test material security, proper test preparation guidelines and administration procedures, consequences for confirmed violations of test security, and annual training requirements at the district and school levels for all individuals involved in test administration. Consequences for adult-initiated test irregularities may be severe, including placing teaching licenses in jeopardy (see Test Administration Manual).

2.5B Detection of Test Irregularities

The ODE utilizes a localized monitoring system where school test coordinators oversee building-level administration by trained, Qualified Assessors, and report to centralized district test coordinators, who are then responsible for reporting any confirmed violations to ODE. Improprieties are defined as adult-initiated or student-initiated and investigated accordingly (see Test Administration Manual).

2.5C Remediation Following Test Security Incidents

ODE’s alternate assessment program manager investigates and remediates substantiated test security incidents for the ORExt by working with district test coordinators. Additional information regarding this process can be located in the general assessment system Peer Review evidence submission.

2.5D Investigation of Test Irregularities

School and district test coordinators conduct initial investigations into all alleged test irregularities. Once reported to ODE, all alleged test irregularities are investigated in consultation with district test coordinators and the test vendor, as appropriate (see Test Administration Manual). In the event that a test irregularity is determined to be factual, consequences are determined based upon contextual issues that are brought to light during the investigation. Additional information regarding this process can be located in the general assessment system Peer Review evidence submission.

2.6 Systems for Protecting Data Integrity and Privacy

2.6A Integrity of Test Materials

Test materials for the ORExt are maintained throughout development, dissemination, and administration via multiple mechanisms. All items under development are stored in secure file servers managed by Behavioral Research & Teaching at the University of Oregon, the test vendor for the ORExt. Item reviews necessary to provide alignment, bias, and sensitivity information are conducted online using the secure Distributed Item Review (DIR) platform (secure website, but see DIR Overview for a system overview).

For the 2021-2022 school year, all paper/pencil secure test distribution and data entry was hosted by BRT through the secure training site.

The secure tablet application and web-based platform distribution and data entry were hosted by BRT servers. All technology based secure administration and data entry was password-protected. Download of the tablet app was dependent on the type of device, all instructions and download links are available in the Oregon K12 website secure platform. Additional information regarding test security can be located in the general assessment system Peer Review evidence submission.

Data Entry Page

2.6B Secure Student-Level Assessment Data

Student level data is protected by relevant training and through a secure data system in which all data entry is conducted online using password-protected, secure procedures on the Oregon K12 website. Only trained users with a vested educational interest who have signed test security agreements are authorized to access to online data entry systems.

2.6C Protecting Personally Identifiable Information

All confidential, personally identifiable student information is protected by policy and supported by training (see Test Administration Manual). The minimum number of students necessary to allow reporting of students and student subgroups varies by rating (i.e., achievement, growth, graduation, and school size), by level (i.e., school/district/state), and by number of years of assessment data available. For example, to receive an achievement rating, schools must have at least 40 tests for the two most recent school years in reading or mathematics. Alternatively, small schools receive an achievement rating if they have at least 40 tests over the most recent four years. If a school does not have at least 40 tests over a four-year period, they will not receive an achievement score (see State Annual Report Card). Similar rules are applied to student subgroups, including students with disabilities, English learners, and students from diverse racial/ethnic backgrounds (see State Annual Report Card).