Comparing mapping methods

 

Statistical

Non-statistical

 

Rosetta Stone

AMPL module

Policy linking

Pairwise comparison

Data collection Ex-ante Ex-ante Ex-post Ex-post
Students Same Different Different Different
What A concordance table which translates the scores resulting from national or regional assessments to scores on international assessments. An AMPL module calibrated to the MPL is inserted either as an additional booklet or by running parallel assessments. Matches up definitions of the MPL descriptor using subjective judgment and under certain conditions, allows those assessments to be aligned across countries.
Items/test Different assessments Same AMPL module across different assessment programmes Different assessments
Calibration Calibration needs various steps Accurate to report on the MPL Depends on assessment programme
Alignment with Global MPL Yes, but needs standard setting to define accurate alignment. Yes Depends on alignment and sufficiency of items
Sufficient # of items Yes Yes Depends on each assessment tool
Measurement skills continuums Yes Not now but possible with current developments Depends on each assessment tool
Track progress over time Yes Yes Not clear; depends on quality of tools
Frequency Cycle depending on each assessment On demand Once per assessment n/a
Output Concordance table Calibrated to the MPL Identifies the MPL cut-off points Identifies the MPL cut-off points
How Relies on the participation of countries in two assessments.

Students take the two assessments to help link between the results of both assessments.

Insert the booklet either as a standalone running parallel test or as a rotating booklet. Group of experts provide judgment about each item on the test and set initial cut scores based on their understanding of the proficiency levels and the student population.
Country ownership Very low High High Medium
Needs Tests have enough items that could identify linking. A tool built with items that are aligned and sufficient to measure the MPL. Good-quality cognitive tool and procedures.

Strong alignment of assessment tools to GPF.

Pros Technically rigorous Technically rigorous Cost-effectiveness
Cons Costly. Efficient if done between a regional and a global assessment. Does not allow deep investigation of the construct. Relatively subjective (less for pairwise). Depends on the quality of the assessment tool and implementation of the linking process.
Achieved so far Rosetta Stone: ERCE (LAC) and PASEC (SSA) participated with IDEA in the Rosetta Stone exercise. AMPL-b administered

AMPL-c under development (PISA)

AMPL-a under development

First phase of pilots around 16 countries completed. Standard setting exercise for MILO (ACER, 2022).
Next/remaining steps Potential expansion to other regions and national assessments. Scale-up depends on country’s interest and development partners’ support. Revision of toolkit. Methodology guidance and analysis.
National cost Between US$250,000 and $400,000. Printing cost of a booklet. Extra administration costs depends on modality.  Between US$30,000 to $50,000 for national workshop. None
International cost International US$1 million per region. Regional – US$500,000 Averages US$100,000 for technical assistance  Between US$50,000 and $75,000 per country US$40,000

 

Section  Resource Link
1 SDGs
SDG4
Minimum Proficiency Levels used to report for indicator 4.1.1
Global Content Framework
GPF for Reading
GPF for Mathematics
3 SDG4 Global Tables 2021
4 Zambia Conference Documents
PIRLS
TIMSS
PASEC
PILNA
SEA-PLM
SACMEQ
PISA
Rosetta Stone Study
MILO: Monitoring Impacts on Learning Outcomes
Policy linking