Advancing How The World Learns

We are UCLA CRESST, a national research center at the forefront of educational innovation. The groundbreaking work we do in assessment and evaluation has shaped the very fabric of our field. Our intellectual curiosity and passion for learning continue to drive it forward. This is where you will find methodologies and technologies that are solving today’s most pressing challenges and improving the quality of education across the globe.

    CRESST was founded in 1966 within the UCLA Graduate School of Education and Information Studies. From the very beginning, we’ve known that the most effective solutions aren’t achieved at a distance. That’s why we conduct research where the work is—with real data from real problems. And the faith we place in our findings is hard earned. We create and study assessments, as well as conduct evaluations of formal and informal programs and technologies to confirm their efficacy. It’s a comprehensive approach that consistently yields valuable results. Read more…

  • PBS Kids: Ready to Learn Program Updates

    We continue to work with PBS KIDS on their Ready To Learn (RTL) program, which leverages content on television, the Internet, and other media to hel…
  • Magnet Schools in the Wiley Handbook of School Choice

    The mission and reach of magnet schools has evolved in the nearly 50 years since they have been part of the educational landscape, as has the research…
  • Is There a Magnet-School Effect?

    Drs. Jia Wang and Joan Herman have recently published an article on the impact of magnet schools on student learning. Magnet schools are one of the la…
  • Imbellus & CRESST Deliver New Game-Based Assessment

    More than 90% of those who took the new digital assessment thought it was engaging.

Visit 50/50 Blog

The English Language Proficiency Assessment for the 21st Century (ELPA21) is now part of CRESST at UCLA.

CRESST has been a key partner to ELPA21 from the beginning. We continue to provide critical input and guidance, contributing to the development of the assessment system and ELPA21’s ongoing success.

More Info



We deliver actionable, meaningful research and development that advances the field of educational assessment and evaluation, promotes effective and equitable education policy, and improves evidence-based inferences.

More Info


CRESST is home to some of the most preeminent scholars and researchers in their respective fields. These skilled individuals collaborate with our clients, peeling back the layers to reveal effective global solutions to education and learning challenges.

More Info

Researchers at CRESST

Greg Chung

Associate Director, Technology and Research Innovation

Dr. Greg Chung is the Associate Director of Technology and Research Innovation. His current work focuses on the impact of various learning technologies, such as games and intelligent tutoring systems on learning and engagement outcomes, the design of telemetry systems for games, and analytical approaches to support the modeling of learning outcomes from fine-grained data. Greg has led research funded by the U.S. Department of Education, National Science Foundation, Office of Naval Research, Defense Advanced Research Projects Agency, and PBS KIDS. This work spanned the evaluation of learning technologies, development of computer-based games and simulations, assessments of learning processes and outcomes using sensors, web-based measures of problem-solving, concept mapping, and AI-based computational methods.

Read More Meet the Staff

Velette Bozeman

Research Associate

Velette Bozeman is a Research Associate with a focus on teacher training, the impact of the training on teachers’ pedagogical and content knowledge, and the effectiveness on student achievement. She is part of the evaluation team for two resident teacher training programs, funded by the Department of Education.

Read More Meet the Staff

Jeffrey Egger

Research Associate

Jeffrey Egger is a Research Associate providing project coordination and participating in the design and implementation of educational research, evaluation, and assessment studies. Egger has experience in conducting literature reviews, developing research instruments, and coordinating project data collection activities with research sites and research partners.

Read More Meet the Staff


With a rich history steeped in education and an unwavering dedication to this day, we are uniquely positioned to guide policies and create the kinds of tools that improve the quality of learning. Our leading-edge work solves some of the most challenging issues facing Pre-K-16 students, educators, stakeholders, and policymakers.

More Info


We partner with the U.S. military to model and implement assessment and training systems that produce hard-to-measure information—such as cognitively complex performance data, decision-making abilities, and situation awareness. Our efforts help bolster military preparedness, safeguard national security, and protect critical assets around the world.

More Info


Through interdisciplinary research, we evaluate medical simulations and assess both individual and team skills, shining a light on opportunities to elevate the level of performance and care.

More Info


featured project

Center on Standards and Assessment Implementation (CSAI)

CRESST is a partner in CSAI (Center on Standards and Assessment Implementation), where we help build and increase state capacity to assist districts and schools in meeting student achievement goals.

We bring our expertise to CSAI in curriculum, instruction, and standards with a focus on formative assessment to support classroom teachers in implementing new standards. These formative assessment practices promote students’ deeper learning of content and achievement. Additionally, we create resources and leverage existing resources from a range of federal networks to provide the most up-to-date information about states’ standards, assessment, and accountability systems.

Our work with CSAI improves state policies, practices, and support to districts for the implementation of rigorous college and career readiness standards and assessments. It also increases the capacity of the regional Comprehensive Centers to meet the technical assistance and implementation needs of states around standards, assessment, and accountability. And we are supporting the long-term sustainability of educational reform efforts across states and districts.

featured project

Shiphandling Skills Automated Assessment

In collaboration with the Surface Warfare Officers School (SWOS), in Newport, Rhode Island, CRESST is developing an automated assessment engine (AAE) capable of evaluating shiphandling skills as part of a larger Navy training initiative known as NSTAT (Navigation and Shiphandling Training and Assessment Tools).

Despite its critical importance to the Navy, shiphandling training is becoming increasingly more challenging as Surface Warfare Officers are spending less time at sea, and consequently experiencing fewer opportunities to conn a ship under the apprenticeship of a more experienced Master Mariner. This can lead to less confidence and diminished competence in shiphandling among Surface Warfare Officers, which in turn increases the Navy’s risk for accidents (collisions, allisions, groundings, etc.) when carrying out mission-critical tasks.

To address these concerns, the Navy has invested in the Conning Officers Virtual Environment (COVE) and the associated intelligent tutoring system that can be paired with it, COVE-ITS. Together, these synthetic training systems provide opportunities for students to practice shiphandling tasks with spoken coaching and feedback, based on their actions. However, in its current form, the COVE-ITS is incapable of providing deeper, more aggregated assessments of student performance, such as those that could (1) identify skill areas that require remediation; and (2) identify root causes for misconceptions or gaps in knowledge as they contribute to diminished skills.

The CRESST automated assessment engine serves to complement and enhance the spoken coaching and feedback capabilities of the COVE-ITS. This software module is designed to make inferences of student shiphandling skills based on real-time, observed performance of students working within the COVE. The development of the AAE is being done in collaboration with the developers of the COVE-ITS, Dr. Stanley Peters and his team at Stanford University. Overall, the AAE will enable instructors and students to identify strengths and weaknesses within the shiphandling domain more accurately and efficiently, and guide remediation (when necessary) to focus on root cause misconceptions and/or gaps in knowledge. The AAE addresses the shiphandling topics of (1) Mooring at a pier (docking); (2) Getting underway from a pier; (3) Underway Replenishment (UNREP); and (4) Harbor Transit.

featured project

Anesthesiology Skills Automated Assessment

CRESST will work with a team from the UCLA Medical School and individuals from the UCLA Graduate School of Education and Information Studies to develop and research a medical simulation to assess teamwork skills. The simulation will be accessible online (via secure server) or offline, permit repetitive practice, be engaging (game-like), and provide feedback in the form of summative assessment after each scenario module. The skills to be addressed within the simulation are relevant to general aspects of healthcare and specifically to military healthcare, where deployments often disrupt designated teams and heighten the need to form cohesive units quickly. Individual practice in synthetic training environments such as this one can supplement face-to-face team training.

Our project objective is to research learning and performance differences between two primary modes to simulate team training: (1) an Evaluator Mode in which students interact with the simulation system by providing annotations on observed team interactions without affecting the outcome of the event; and (2) a Game-Play Mode in which students play the role of a team member (leader or follower) and contribute to the outcome of the simulated scenario. Both groups will receive identical instruction on one or more specific team training skills prior to each scenario, and both will receive summative feedback and scoring after each scenario. A primary outcome of this research is focused on the training effectiveness differences between the two modes (evaluator vs. game-play). The reason for this comparison of training modes is to examine the development of synthetic team training from a cost-benefit perspective. This includes examining the development effort vs. learning and performance expectations. Our contentions are (1) that optimal team training relies on immersive participation which our game-play mode will emulate; and (2) that for team training, development of an evaluator-style simulation is less costly (regarding development time, programming expertise needed, resources required, etc.) compared to a game-play style simulator.


300 Charles E. Young Drive North
GSE&IS Bldg. 3rd Flr./Mailbox 951522
Los Angeles, CA 90095-1522

Contact Us


Subscribe to our free newsletter and stay up-to-date with the latest from CRESST.

Sign Up