AI Education Thesis Ideas

Multimodal automated assessment (text + diagrams + sensors) emerging

Multimodal automated assessment (text + diagrams + sensors) emerging

Key Questions

What is multimodal automated assessment?

Multimodal automated assessment involves evaluating student work using text, diagrams, and sensor data through ML systems. Frameworks like those in intelligent educational systems use autoencoders, RBMs, and graph recommendations for grading multimodal inputs. LLMs struggle with pedagogical signals, prompting a push for specialized benchmarks and datasets like MATH-Vision.

What is the MATH-Vision dataset?

MATH-Vision is a dataset for measuring multimodal mathematical reasoning in large multimodal models, focusing on geometry and proofs. It supports visual math reasoning tasks and benchmarks for LMMs. It addresses gaps in current LLM capabilities for handling diagrams in math assessments.

What are the main challenges in multimodal assessment?

Challenges include artifact embeddings, alignment between modalities, noise in sensor data, and developing robust metrics. LLMs fail on pedagogical signals, requiring fine-tuning and sensor-fusion benchmarks. Emerging solutions emphasize task-specific ML and calibration for accurate grading.

2026 preprint, arXiv benchmarks, task-specific ML sensor system for placement/task assessment, MATH-Vision dataset for LMM visual math reasoning (geometry/proofs), and intelligent ed systems w/ multimodal behavior modeling (autoencoders/RBMs/graph recs; ex-1Ae06gAk) show frameworks grading multimodal student work; LLMs fail ped signals. Push for ed leaderboards/datasets. Challenges: artifact embeddings, alignment, noise, metrics. Thesis: fine-tuning, sensor-fusion benchmarks (MATH-Vision), calibration.

Sources (2)
Updated Apr 21, 2026