Explore the critical challenge of uncertainty quantification in large language models. Learn about confidence estimation techniques, calibration metrics like ECE and MCE, and practical methods to improve model reliability from logit-based approaches to ensemble methods and post-hoc calibration.
Tracking benchmarks and datasets for table understanding, visual table QA, and LLM/VLM-era table evaluation. Scope: downstream reasoning and multimodal understanding over tables, not structural recognition.
A comprehensive survey of multimodal large language models from 2021 to 2024, covering encoder-only models, encoder-decoder architectures, decoder-only models, and specialized applications for documents and screens.