Benchmark Dataloader
A small benchmarking harness for multimodal dataloaders, built to surface throughput bottlenecks before they become expensive training-time surprises.
Member of Technical Staff at Datology
At Datology, I work on evaluation, curation, and training systems for multimodal models. I want benchmark results, pipeline choices, and training outcomes to agree.
Good research infrastructure shortens the distance between a data choice and a trustworthy comparison.

A benchmark designed to make VLM evaluation more discriminative and decision-useful.
Read paper
Research engineer focused on multimodal curation, evaluation, and training systems.
Three lanes: evaluation, curation, and training systems. The point is to make the next experiment clearer and harder to fool.
Designing benchmarks and scoring paths that are selective enough to inform curation and model choices, not just report a leaderboard.
Working on embeddings, filtering, and dataset export paths for large image-text corpora, with attention to alignment and redundancy.
Building the dataloading, evaluation, and multi-node launch infrastructure that makes multimodal iteration fast enough to matter.
A smaller public tool, but the same bias toward practical research infrastructure.
A small benchmarking harness for multimodal dataloaders, built to surface throughput bottlenecks before they become expensive training-time surprises.