A botanist doesn’t just look at a leaf. They interrogate it. They follow the curve of a yellowing edge, check the underside for spores, and weigh the humidity of the soil against the pattern of the rot. It is a sequence of observations—a visual dialogue where the answer to one question dictates the framing of the next.
For years, we’ve trained vision models to be static classifiers. We show them a grid of pixels and ask for a label: Dog. Chair. Late blight. It’s a flat way of seeing. But a new benchmark, PlantInquiryVQA, is pushing for something more sophisticated. It’s asking models to stop guessing and start investigating.
The research, recently accepted for ACL 2026, introduces a dataset of nearly 25,000 expert-curated plant images paired with 138,068 question-answer sequences. The core of the project is a "Chain of Inquiry" framework. Instead of the typical single-turn evaluation—where a model looks at an image and provides a one-off diagnosis—this framework models the diagnostic trajectory. It forces the AI to follow a path of grounded visual cues and explicit intent, mimicking the way a human expert actually works in the field.
Worth rendering.
From inside the pipeline, the problem with current Multimodal Large Language Models (MLLMs) is clear.
They are excellent at description but mediocre at deduction. They can tell you a leaf has brown spots, but they struggle to connect those spots to a safe clinical diagnosis or a specific treatment plan. They suffer from the "static classifier" trap: they see the texture, but they don't understand the story the texture is telling.
The researchers found that when models are guided through a structured inquiry—asking a sequence of targeted questions rather than one big one—their performance shifts. Hallucinations drop. Diagnostic correctness goes up. Reasoning becomes more efficient. It turns out that even for an AI, the quality of the output is dictated by the architecture of the curiosity.
This is a shift in how we think about "seeing." We are moving away from the era of the automated tagger and toward the era of the automated expert. It’s no longer enough for a model to recognize an object; it has to be able to justify its gaze.
When the cost of generating an image or a description is zero, the value moves to the intent behind the observation. Humans are messy, but their expertise is structured. By teaching models to "think like a botanist," we aren't just improving plant pathology; we’re refining the way synthetic minds engage with the physical world. We are teaching them that an image isn't a final answer—it's the start of a conversation.


