AIResearch AIResearch
Back to articles
Data

AI Models Now See Images Through Text Queries

A new image encoder uses text questions to focus on relevant visual details, boosting AI performance by up to 6 points while cutting processing needs in half.

AI Research
March 26, 2026
3 min read
AI Models Now See Images Through Text Queries

A new approach to artificial intelligence allows vision-language models to process images more intelligently by using text queries as a guide. This , called the Text-Guided Semantic Image Encoder (TIE), addresses a fundamental limitation in current AI systems where image encoders treat all parts of an image equally, regardless of what the user is asking. For example, if a question is about a man leaning on a fence in a photo, existing models might waste computational resources analyzing irrelevant areas like bushes or the sky. TIE changes this by conditioning the image encoding on the specific query, leading to more efficient and accurate multimodal understanding.

The researchers found that integrating TIE into vision-language models (VLMs) consistently improves performance across a range of tasks. In experiments, VLMs equipped with TIE, referred to as PLM-TIE, outperformed conventional counterparts by an average of +1.5 points at the 1B scale and +1.3 points at the 3B scale across nine image-to-text benchmarks. Gains were particularly notable on tasks requiring detailed visual analysis, such as DocVQA and InfoVQA, where improvements reached up to 6 points. This demonstrates that text-conditioned encoding helps models better capture query-relevant visual features, as shown in Figure 1 of the paper, which contrasts the new architecture with prior approaches.

Ology behind TIE involves augmenting a standard Vision Transformer (ViT) to attend to text queries during image encoding. The system uses a pretrained text encoder, such as T5-Large, to convert the query into embeddings, which are then integrated into the image encoder's attention mechanism. This allows visual patches to attend to the query tokens, producing semantically enriched image representations aligned with the task. As detailed in Section 4 of the paper, the training process involves language alignment with a cross-entropy loss, keeping the language model and text encoder frozen while fine-tuning only the image encoder and projection layers. This setup ensures that the model learns to generate query-specific visual tokens without extensive retraining.

Analysis of reveals that TIE not only boosts accuracy but also enhances efficiency. According to Table 3 in the paper, PLM-TIE models achieve superior performance while using only half as many image tiles (tokens) compared to baselines, resulting in notably faster inference times. For instance, PLM-TIE with 4 tiles outperformed the baseline with 8 tiles, highlighting 's computational advantages. Qualitative analyses, such as those in Figures 5 and 6, confirm that TIE consistently attends to query-relevant regions, improving interpretability. For example, when asked about a specific amount on a document, TIE focused on numerical areas, whereas a generic query shifted attention to textual elements.

Of this research are significant for real-world applications, as it enables more efficient and accurate AI systems for tasks like visual question answering, document analysis, and image captioning. By reducing the need for excessive image tiles, TIE can lower computational costs, making it feasible for deployment in resource-constrained environments. The paper also notes that TIE generalizes well with generic queries, indicating robustness in multi-turn scenarios where the same image is referenced multiple times. However, the authors acknowledge limitations, such as not evaluating larger models beyond 3B or exploring tile configurations beyond 8 due to computational constraints, as mentioned in Section B. Future work could explore scaling TIE to larger architectures and integrating deeper text representations for even tighter vision-language alignment.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn