Content
summary Summary

Researchers from UC Berkeley, Saudi Arabia's King Abdullah City for Science and Technology, and the University of Washington took a close look at how LLMs generate questions. Their findings show some clear differences between AI and human questioning patterns.

Ad

The research team started by developing categories for different types of questions, ranging from basic fact-checking to complex queries requiring detailed explanations. They then used these categories to analyze both AI-generated questions and existing datasets of human questions.

To test their theories, the team looked at how LLMs answered questions both with and without supporting context. By experimenting with different answer lengths, they could measure how much information each question really needed, giving them insight into the complexity of different question types.

To compare AI-generated questions, the researchers used two different datasets based on Wikipedia articles, each created using a different method. In one dataset, questions were created based on specific text passages, while in the other, researchers matched existing questions to relevant Wikipedia sections.

Ad
Ad

LLMs questions cover context more evenly

The team discovered that AI models heavily favor questions that need detailed explanations - about 44% of AI-generated questions fall into this category. Humans, on the other hand, tend to ask more straightforward, fact-based questions.

These AI-generated questions typically require longer answers to be complete, even when keeping responses as concise as possible. The difference in required answer length was significant compared to human-created questions.

Infographic: Two-column layout for analyzing LLM questions with context, generated questions and characteristic evaluation including evaluation metrics.
LLMs like to ask questions that require detailed answers. | Image: Zhang et al.

While humans often focus their questions on information that appears early in a text, AI models spread their questions more evenly across the entire content. This is particularly interesting because LLMs typically shows positional bias when answering questions.

The researchers believe their findings have practical applications. Since AI questions have such unique patterns, they could help test RAG systems or identify when AI systems are making things up. The insights could also help users write better prompts, whether they want AI to generate more human-like questions or questions with specific characteristics they're looking for.

AI-generated questions are becoming more common in commercial products. For example, Amazon's shopping assistant Rufus suggests product-related questions, while Perplexity's search engine and X's Grok chatbot use follow-up questions to help users dig deeper into topics. These tools let users either select from AI-generated questions to learn more about specific posts or topics.

Recommendation

For anyone interested in exploring further, the research team has shared their code publicly on GitHub alongside their published paper.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Researchers from the University of California Berkeley, the King Abdullah City for Science and Technology (KACST) and the University of Washington have investigated the properties of questions generated by large language models.
  • The questions generated by language models show a strong preference for descriptive questions that require explanations or characterizations, while human questioners tend to focus on factual questions. In addition, AI questions cover context more evenly than human questions.
  • These findings can help optimize prompt engineering for question generation with language models and generate questions that are more similar to human questions or have certain desired characteristics.
Sources
Jonathan works as a freelance tech journalist for THE DECODER, focusing on AI tools and how GenAI can be used in everyday work.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.