This chapter corresponds to code in the researcherRAG repository
View Stage 6 Prompt

Research Conversation & Analysis

Now that your RAG system is built, learn how to conduct systematic literature analysis through effective conversational AI interactions. This chapter covers query strategies, iterative research workflows, citation management, and best practices for extracting insights from your knowledge base.

๐Ÿ“‹ Prerequisites

  • โœ“ Completed Stages 1-5 (Vector DB built and validated)
  • โœ“ One of the chat interfaces installed (Claude Code, Streamlit, or FastAPI)
  • โœ“ Familiarity with your research domain and questions

โš ๏ธ IMPORTANT: RAG vs General Knowledge

Before starting your research conversations, it's critical to understand the difference between asking Claude directly versus using your RAG system interface.

โŒ WITHOUT Interface

Direct Claude Code chat uses general knowledge:

You: "Which methodologies are most commonly used?"

Claude: "Based on my training data, common methodologies include surveys, experiments..."

โ† NOT from your database!

โœ… WITH ScholaRAG Interface

Interface searches YOUR Vector Database:

System: Loaded 150 papers from Vector DB

You: "Which methodologies are most commonly used?"

System: ๐Ÿ“š Found 5 relevant papers:

  • Smith et al. (2023)
  • Jones & Lee (2022)

Claude: "Based on 5 papers in YOUR database: Qualitative methods: 3 papers [Smith, 2023]..."

Key Differences

AspectDirect Claude ChatScholaRAG Interface
Data SourceGeneral knowledge (training data)YOUR Vector Database
TransparencyNo visibility into sourcesShows which papers retrieved
CitationsNo paper citationsEvery claim linked to papers
VerificationCannot verify sourcesTrace back to originals
LimitationsDoesn't know what it doesn't haveSays "not in your database"

๐Ÿ” How to Run the RAG Interface

Step 1: Open your terminal and navigate to your project:

cd /path/to/your/researcherRAG-project

Step 2: Run the interface script:

python interfaces/claude_code_interface.py

Step 3: You'll see this prompt:

๐Ÿ“‚ Loading Vector DB from ./chroma_db...
โœ… Loaded 137 papers from collection 'papers'
โœ… Connected to Claude API

Type your questions (or 'exit' to quit)

You: 

Step 4: Ask questions!

You: What are the main adoption barriers?
You: Show me papers from 2023
You: exit  # when done

โœ“ How to verify it's using YOUR papers:

  • โ€ข System shows "๐Ÿ“š Found X relevant papers"
  • โ€ข Answers include [Author, Year] citations
  • โ€ข Can query specific papers you know are in your DB

Effective Query Strategies

๐Ÿ’ก New: Ready-to-Use Prompt Library

We've created 7 specialized prompt templates for common research scenarios. Copy, customize, and use them directly in your RAG interface.

โ†’ Browse Prompt Library

Different types of queries serve different research purposes. Use this framework to formulate effective questions for your literature analysis:

1. Exploratory Queries - Understand the Landscape

Start broad to get an overview of your corpus:

"What are the main research themes in my corpus?"
"Which methodologies are most commonly used?"
"Who are the key authors and their contributions?"
"What time periods are covered?"
"Show me the most cited papers"

When to use: Beginning of analysis, understanding domain coverage

2. Specific Information Queries - Deep Dive

Ask focused questions about specific topics:

"What factors influence technology adoption in healthcare?"
"What are the reported adoption rates in developing countries?"
"Which theoretical frameworks are most cited?"
"What limitations are mentioned in recent studies?"
"How is 'organizational readiness' defined?"}

When to use: Targeted investigation, extracting specific data points

3. Comparative Queries - Find Patterns

Compare across different dimensions:

"How do adoption rates differ between developed and developing countries?"
"Compare quantitative vs qualitative studies on this topic"
"What changed in the literature before and after 2020?"
"How do TAM and UTAUT frameworks compare?"
"Compare findings from healthcare vs education sectors"

When to use: Identifying trends, contrasting approaches, temporal analysis

4. Gap-Finding Queries - Identify Future Research

Discover under-researched areas:

"What populations or contexts are under-represented?"
"Which methods have NOT been used to study this?"
"What gaps or limitations do authors identify?"
"Are there contradictory findings that need resolution?"
"What future research directions are suggested?"

When to use: Writing discussion/future work sections, identifying research opportunities

Iterative Research Process

Effective literature analysis is iterative, not linear. Use this session-based approach to systematically explore your knowledge base:

Session 1: Initial Exploration (30-60 min)

Goal: Get familiar with your knowledge base, understand scope and coverage

Example Session Flow:

Q1: "How many papers are in my knowledge base?"

Q2: "What are the 5 most common research topics?"

Q3: "Show me the most cited papers"

Q4: "What time range do these papers cover?"

Q5: "Which methodologies appear most frequently?"

Session 2: Deep Dive by Theme (1-2 hrs)

Goal: Investigate specific themes or topics in depth

Example: Deep Dive on "Adoption Barriers"

Q1: "What barriers to technology adoption are mentioned?"

Q2: "For each barrier, which papers discuss it?"

Q3: "How frequently is cost mentioned as a barrier?"

Q4: "Do barriers differ by geographic region?"

Q5: "What solutions or mitigation strategies are proposed?"

๐Ÿ’ก Pro Tip: Follow-up Questions

Don't stop at the first answer. Ask follow-ups like "Which paper provides the most detail on this?" or "Are there contradictory findings?" to dig deeper.

Session 3: Cross-Paper Synthesis (1-2 hrs)

Goal: Synthesize findings across multiple papers, identify patterns

Example: Synthesizing Adoption Factors

Q1: "Create a table of all adoption factors mentioned, with paper citations"

Q2: "Which factors are mentioned in 3+ papers?"

Q3: "Compare how developed vs developing countries differ"

Q4: "Organize findings by theoretical framework (TAM, UTAUT, etc.)"

Q5: "Summarize consensus vs contradictions in the literature"

Session 4: Gap Analysis (1 hr)

Goal: Identify gaps, limitations, and future research directions

Example: Finding Research Gaps

Q1: "What limitations do authors mention in their studies?"

Q2: "Which populations or contexts are under-studied?"

Q3: "What methodological approaches are missing?"

Q4: "What future research directions are suggested?"

Q5: "Where are the contradictions that need resolution?"

Citation Management & Verification

Maintaining accurate citations is critical for academic integrity. Your RAG system should provide citations for every claim, but you must verify them.

Citation Validation Workflow

Citation Quality Checklist

โœ“

Every Claim Has a Citation

No unsourced statements in your notes

โœ“

Citations Use Correct Format

[Author, Year] or [Author et al., Year] for 3+ authors

โœ“

Paper Actually in Your Database

No hallucinated citations

โœ“

Claim Matches Source Material

Spot-check by reading original paper section

โœ“

DOI or URL Accessible

Test links to ensure readers can access papers

Managing Research Sessions

Keep your research organized by documenting sessions, tracking questions, and exporting findings.

Session Documentation Template

# Research Session Log

## Session 1: Initial Exploration
**Date:** 2024-01-15
**Duration:** 45 minutes
**Goal:** Understand corpus coverage and identify main themes

### Questions Asked:
1. Q: "How many papers in my database?"
   A: 137 papers (2010-2024)

2. Q: "What are the top 5 research themes?"
   A: 1) Technology adoption (45 papers)
      2) Implementation barriers (38 papers)
      3) User acceptance (32 papers)
      ...

### Key Findings:
- Adoption barriers most discussed topic
- Limited research on developing countries (only 18 papers)
- Qualitative methods dominate (67 papers vs 45 quantitative)

### Follow-up Questions for Next Session:
- Deep dive on adoption barriers
- Compare developed vs developing country findings
- Investigate methodological gaps

### Papers to Read in Full:
- [Smith, 2023] - Comprehensive systematic review
- [Johnson, 2022] - Novel framework for barriers

Export & Organization Tools

๐Ÿ“Š Export Conversation Logs

# Save conversation history
python interfaces/export_logs.py \
  --session session1 \
  --output research_notes.md

๐Ÿ“‘ Generate Citation List

# Extract all cited papers
python interfaces/export_citations.py \
  --format bibtex \
  --output references.bib

๐Ÿ“ˆ Create Summary Statistics

# Generate corpus statistics
python interfaces/generate_stats.py \
  --output corpus_stats.html

๐Ÿ” Find Specific Citations

# Query specific papers
python interfaces/claude_code_interface.py \
  --query "Show papers by Smith"

Research Conversation Best Practices

โœ… Do This

  • โœ“ Always use the interface scripts, not direct chat
  • โœ“ Start broad, then narrow down
  • โœ“ Ask follow-up questions to dig deeper
  • โœ“ Document sessions as you go
  • โœ“ Verify citations by spot-checking sources
  • โœ“ Export findings regularly
  • โœ“ Organize by research themes

โŒ Avoid This

  • โœ— Using direct Claude chat (gets general knowledge)
  • โœ— Accepting answers without verifying citations
  • โœ— Asking only one question per topic
  • โœ— Forgetting to document your sessions
  • โœ— Not following up on interesting findings
  • โœ— Ignoring contradictory results
  • โœ— Skipping spot-checks of source material

Next Steps

Once you've completed your research conversations and gathered insights, you're ready to write up your findings. Continue to the next chapter to learn how to structure your literature review, generate PRISMA diagrams, and create publication-ready documentation.

๐ŸŽฏ Ready to Start Researching?

Fire up your interface script (python interfaces/claude_code_interface.py) and start with exploratory queries. Remember: transparency and verification are key to trustworthy research.