IEEE SPS @ UF

Plato’s Cave

Tooling that helps reviewers read faster by turning papers into structured ‘claims and evidence’ graphs with audit-friendly outputs.

Plato’s Cave project logo.

Overview

System that extracts structured nodes (claims, evidence, limitations, etc.) from papers and scores them to support literature review and comparison across a collection.

Reading dozens of papers is slow partly because the structure is inconsistent. Plato’s Cave uses modern language models to extract a consistent structure (e.g., claims, evidence, limitations) and then runs a scoring pipeline so papers can be compared more systematically.

What I built

  • A batch pipeline that processes PDFs, extracts structured ‘nodes’ (e.g., claims/evidence), and stores results in machine-readable formats.
  • Scoring and normalization routines so outputs are comparable across papers.
  • Run outputs designed for auditing (logs, summaries, and artifacts).

How it works

  • Convert a paper to text.
  • Use a language model to label and structure key statements.
  • Score the resulting graph for quality and consistency.
  • Export summaries so humans can review quickly.

Snapshot

Track
Tools · language + structure · reproducibility
Status
Active; evolving with new evaluation experiments.
Focus
Large language models (LLMs)Research toolingReproducibility

Stack

  • Python
  • Large language model APIs
  • Experiment logging

Glossary

LLM
Large language model: a model trained to generate and analyze text.
NLP
Natural language processing: methods for working with text.