Paper Compass Score. Evaluate a paper's research value with a strict 10-point rubric, evidence-backed reasoning, and a mandatory 5-paper arXiv comparison set. Use when user wants to score a paper, decide whether it is worth reading, or judge whether it has field-shaping potential.
Keep direct quotes short: at most 25 English words.
If evidence is missing, explicitly write 信息不足 or .
Verwandte Skills
insufficient information
C1: Fixed Scoring Rubric
Total score must be 10.0.
Final score must use 0.1 granularity.
Read references/scoring-rubric.md and use that rubric exactly.
Do not invent new dimensions.
Do not change weights dynamically.
C1.5: Arithmetic Consistency Is Mandatory
The displayed final score MUST equal the arithmetic sum of the 7 dimension scores.
In the report, the Section 1 final score is DEFINED by Section 2: it must be computed from the 7 itemized scores in ## 2. Score Breakdown / ## 2. 分项评分.
There is no hidden bonus, penalty, or manual override after summation.
Derive the rating band from the final arithmetic score, not the other way around.
Before writing the report, recompute the sum once and verify:
If the sum and displayed final score differ even by 0.1, fix the dimension scores or the final score before writing.
Never write Section 1 first and then backfill Section 2. Score Section 2 first, sum it, then write Section 1.
C2: Mandatory 5-Paper Comparison Set
You MUST build a comparison set of exactly 5 arXiv papers found online.
The set must contain:
3 recent papers from the last 2 years
2 classic papers in the same field
Read references/comparison-selection.md before selecting them.
Every selected peer paper must include:
title
arXiv link
year
venue if available
citation count if available
one-sentence reason for inclusion
If fewer than 5 qualified arXiv papers can be verified, mark comparison_status=incomplete and set confidence to low.
In the incomplete-comparison case, score uncertain dimensions conservatively so that the arithmetic final score does not exceed 7.4/10.0.
Do NOT apply a hidden post-hoc cap after summing, because that would break score consistency.
C3: Conservative Forward-Looking Judgment
High score does NOT automatically mean field-defining.
Only use 有潜力奠基新方向 / potentially field-shaping when evidence supports it.
Strong field-shaping language requires at least 2 verified signals:
follow-up work from multiple groups
reused terminology / paradigm / benchmark
strong citing-paper quality
real industry or open-source adoption
If the paper is too new, write early promise, not yet established.
C4: Metadata Verification Is Required
For all papers, verify publication signal and citation signal with APIs.
Never call a paper unpublished, preprint only, or classic without verification.
For arXiv papers, use arXiv API first, then Semantic Scholar, then OpenAlex as needed.
For non-arXiv papers or local PDFs, extract title first, then search APIs by title.
For published-paper discovery, IEEE/ACM/Springer venue lookup, or ambiguous title search, you MAY use /semantic-scholar because it has better built-in filtering for venue papers and citation metadata.
C4.5: Citation Resolution Across Versions
Citation retrieval must consider both:
the arXiv preprint record
the later published journal/conference record
You MUST try to resolve whether these are the same scholarly work before using citation counts.
Never blindly add arXiv citations and published-version citations together, because that may double-count the same citing papers.
Prefer citation counts from a canonical merged record when the provider already links arXiv and DOI versions.
If providers disagree, write the split explicitly and explain which count is used for scoring.
Google Scholar may be used only as a manual cross-check when other sources disagree materially; do not rely on it as the primary automated source.
C5: Output Structure Is Fixed
Read the selected template before writing.
Output must follow the template exactly.
Section 7. **Sources**: must always be present.
The report must present itemized scores before the final verdict.
In other words: write the score table first, then write the overall conclusion derived from it.
Input Normalization
User Input
Rule
2010.11929 or arxiv:2010.11929
Convert to arXiv ID, then fetch metadata and content
https://arxiv.org/abs/...
Extract arXiv ID
https://arxiv.org/pdf/...
Extract arXiv ID
https://arxiv.org/html/...
Extract arXiv ID
Local PDF path
Parse title/abstract from file, then search online metadata by title
Other paper URL
Fetch readable content, then search metadata by title/DOI
Required Data Sources
Use Bash + curl or small scripts for metadata calls. Prefer APIs over generic web search.
Required sources:
arXiv API
title
authors
abstract
year
arXiv categories
Semantic Scholar API
venue
citation count
influential citation count
TLDR
author list
external IDs when available
OpenAlex API or another verified scholarly API
cited-by trend or backup citation metadata
citing paper sample when needed
DOI/arXiv version linkage or location metadata when available
Full paper text
downloaded PDF or readable HTML
Do not rely on a single unstable source for all metrics.
Recommended source roles:
arXiv API: authoritative for arXiv identity and abstract metadata
arXiv PDF/HTML: authoritative for paper content and quote extraction
/semantic-scholar or Semantic Scholar API: best for venue discovery, citation counts, DOI, and published-paper retrieval
OpenAlex API: backup cross-check for citations and related-paper metadata
Google Scholar: manual-only tie-breaker for citation visibility across versions; not a primary automated source
Workflow
Step 1: Fetch Target Paper Metadata and Content
Extract and verify:
title
authors
year
venue and publication signal
citation count
influential citation count if available
abstract or TLDR
full text sections when accessible
version linkage between arXiv and published record when available
For arXiv papers:
Use arXiv API for title, authors, abstract, year
Download PDF with retry, timeout, redirect-following, and a browser-like user agent
Verify the local PDF is non-empty before treating the download as successful
If PDF fails, try arXiv HTML as the content fallback
Use Semantic Scholar for venue, DOI, external IDs, and citation metadata
Use OpenAlex as a backup or cross-check for citation data and version linkage
Keep all downloads inside the current workspace under ./papers/, never /tmp
Citation-resolution procedure for arXiv papers:
Query Semantic Scholar by ARXIV:{id}.
If a DOI or published venue is returned, also query the published record by DOI or exact title.
Query OpenAlex by DOI when available; otherwise query by title and arXiv ID.
Decide whether the arXiv and published entries refer to the same work:
same DOI
explicit arXiv external ID on the published record
matching title/authors/year with clear venue linkage
If they are the same work and one provider already merges them, use that canonical citation count.
If they remain separate and unresolved, report both counts and use a conservative scoring basis:
prefer the published-version count when the paper is clearly published
otherwise use the higher-confidence provider count
never sum the two counts unless you have explicit evidence they are disjoint