Add a new research output (model, paper, library, dataset) to a lab's output directory
If an arxiv paper exists, read the HTML version (arxiv.org/html/{id}v1) and extract:
Architecture:
Training:
— new techniques introduced by this model:
Prior innovations used — established techniques the model builds on:
Benchmarks — extract exact scores:
Many models (especially proprietary ones) are announced via blog posts without arxiv papers. In this case:
model.parameters if confirmed by a primary sourceFor every model output, check:
site:artificialanalysis.ai [model name]. Fetch the page and extract the Intelligence Index score. Check for reasoning/adaptive variants — many models have multiple AA entries (e.g., model-name, model-name-reasoning, model-name-adaptive). Use the highest score and link to that variant's page.site:openrouter.ai [model name]. Add the canonical model URL (without date suffix).huggingface.co/org/model)huggingface.co/blog/[org]/[post-slug]. These often contain detailed benchmarks, architecture explanations, and usage guides not found in the model card. Search site:huggingface.co/blog [model name].Some frontier models (GPT-4, Claude, Gemini) are closed-source but have published technical reports:
type: paper (not type: model) since there are no downloadable weightsPapers that introduced techniques now used industry-wide (e.g., "Attention Is All You Need", CLIP, InstructGPT/RLHF, Chain-of-Thought, DPO):
type: paperpapers.ssrn.com) instead of arxiv, especially for economics/policy/interdisciplinary AI research. SSRN blocks automated fetching via Cloudflare — if you encounter this, ask the user to share the paper title/authors directly. Use the SSRN abstract URL as the source link, and paper: (with no arxiv ID) for the structured field.We track scientific models that demonstrate frontier-level capabilities in scientific domains:
Create data/outputs/{lab-slug}/{output-slug}.yaml: