Trigger this skill when the user asks for questions about how far something is, how big it is, which point is closer, how tall a desk is, or whether an embodied agent can estimate physical size from what it sees. Plain-language triggers include: 'near or far,' 'how deep is that point,' 'how tall is the object,' 'measure the desk,' and 'questions about exact size instead of just naming things.'
[Case 1]
[Case 2]
[Case 3]
To synthesize data for this capability, you must strictly follow a 3-phase pipeline. Do not hallucinate steps. Read the corresponding reference file for each phase sequentially:
Phase 1: Environment Exploration
Read the exploration guidelines to discover raw knowledge seeds:
references/EXPLORATION.md
Phase 2: Trajectory Selection
Once Phase 1 is complete, read the selection criteria to evaluate the trajectory:
references/SELECTION.md
Phase 3: Data Synthesis
Once a trajectory passes Phase 2, read the synthesis instructions to generate the final data:
references/SYNTHESIS.md