Apply link: https://t.mercor.com/srwnl
Key Responsibilities
- Convert high-level objectives into tightly scoped, testable deliverables with clear inputs/outputs and measurable success criteria.
- Create structured documentation that defines expected behavior, constraints, and edge cases in a way other evaluators can reuse.
- Build lightweight automation scripts to support evaluation flows (e.g., generating required artifacts, validating outputs, enforcing format rules).
- Write deterministic Python verifier scripts that check completion via final state or output validation (files, directories, content assertions).
- Design prompts/tasks that reliably elicit the target workflow behavior while avoiding leakage of internal instructions or implementation details.
- Implement robust error handling and actionable failure messages in verification tooling.
- Develop plausible but ineffective “baseline” or “distractor” approaches to confirm evaluation discrimination (i.e., the solution must use the intended approach).
- Maintain clean artifact hygiene: versionable structure, consistent naming, minimal ambiguity, and reproducible execution.
Ideal Qualifications
- Strong Python skills (file system operations, parsing, validation, test-style assertions, deterministic execution).
- Experience with evaluation harnesses, automated grading, or QA-style verification (unit/integration test mindset).
- Familiarity with prompt design and LLM evaluation methodologies (closed-ended tasks, leakage avoidance, reliability testing).
- Comfort with structured specs and documentation conventions (Markdown, YAML frontmatter patterns, well-scoped requirements).
- Working knowledge of common developer tooling: Git, CLI workflows, virtual environments, dependency management.
- Bonus: embeddings/similarity concepts (e.g., cosine similarity) for “looks relevant but fails” negative-control design.
- Ability to communicate clearly and keep scope controlled without relying on domain-specific context.
#J-18808-Ljbffr…
