Innodata Inc.

Innodata solves your toughest data engineering challenges using artificial intelligence and human expertise.

Applied Research Scientist, LLM Evaluation – Post-Training

Full TimeRemoteTeam 1,001-5,000H1B No SponsorCompany SiteLinkedIn

Location

New Jersey

Posted

11 days ago

Salary

Not specified

Postgraduate Degree5 yrs expEnglishPythonPy TorchTensorflow

Job Description

• Define and execute a research agenda focused on LLM evaluation and post-training, especially evaluation-driven model improvement • Design rigorous experiments to study how evaluation methodologies impact fine-tuning and post-training outcomes • Develop and validate evaluation frameworks for LLM and multimodal systems, including: benchmark/task design scoring methods judge/model-assisted evaluation human evaluation protocols robustness/stress testing • Lead research on advanced evaluation domains, including long-context, cross-modal, and dynamic multi-turn evaluations • Study the effectiveness and limitations of existing evaluation techniques, and propose improved methodologies with clear validity and scalability tradeoffs • Analyze model behavior and failure patterns; generate actionable recommendations for model improvement and evaluation redesign • Collaborate with AI/ML Research Engineers to translate research methods into scalable evaluation and post-training pipelines • Collaborate with Language Data Scientists to integrate human-in-the-loop and synthetic data/evaluation strategies into research programs • Engage with customer technical stakeholders to understand evaluation goals, review methodologies, and provide expert recommendations • Contribute to internal benchmark datasets, evaluation frameworks, and reusable research assets • Produce high-quality technical documentation, internal research reports, and client-facing materials explaining methods, results, assumptions, and limitations • Contribute to thought leadership and best practices in LLM evaluation, post-training, and GenAI quality measurement

Job Requirements

  • MS/PhD in Computer Science, Machine Learning, Statistics, Applied Mathematics, AI, or a related quantitative scientific field (PhD strongly preferred)
  • 5+ years of relevant experience in applied research / research science in ML/AI, with substantial work in LLMs or foundation models
  • Demonstrated experience with LLM evaluation, benchmarking, alignment, post-training, or model quality research
  • Strong foundation in experimental design, statistical analysis, and scientific reasoning for ML systems
  • Strong coding skills in Python for research experimentation and analysis (e.g., data processing, evaluation pipelines, statistical analysis, visualization)
  • Experience working with modern ML tooling/frameworks (e.g., PyTorch, Hugging Face, JAX/TensorFlow as applicable)
  • Ability to evaluate and compare human and automated evaluation methods, including tradeoffs in cost, reliability, validity, and scalability
  • Experience designing evaluation studies and protocols that are reproducible across datasets, model versions, and evaluation runs
  • Ability to collaborate directly with technical stakeholders including research scientists, ML engineers, data scientists, and customer technical counterparts
  • Strong communication skills and ability to present nuanced technical conclusions, assumptions, and limitations clearly.

Benefits

  • Health insurance
  • Retirement plans
  • Paid time off
  • Flexible work arrangements
  • Professional development opportunities

Related Categories

Related Job Pages