Search by job, company or skills

Innodata Inc.

Chinese Language Expert

Early Applicant
  • Posted 5 days ago
  • Be among the first 10 applicants
Fresher

Job Description

We are looking for native Chinese Experts from Taiwan to contribute on our on-going AI/LLM project.

Position Overview:

  • Job Title: Chinese Language Expert
  • Experience Level: Degree in Linguistics or a related field
  • Location: Remote | Freelance
  • Language: Chinese
  • Location: Taiwan (native)
  • Mandatory: To complete LLM Evaluation assessment by coming Sunday, September 28th.
  • LLM Evaluation Assessment Link: https://icap.innodata.com/registerfreelancerenc=oUTZVsr/Pnz/0Xygc2EK32MdtinqnjC9vy8RU3Ha4EP5XfySiaGIBP8Cxl4qcKu2lYthJblueSandPblueSandddNWdugfpU7t2wezikielezikiel

LLM Evaluation Guidelines:

  • No. of Questions: 122
  • No. of Sections: 7
  • Test Duration: 120 minutes (2 hours)
  • Total Marks: 122
  • Passing Criteria: High (Get at least 100+ questions correct to get passed)
  • Required: Complete by Sunday, 28th September
  • Note: Assessment gets auto submitted after 120 minutes, maintain speed with accuracy.

Job Description:

We are seeking highly analytical and detail-oriented professionals with hands-on experience in Red Teaming, Prompt Evaluation, and AI/LLM Quality Assurance. The ideal candidate will help us rigorously test and evaluate AI-generated content to identify vulnerabilities, assess risks, and ensure compliance with safety, ethical, and quality standards.

Key Responsibilities:

  • Conduct Red Teaming exercises to identify adversarial, harmful, or unsafe outputs from large language models (LLMs).
  • Evaluate and stress-test AI prompts across multiple domains (e.g., finance, healthcare, security) to uncover potential failure modes.
  • Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
  • Collaborate with data scientists, safety researchers, and prompt engineers to report risks and suggest mitigations.
  • Perform manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence.
  • Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
  • Document findings, edge cases, and vulnerability reports with high clarity and structure.

Requirements:

  • Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
  • Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
  • Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
  • Understanding of LLM behaviors, failure modes, and model evaluation metrics.
  • Excellent critical thinking, pattern recognition, and analytical writing skills.
  • Ability to work independently, follow detailed evaluation protocols, and meet tight deadlines.

Preferred Qualifications:

  • Prior work with teams like OpenAI, Anthropic, Google DeepMind, or other LLM safety initiatives.
  • Experience in risk assessment, red team security testing, or AI policy & governance.
  • Background in linguistics, psychology, or computational ethics is a plus.

More Info

Industry:Other

Function:Ai Llm Project

Job Type:Permanent Job

Date Posted: 28/09/2025

Job ID: 127546941

Report Job

About Company

View More
Last Updated: 02-10-2025 09:08:53 PM
Home Jobs in Taiwan Chinese Language Expert