We are seeking multilingual raters to assist in creating evaluation sets for LLM model by using the product based on a given scenario. Raters will simulate real-world research tasks, collect source documents, generate prompts, and evaluate responses to ensure high-quality model outputs.
Responsibilities:
Provide feedback on usability, accuracy, and comprehensiveness of LLM outputs.
Requirements: