You will play a key role in advancing how artificial intelligence understands and produces software by guiding large-language models through expert-level code evaluation. Your primary responsibility is to analyze, compare, and rank AI-generated C# code snippets, identifying strengths and flaws in logic, structure, and performance.
What You'll Do
- Assess multiple code solutions for the same problem and determine the most effective one based on correctness, efficiency, and maintainability
- Refactor and correct code to meet production standards, fixing bugs and improving design
- Provide clear, well-reasoned written justifications for your evaluations
- Feed detailed feedback—including edits, ratings, and test outcomes—into the model training pipeline
- Help the system learn to critique and enhance its own code suggestions through iterative human input
What We're Looking For
- At least three years of professional experience developing in C#
- Proven ability to quickly identify issues in code, including logic flaws, inefficiencies, and security concerns
- Exceptional attention to detail and a methodical approach to problem-solving
- Strong written communication skills, especially in explaining technical trade-offs
- Comfort reading language documentation and technical specifications
- Ability to work independently in an asynchronous, low-interruption environment
Nice to Have
- Experience with constraint programming or formal logic systems
Environment & Benefits
- Fully remote—work from any location
- Direct impact on model performance with no organizational overhead
- Asynchronous culture that values clarity and precision in writing
- Minimal meetings and bureaucracy—focus on meaningful work
Compensation
Hourly rate ranges from $30 to $70, based on experience and regional benchmarks. The role operates on a global hiring model with full remote flexibility.


