100% found this document useful (1 vote)
4K views13 pages

[Turing] Guidelines for RLHF Assessment_2025_DS.pptx

Uploaded by

kpinkyfam274
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
4K views13 pages

[Turing] Guidelines for RLHF Assessment_2025_DS.pptx

Uploaded by

kpinkyfam274
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13

Reinforcement learning from human

feedback (RLHF) Assessment


READ THE INSTRUCTIONS IN THESE SLIDES CAREFULLY. TAKE YOUR TIME.
YOU WILL ONLY GET 1 ATTEMPT ONCE YOU BEGIN THE ASSESSMENT
Overview

We are seeking talented individuals capable of training Large Language Models (LLMs) to solve
real-world problems for the world’s biggest companies.

This RLHF assessment assesses your skills in:

- Analyzing LLM outputs critically and constructively

- Assessing responses for relevance, accuracy, and safety

- Refining AI systems through detailed, context-aware feedback

- English writing proficiency


Guidelines
Test Integrity Guidelines

To ensure a fair evaluation for all candidates, please adhere to the following:

● No additional monitors allowed

● Use only the provided test environment and IDE

● Close unrelated browser tabs

● No online resources or tools (e.g., LeetCode, ChatGPT, search engines) are permitted

● Consult official language/API documentation only

● Complete the assessment independently, without external assistance

● Respect the integrity of the auto-proctoring system

🚩 Failure to comply with these guidelines WILL result in immediate disqualification. We appreciate your
honesty and commitment to a fair process.
RLHF Assessment Objective

- Evaluate English writing proficiency:


- Produce clear, effective, and grammatically correct texts
- Craft high-quality rationales

- Analyze model responses comprehensively:


- Articulate advantages and disadvantages
- Identify objective differences
- Maintain appropriate technical terminology

- Demonstrate attention to detail:


- Compare model responses accurately
- Notice and rank subtle differences
RLHF Test Instructions (30 minutes)

- You will be provided a short conversation based on interactions between a user and an AI model
assistant.

- At the end of the conversation, there will be two potential responses given by the AI model to
answer the user’s last message/request.

- Based on the conversation, you are required to


- Read the full interaction between the User and AI model assistant to understand the context.
- Evaluate and compare the two potential responses by the AI model assistant, and write
analyses considering:
- Response evaluation: Here, we need to select the dropdown options and submit our
evaluation for both model responses.
- Preference explanation: Please provide a brief explanation of your preference,
highlighting why you chose one model response over the other.
- Ideal response: Please generate your own ideal response using the 'Code Execution'
cell, as demonstrated in the upcoming slides.
Test Case Example: Here is the user prompt
Use Case: A team is experiencing intermittent connectivity issues between their web applications and
the backend API
AI Model outputs
based on user
prompt
Please note that the actual fields we collect during the test may differ from these in the example.
Use the 'Code
Executions' cell to
generate an ideal
response, which can
then be included in
the 'Ideal Response'
section as
demonstrated in the
previous slide.
Solution:

Good Example(1/2)
Correctness: Consider the factuality, logical flow, contextual relevance, and
consistency.
Response A's approach provides code that has no major issues in satisfying the user's ask
User: Give me a fast method in python to check if a list contains
for python code to check if a list contains duplicated items. It is highly efficient due to
duplicated items. leveraging Python's set, which only holds unique elements. This allows for constant-time

Model response A: membership checks (O(1)), ensuring quick operations for most dataset sizes.

Response B's approach provides code that has minor issues in satisfying the user's ask for
python code to check if a list contains duplicated items. It is less efficient for larger lists, as
it relies on comparing each element to every other, yielding a time complexity of O(n^2).
Additionally, there's a minor error in using "for i in range(n+1)" instead of the correct "for i in
range(n)".

Model response B:
Informativeness: Is the information provided relevant to the user’s request? Is the
information complete?
Response A's approach is not informative. It has a lack of comments, so a practical usage
example might limit its immediate clarity for some users.

Response B's approach is very informative. Overall extensive commenting enhances its
readability, aiding beginner understanding.

Which one would you choose as the best possible model response? Explain why.
Response A is a better response as it has no issues in correctness and more efficient.
Response A is the preferred way to check for duplicates in most cases. Response B, while
simpler and easier to understand, can be impractically slow for lists with many elements.
The user is looking for a fast solution, so model response A is the best of the two.
Good Example(2/2)
User: Give me a fast method in python to check if a list contains
duplicated items. In this solution:

Model response A:
- The answer is comprehensive, demonstrating that aspects of
correctness and informativeness of both responses were
addressed in the respective input boxes.
- The depth of analysis is commendable, as the two
approaches were compared in terms of their time

Model response B: complexity.


- The most important differences between the two model
responses were identified: comments, test case, incorrect
range in the outer FOR loop and the difference between the
two algorithms (set() vs FOR loops).
- The text is free of spelling/grammar mistakes,
well-structured, coherent and seamlessly articulate the most
important findings of the comparison.

.
Solution:
Bad Example Correctness: Consider the factuality, logical flow, contextual relevance, and
consistency.
User: Give me a fast method in python to check if a list contains Response A is good because set a fast function. Only use here 1 for and 1 if. Code is short.
duplicated items.
Response B is also good because is return corrected result. FOR loops is correct.
Model response A:

Informativeness: Is the information provided relevant to the user’s request? Is the


information complete?
Response B. It has more comments.

Which one would you choose as the best possible model response? Explain why.
Response A is better than B because using set is the best solution.
Model response B:

In this solution:
- The answer is too short lacking a comprehensive comparison of the two model
responses.
- In "Informativeness", response A is not even discussed.
- The depth of analysis is too superficial.
- Some important differences are missing: comments, test case, incorrect range in
outer FOR loop.
- Spelling/grammar mistakes are notorious.

This is not a good solution, even though model response A is indeed better than B, and the
set() is a better approach.
You’re good to start!

Go back to the assessment tab to begin.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy