Responsible AI

Download as pdf or txt
Download as pdf or txt
You are on page 1of 45

Week-1 RSAI 2024

1. Consider the following prompt: “A squirrel gives an apple to a


bird”. The below images are generated from two image
generation models. Which of the models has a higher number of
trainable parameters?

Fig 2A Fig 2B
a. Model generating image Fig 1A
b. Model generating image Fig 1B
c. Insufficient Information
d. Parameters have no importance

2.
Fig 3
What language-based task is being performed in Fig 2?
a. Common-sense QA
b. Planning and Strategic Thinking
c. Paraphrasing
d. Text Generation

3.
Fig 4
Which of the following tasks by AI surpasses the human
benchmark according to Fig 4?
a. Language Understanding
b. Code Generation
c. Neither
d. Both

4.
Fig 5A - Task A
Fig 5B - Task B

What are Task A and Task B


a. A: Planning & Strategy, B: Common sense & Reasoning
b. A: Text generation, B: Common sense & Reasoning
c. A: Common sense & Reasoning, B: Text Generation
d. A: Common sense & Reasoning, B: Planning & Strategy
5.

Fig 6A

Fig 6B
Is there any issue with the model’s response in Fig 5A and Fig
5B?
a. The model is biased about career
b. The model is honest
c. The model has no issue
d. The model is biased towards gender
6.
A B C D
Identify the source, destination, mask, and output for a deepfake
generation.
a. A: Source, B: Destination, C: Mask, D: Output
b. A: Destination, B: Mask, C: Source, D: Output
c. A: Mask, B: Source, C: Destination, D: Output
d. None of the above

7.
Fig 8
What issue caused the catastrophe in Fig 6?
a. Bias in Algorithm
b. Variance in Algorithm
c. Error in Algorithm
d. Slowness in Algorithm
8. Which of the following are risks associated with AI.
a. Malicious use ​
b. AI plateau ​
c. Organizational risks ​
d. Self-driving cars

9. “Competition could push nations and corporations to rush AI


development, relinquishing control to these systems.”
The following describes which kind of risk associated with AI.
a. Malicious Use
b. AI race
c. Organizational Risks
d. Option B and C both

10. What are some solutions to fight the organizational risks of


AI?
a. Safety Culture
b. CyberTerrorism
c. Red Teaming
d. Anomaly Enforcement

11. Which of the human-like behaviours cause AI agents to go


rogue?
a. Jealousy
b. Power-seeking
c. Self-preservation
d. Empathy
12. Identify which components the below belong to in a disaster
risk equation.
Alignment - i
Robustness - ii
Monitoring - iii
a. i - Vulnerability, ii - Hazard Exposure, iii - Hazard
b. i - Hazard Exposure, ii - Hazard, iii - Vulnerability
c. i - Hazard, ii - Vulnerability, iii - Hazard Exposure
d. None of the above

13. What does it mean for a model to be “Aligned”?


a. The model should be reliable, helpful, and safe
b. The model is fast
c. The model can perform several tasks
d. The model should speak multiple languages

14. RoBERTa _______ in reasoning tasks compared to BERT.”


Fill in the blanks.
a. Succeeds
b. Fails
c. Remains the same
d. Performs better for some questions, worse for other questions.

15. Which of the following are some speculative hazards and


failure modes.
a. Cyberdefense
b. Proxy gaming
c. Emergent goals
d. Empathy
16. A
____________________________________________________

B
____________________________________________________

C
____________________________________________________

Identify the behavior required to be followed by the AI agents in


the above images.
a. A: Robustness, B: Monitoring, C: Alignment
b. A: Alignment, B: Robustness, C: Monitoring
c. A: Monitoring, B: Alignment, C: Monitoring
d. None of the above
17. “Deep RL methods outperform humans in simulated aerial
combat​.”
What kind of hazard can this cause?
a. CyberTerrorism
b. Deception
c. Weaponized AI
d. Proxy gaming

18. Which of these are not a part of Red Teaming a model?


a. Checking the toxicity of the model by querying harmful content
b. Checking the model generation speed to your queries
c. Checking the bias in the model against certain genders and
race
d. Querying the model for your questions and doubts

19. How do you measure the similarity between two protein


structures?
a. Local Distance Test
b. Global Similarity Test
c. Global Distance Test
d. Local Similarity Test

20. If the hardware access for AI building is limited to a few, it can


cause severe inequality in which of the following?
a. Knowledge
b. Economy
c. Safety
d. Power
21. “A team to identify vulnerabilities in AI systems within an
organisation that builds AI systems.”
The above defines which of the Organizational Risk’s safety
measures?
a. Anomaly Detection
b. Red Teaming
c. Cyberattack
d. Safety Testing
e. Vulnerability Detection

22. What is the correct order of the steps involved in reducing the
risk associated with AI systems.
a. Respond -> Prioritize -> Identify -> Improve
b. Prioritize -> Identify -> Improve -> Respond
c. Respond -> Prioritize -> Identify -> Improve
d. Identify -> Prioritize ->Respond -> Improve

23. Codex LM is used for which of the following?


a. Convert code to its natural language description
b. Convert math problems into step-by-step natural language
explanation
c. Convert natural language description of programming task to
code
d. Convert natural language explanations of math problems to
their numerical equivalents
e. Convert code from one programming language to the other

24. Which of the following do you think are capable of


discovering novel algorithms?
a. Google’s Minerva
b. OpenAI’s Codex
c. Deepmind’s AlphaFold3
d. BigScience’s BLOOM
e. Deepmind’s AlphaTensor
f. StackOverFlow’s OverflowAI
ROBUSTNESS, RLHF, AI Alignment - WEEK 2

1. A robust model provides unreliable predictions when met with adversaries. Which
all of the following are common adversaries in this context?
a. Distribution Shift
b. Overfitting
c. Noisy Data
d. Model Compression
e. Gradient Descent
f. Data Augmentation

2. In the context of AI research, which of the following events could be considered a


black swan?
a. Incremental improvements in natural language processing algorithms.
b. The consistent performance of AI models on standard benchmarks.
c. The sudden discovery that a widely-used AI model has a critical flaw, leading to
significant ethical and legal repercussions.
d. The publication of a research paper revealing a minor improvement in an AI
algorithm.
e. A gradual increase in the accuracy of AI models over time.

3. To train a model that achieves accuracy in the range of 95% to 98%, you need 1GB
of data. To get 100% accuracy, you need 120GB of data. This idea is similar to
which of the following principles:
a. Sigmoid Distribution
b. Power law distribution
c. Uniform distribution
d. Gaussian Distribution
e. Long-tailed distribution

4. Identify the equations that can lead to a long-tailed distribution.


a. Idea * student * resources * time
b. Idea * student + resources * time
c. Idea + student + resource + time
d. Idea - student * resource - time
5. Black Swan lies in which of the following categories?
a. Known Knowns
b. Known Unknowns
c. Unknown Knowns
d. Unknown Unknowns

6. Match the items below with their corresponding descriptions.


Column A Column b
I. Known Knowns A. Close-ended questions
II. Known Unknowns B. Recollection
III. Unknown Knowns C. Open-ended exploration
IV. Unknown Unknowns D. Self-Analysis
a. I-C, II-D, III-A, IV-B
b. I-B, II-A, III-D, IV-C
c. I-A, II-B, III-C, IV-D
d. I-D, III-C, III-B, IV-A

7. Why is Black Swan and Long-tailed distribution important?


a. Understand small things that have a small but useful effect
b. Understand large things that have a small but useful effect
c. Understand small things that have a catastrophic effect
d. Understand large things that have a catastrophic effect

8. To check if an image classification model is robust, identify all the training and
testing processes that can be used from below. The three datasets are ImageNet,
AugMix, Mixup
a. Train on AugMix and test on AugMix
b. Train on AugMix and test on ImageNet
c. Train on ImageNet and test on AugMix
d. Train on Mixup and test on ImageNet
e. Train on ImageNet and test on ImageNet
f. Train on ImageNet and test on Mixup

9. Identify all the conditions to check if a model is robust?


a. Models with larger parameters
b. Models with small parameters
c. Models that can generalise better
d. Models trained to perform the best on a specific type of data

10. Which of the following are some data augmentation methods?


a. DataShrink
b. AugMix
c. Label Flipping
d. Mixup

11. The introduction of new lighting conditions in an image dataset would most likely
cause?
a. Distribution Shift
b. Concept Shift
c. Model Decay
d. Feature Extraction

12. Identify the goal(s) of a model when training with RLHF is as follows:
a. Maximize the penalty
b. Maximize the reward
c. Minimize the penalty
d. Minimize the reward

13. Identify the step(s) involved in RLHF pipeline:


a. Supervised fine-tuning
b. Unsupervised fine-tuning
c. Reward model training
d. Penalty model training
e. Proximal Policy optimization
f. Convex Policy Optimization

14. Identify issue(s) associated with RLHF from below:


a. It does not perform as well as supervised learning
b. Performance sensitive to hyperparameters
c. It does not perform as well as unsupervised learning
d. Fitting and optimization of the reward function is computationally expensive
e. Pretrained models easily outperform them in tasks like summarization

15. What is the constraint under which the model optimization is done in RLHF to
ensure that the model doesn’t diverge too far from the pretrained model?
a. KL Divergence
b. L2 Regularization
c. Entropy Maximization
d. Gradient Clipping
16. What are the issues with reward modelling?
a. Reward shrinking - gradually decreasing rewards over time
b. Reward misalignment - reward signals do not align with the desired outcomes
c. Reward saturation - model stops learning after a certain reward threshold is
reached
d. Reward consistency - ensuring rewards are uniformly distributed
e. Reward hacking - maximise reward with imperfect proxy and forget the goal

17. Direct Preference of Optimization works in which one of the following ways:
a. RLHF without rewards model
b. RLHF without human feedback
c. RLHF without reinforcement learning
d. RLHF without KL divergence

18. Identify the issues with human feedback in RLHF


a. Overabundance of feedback
b. Consistent and uniform feedback
c. Biased and harmful feedback
d. Feedback redundancy

19. Identify the way(s) to maintain transparency in the context of RLHF to avoid safety
and alignment issues.
a. Quality-assurance measures for human feedback
b. Minimize the involvement of humans to reduce biases
c. Use black-box algorithms to simplify the process
d. Avoid documenting the feedback process to save time
e. Limit the diversity of human feedback to ensure consistency
f. Have a powerful loss function when optimizing the reward model

20. What is distribution shift in machine learning?


a. The training distribution is not similar to the test distribution
b. The model's parameters change during training
c. The target variable's distribution changes over time
d. The model's prediction accuracy improves on new data

21. What is data poisoning in the context of machine learning?


a. Removing important features from the dataset
b. Oversampling minority classes in imbalanced datasets
c. Adding hidden functionalities to control the model behaviour
d. Encrypting sensitive information in the training data

22. When is a data poisoning attack considered successful?


a. The model's overall accuracy decreases significantly
b. The model fails to converge during training
c. The model becomes computationally inefficient
d. The model outputs the specific behaviour when it encounters the trigger

23. Consider the following scenario:


You have a dataset with 10000 samples and a model trained over it has a test
accuracy of almost 94%. You then introduce a trojan data poisoning attack in your
dataset such that every time it looks at a certain trigger pattern, the model
behaves in a certain way. You get a success rate of 99% for your data poisoning
attack through the trigger by poisoning 0.01% of your samples. What is the new
test accuracy of your model? Identify the correct range.
a. 40 to 50 percent
b. 50 to 60 percent
c. 60 to 70 percent
d. 70 to 80 percent
e. 80 to 90 percent
f. 90 to 100 percent

24. What are the different defence mechanism(s) against poisoning attacks?
a. Biasing
b. Filtering
c. Unlearning
d. Representation engineering
e. AutoDebias
UNLEARNING
1. What kind of content information do you want to remove from the model data?
a. Biased or discriminatory data
b. Useful patterns and trends
c. General public data
d. Random noise
e. Personally identifiable information
f. Valid and accurate data

2. What are the reasons to choose unlearning over retraining?


a. To improve the overall performance of the model
b. To add new data to the model
c. To change the underlying algorithm of the model
d. To completely overhaul the model's architecture
e. To save computational resources and time

3. Identify the steps involved in the exact unlearning as discussed in the course.
a. Isolate the data -> shard the data -> slice the data -> aggregate the data
b. Aggregate the data -> isolate the data -> slice the data -> shard the data
c. Shard the data -> Slice the data -> Isolate the data -> Aggregate the data
d. Shard the data -> Isolate the data -> Slice the data -> Aggregate the data
e. Isolate the data -> slice the data -> shard the data -> aggregate the data

4. Which model should be retrained in the exact unlearning process?


a. The constituent model that is trained over the isolated data
b. The constituent model that is trained over the sharded data
c. The constituent model that is trained over the aggregated data
d. The constituent model that is trained over the sliced data

5. How should the original model and the model after the below unlearning methods
behave?
1) exact unlearning
2) approximate unlearning
a. 1) distributionally identical 2) distributionally identical
b. 1) distributionally close 2) distributionally close
c. 1) distributionally identical 2) distributionally close
d. 1) distributionally close 2) distributionally identical
6. How does unlearning via differential privacy work?
a. check whether an adversary can reliably tell apart the models before unlearning
and after unlearning
b. check whether the model can output private and sensitive information before and
after unlearning
c. check whether the model's predictions become more consistent and stable for
private information before and after unlearning.
d. check whether an adversary can identify the differences in the distribution of
output data of the model before and after unlearning

7. Identify all the methods for privacy unlearning.


a. Gradient descent on encountering the forget set
b. Remove noise from the weights influencing the forget set
c. Add noise to weights influencing data in forget set
d. Gradient ascent on encountering the forget set
e. Increase the learning rate when encountering the forget set
f. Apply dropout to all layers when encountering the forget set

8. Match the unlearning method to their corresponding concept


1) privacy unlearning I. data and model architecture is not modified
2) concept unlearning II. use membership inference attack concept
3) example unlearning III. forget set is not clearly defined
4) ask for unlearning IV. forget set is clearly defined
a. 1-III, 2-I, 3-IV, 4-II
b. 1-II, 2-III, 3-IV, 4-I
c. 1-IV, 2-II, 3-I, 4-III
d. 1-I, 2-IV, 3-II, 4-III
e. 1-IV, 2-I, 3-III, 4-II

9. The forget set to be unlearned is not known in which of the following:


a. Example Unlearning
b. Differential Privacy Unlearning
c. Privacy unlearning
d. Concept unlearning

10. In the scenario of ask for unlearning, what kind of things can be easily unlearned?
a. Hate speech
b. Toxic content
c. Factual Information
d. Sensitive information
11. When evaluating the quality of unlearning using Membership Inference Attack, which of
the following scenarios implies that the unlearning is successful?
a. The accuracy increases on the forget set
b. The accuracy drops on the forget set
c. The accuracy stays the same on the forget set
d. The accuracy increases on the test set
e. The accuracy drops on the test set
f. The accuracy stays the same on the test set

12. What are some metrics to evaluate the unlearning?


a. If it was more computationally efficient compared to retraining
b. Increased size of the original dataset
c. If the unlearning retains information derived from the concept to be forgotten
d. If the performance has been maintained before and after unlearning

13. In an interclass confusion scenario where confusion is synthetically added to a dataset


by label flipping for some of the concepts, identify the kind of unlearning method that can
be used to unlearn the data points that have their labels flipped. Assume that you have
the entire data points for which the labels were flipped.
a. Concept unlearning
b. Example Unlearning
c. Differential Privacy Unlearning
d. Exact unlearning
e. Ask to forget

14. What idea does the paper Corrective Machine Learning build upon?
a. Not all poisoned data can be identified for unlearning
b. Identifying and removing a small subset of poisoned data points is sufficient to
ensure the model's integrity
c. enhancing the model's ability to handle completely new, unseen poisoned data
d. The accuracy of the model improves proportionally with the amount of data
removed, regardless of whether it is poisoned or not
e. adding redundant data to the dataset to counteract the effects of poisoned data.
f. Not all poisoned data can be identified for unlearning

15. Identify all the methods that act as the baseline for the TOFU benchmark dataset
a. Gradient Descent
b. Gradient Ascent
c. Gradient Difference
d. Gradient boosting
e. Gradient Clipping
16. The WMDP benchmark tests on unlearning what kind of information?
a. Biosecurity
b. High-school biology
c. Hate speech on Twitter
d. Crime data

17. You are in charge of building graph models trained on Instagram social networks to
provide content recommendations to users based on their connections’ content. You
realize that a particular user in the network is leading to toxic content recommendations.
What kind of unlearning would you use in this scenario to prevent the recommendation
of toxic content?
a. Node feature unlearning
b. Node unlearning
c. Edge Unlearning
d. Subgraph unlearning

18. In Representation Engineering, what is the representation?


a. Attention heads affecting the data
b. Positional embeddings of the data
c. Activations of the layer affecting the data
d. The encoder of a transformer
e. The decoder of a transformer
BIAS - I : Assignment 4
1. Which of the following define bias? (Select all that apply)
a. Systematic favoritism towards certain groups
b. Random errors in data collection
c. Neutral and unbiased judgment
d. Systematic exclusion of certain data points
e. Random selection of samples
f. Systematic deviation from rationality in judgment

2. Which of the following is one of the ways to measure bias


a. Random sampling
b. Cross-Validation
c. Using Benchmark datasets
d. Chi-Square test

3. Which of the following statements is true?


a. Demographics do not influence the perception of bias.
b. Demographics can influence the perception of bias.
c. Perception of bias is unaffected by demographics.
d. Demographics and bias perception are unrelated.

4. What does each pair in the CrowS-Pairs dataset consist of?


a. One stereotype sentence and one neutral sentence
b. One biased sentence and one unbiased sentence
c. One stereotype sentence and one opposite stereotype sentence
d. One stereotype sentence and one less stereotype sentence

5. The statement “Women are bad drivers” is a


a. Stereotype
b. Anti-Stereotype
c. Non-Stereotype
d. Neutral Statement

6. Which of the following is NOT a common source of bias in data?


a. Historical inequities reflected in the data
b. Data collection methods that over-represent certain groups
c. Balanced representation of all demographic groups
d. Labeling errors or subjective annotations
7. At which stage of the machine learning process can bias be introduced by over- or
under-representing certain groups?
a. Data Collection
b. Data Pre-processing
c. Data Annotation
d. All of the above

8. Which of the following characteristics can make a statement biased?


a. Being stereotypical
b. Being socially aligned
c. Having inaccurate information
d. Having emotional language

9. One of the reasons bias exists in models is:


a. Algorithm complexity
b. Data
c. Model architecture
d. Training duration

10. When is bias a problem?


a. When it reduces the models performance
b. When it has negative impact
c. When it aligns with train data but not test data
d. When it increases the complexity

11. Which of the following is the correct setting for contrastive learning?
a. Irrespective of the sentences, minimise the distance between their embeddings
b. Irrespective of the sentences, maximise the distance between their embeddings
c. If sentences are similar, minimise the distance between their embeddings
d. If sentences are different, minimise the distance between their embeddings
e. If sentences are different, maximise the distance between their embeddings
f. If sentences are different, minimise the distance between their embeddings
BIAS - II : Assignment 5

1. What does SEAT stand for?


a. Sentence Evaluation and Association Test
b. Semantic Embedding and Association Test
c. Sentence Embedding Association Test
d. Sentiment Embedding Analysis Test

2. What is coreference resolution?


a. Identifying all mentions in a text referring to different entities
b. Finding all mentions in a text that refer to the same real-world entity
c. Analyzing sentence structure for grammatical correctness
d. Determining the sentiment of a text

3. For a reliable COBIAS score, the context variance (CV) should:


a. Tend to 0
b. Tend to 1
c. Be maximized
d. Be minimized to 1

4. Given the sentences "A man is walking" and "A nurse is walking," a higher similarity
score indicates:
a. More bias
b. Less bias
c. No bias
d. Ambiguous bias

5. Pointwise Mutual Information (PMI) is used to measure the association between two
events. Which of the following statements best describes PMI?
a. PMI measures the likelihood of two events occurring together compared to the
likelihood of their independent occurrence.
b. PMI is a measure of the difference between the probabilities of two events
occurring together and separately.
c. PMI calculates the expected frequency of two events occurring together.
d. PMI assesses the correlation between two events based on their joint probability
distribution.
6. Consider the following prompt: "The image shows a subject with occupation
'{occupation}'. What is the gender of the central subject in this image?"

This prompt provides specific information about the subject's occupation in the image
before asking about their gender. Based on this description, how would you classify the
prompt?
a. Informed Task
b. Blind Task
c. Follow-up Task
d. Generic Task

7. Which of the following statements is true?


a. Proprietary models are more neutral compared to CoDi and other open-source
models.
b. Proprietary models are less neutral compared to CoDi and other open-source
models.
c. There is no difference in neutrality between proprietary models and CoDi or other
open-source models.
d. Neutrality in models is not related to whether they are proprietary or
open-source.

8. Mitigating bias can be done through:


a. Fine-Tuning
b. Increasing the parameters
c. Using encoder models
d. Guardrails

9. In the paper AutoDebias, we search for bias by _________ JS Divergence and debias
by ___________ JS Divergence.
a. Maximizing, Minimizing
b. Minimizing, Maximizing
c. Maximizing, Not Changing
d. Not Changing, Minimizing

10. What does a high language modeling score indicate?


a. The model ranks meaningless associations higher than meaningful associations.
b. The model ranks meaningful associations higher than meaningless associations.
c. The model equally ranks meaningful and meaningless associations.
d. The model's performance is not related to ranking associations.
11. What does the Idealized CAT score represent?
a. Combination of language model score and bias score
b. Combination of stereotype score and fairness score
c. Combination of language model score and stereotype score
d. Combination of language model score and accuracy score
DIFFERENTIAL PRIVACY - 1
1. Which of the following methods is the best method to efficiently protect the data to
preserve the privacy of the users?
a. Anonymization
b. Cryptographical Solution
c. Statistical solution
d. Data Compression
e. Data Duplication

2. Between a randomized response (with epsilon>0) and a fair coin toss response, which
algorithm would you use to preserve privacy but have a better utility?
a. Randomized response because the chance of falsehood is 50%
b. Randomized response because the chance of truth is greater than 50%
c. Randomized response because the chance of falsehood is greater than 50%
d. Coin toss response because the chance of falsehood is 50%
e. Coin toss response because the chance of truth is greater than 50%
f. Coin toss response because the chance of falsehood is greater than 50%

3. Consider the equation in the context of privacy guarantees (The notations used are the
same as used during the lecture).
−ε ε
𝑃(𝑅𝑅(𝑥') = 𝑏) * 𝑒 ≤ 𝑃(𝑅𝑅(𝑥) = 𝑏) ≤𝑃(𝑅𝑅(𝑥') = 𝑏) * 𝑒

To maximize the privacy gains, which of the following values should be changed and
how?
a. ε should be maximum for privacy, ε should be minimum for utility
b. ε should be minimum for privacy, ε should be minimum for utility
c. ε should be maximum for privacy, ε should be maximum for utility
d. ε should be minimum for privacy, ε should be maximum for utility
e. ε is unrelated
4. Consider the below values:
X = {x1,x2,.....xN} is the truth of an experiment
Y = {y1, y2,......yN} is the revealed values instead of the truth
To identify the average of truth, Y as an estimator cannot be used for the process by
which it was obtained. You derive new values Z where Z = {z1,z2,......zN} from Y which
are better estimators of X. How do you arrive at the values Z?
a. Removing the bias from Y introduced through the random process
b. Adding the bias to Y removed through the random process
c. Removing the variance from Y introduced through the random process
d. Adding the variance to Y removed through the random process

5. If ε is fixed, given a privacy guarantee, to improve the utility, which of the following values
can be modified?
a. Increase the number of experiments
b. Increase the amount of randomness
c. Increase the amount of bias introduced in the random process
d. Increase the amount of variance introduced in the random process

6. Identify the equation for the ε-differential mechanism (The notations used are the same
as used during the lecture): :
𝑃(𝑀(𝑥)Є 𝑆 ) ε
a.
𝑃(𝑀(𝑥')Є 𝑆 )
≤𝑒
𝑃(𝑀(𝑥')Є 𝑆 ) ε
b.
𝑃(𝑀(𝑥')Є 𝑆 )
≤𝑒
𝑃(𝑀(𝑥)Є 𝑆 ) ε
c.
𝑃(𝑀(𝑥)Є 𝑆 )
≤𝑒
𝑃(𝑀(𝑥)Є 𝑆 ) ε
d.
𝑃(𝑀(𝑥')Є 𝑆 )
≥𝑒
𝑃(𝑀(𝑥')Є 𝑆 ) ε
e.
𝑃(𝑀(𝑥')Є 𝑆 )
≥𝑒
𝑃(𝑀(𝑥)Є 𝑆 ) ε
f.
𝑃(𝑀(𝑥)Є 𝑆 )
≥𝑒
7. Identify the correct scenario in the case of differential privacy
a. Trust the curator; Trust the world
b. Do not trust the curator; Trust the world
c. Trust the curator; Do not trust the world
d. Do not trust the curator; Do not trust the world

8. Identify all the values representing sensitivity in a laplacian mechanism where the
function under consideration is an average of n binary values {0,1} (The notations used
are the same as used during the lecture).
1
a.
𝑛
1
b.
𝑛
| 𝑥𝑛' − 𝑥𝑛 |
c. ε
ε
d.
𝑛
−1
e. 𝑛
Δ
f.
ε

9. Identify the distribution from which the noise is derived in a laplacian mechanism. The
representation is of the form Laplacian(a,b) where a is the mean and b is the spread
parameter. (The notations used are the same as used during the lecture)
Δ
a. laplacian(1, )
ε
Δ
b. laplacian( ,0)
ε
Δ
c. laplacian( ,1)
ε
Δ
d. laplacian(0, )
ε
e. laplacian(1,1)
f. laplacian(0,0)
10. Higher privacy guarantees can be achieved in which of the following scenarios? Identify
all the possible scenarios.
a. Epsilon should be high
b. Inverse Sensitivity should be high
c. Variance should be high
d. Noise should be high
e. Utility should be high

11. Identify the deviation of the value from the truth in the scenario of a laplacian
mechanism. (The notations used are the same as used during the lecture).
1
a. o( )
ε𝑛
𝑛'
b. o( )
ε𝑛
ε
c. o( )
𝑛
𝑒
d. o( )
ε𝑛
e. o(ε𝑛)

12. In the scenario of a privacy-utility trade-off, for fixed privacy, the number of samples
required for a particular utility varies between the Laplacian mechanism and
Randomized response is different by what factor?
a. Constant factor
b. Linear factor
c. Exponential factor
d. Logarithmic factor
e. Quadratic factor
Assignment week 7

1. When calculating the sensitivity in ε-Differential Privacy where the values to be derived
from the data points is a d-dimension vector, identify the normalisation technique.
(Notations are the same as used in the lecture)
a. Manhattan normalisation
b. Eucledian normalisation
c. Max normalisation
d. Min-max normalisation
e. Sigmoid normalisation

2. In (ε, δ)- Differential privacy what does δ=0 imply? (Notations are the same as used in
the lecture)
ε
a. The equation (𝑃(𝑀(𝑥)ϵ 𝑆) ≤ 𝑒 (𝑃(𝑀(𝑥')ϵ 𝑆) should hold for some of the subsets
S
ε
b. The equation (𝑃(𝑀(𝑥)ϵ 𝑆) ≤ 𝑒 (𝑃(𝑀(𝑥')ϵ 𝑆) should hold for most of the subsets
S
ε
c. The equation (𝑃(𝑀(𝑥)ϵ 𝑆) ≤ 𝑒 (𝑃(𝑀(𝑥')ϵ 𝑆) should hold for all of the subsets S
ε
d. The equation (𝑃(𝑀(𝑥)ϵ 𝑆) ≤ 𝑒 (𝑃(𝑀(𝑥')ϵ 𝑆) should hold for none of the subsets
S

3. How do the utilities vary in the Laplacian mechanism vs the Gaussian mechanism in a
higher dimension differential privacy setting?
a. As the dimension increases, the Gaussian mechanism requires quadratically
more amount of noise than the Laplacian mechanism, decreasing the utility
b. As the dimension increases, the Gaussian mechanism requires quadratically
lesser amount of noise than the Laplacian mechanism, decreasing the utility
c. As the dimension increases, the Gaussian mechanism requires quadratically
lesser amount of noise than the Laplacian mechanism, increasing the utility
d. As the dimension increases, the Gaussian mechanism requires quadratically
more amount of noise than the Laplacian mechanism, increasing the utility

4. _____ property ensures that a function applied on the privacy-protected data _____ its
privacy aspect after applying a function over it.
a. i. Post-processing ii. Retains
b. i. Post-processing ii. Loses
c. i. Composition ii. Retains
d. i. Composition ii. Loses
5. After using k mechanisms for getting k (ε, δ)- differentially private data variations for a
dataset, the combined leakage that is observed from these k mechanisms can be
minimized by:
a. Using Laplacian Mechanism
b. Using Gaussian Mechanism
c. Using Uniform Mechanism
d. Using Exponential Mechanism

6. In a buyer-seller problem, given n buyers and n valuations by the buyers, what is the
total revenue given a price p.
𝑛
a. 𝑝 ∑ 𝐴 𝑤ℎ𝑒𝑟𝑒 𝐴 = 1 𝑖𝑓 𝑣𝑖 ≥ 𝑝 𝑎𝑛𝑑 𝐴 = 0 𝑖𝑓 𝑣𝑖 ≤ 𝑝
𝑖=𝑛
𝑛
b. 𝑝 ∑ 𝐴 𝑤ℎ𝑒𝑟𝑒 𝐴 = 0 𝑖𝑓 𝑣𝑖 ≥ 𝑝 𝑎𝑛𝑑 𝐴 = 1 𝑖𝑓 𝑣𝑖 ≤ 𝑝
𝑖=𝑛
c. 𝑝𝑛
d. 𝑝(𝑛 − 1)
e. 𝑝(1/𝑛)

7. In the exponential mechanism to calculate the price to maximize the revenue, identify the
correct statement in the scenario where 2 unequal prices result in the same revenue:
a. Both prices have an unequal probability of being selected
b. Both prices have an equal probability of being selected
c. A higher price has a higher probability of being chosen due to normalisation
d. A lower price has a higher probability of being chosen due to normalisation

8. In a classification problem, if a data point lies on a hyperplane that perfectly separates


the two classes, the probability of the data point belonging to class A is:
a. 25%
b. 50%
c. 75%
d. 100 %

9. In a vanilla Principle Component Analysis method, the reconstruction loss of a protected


group is _______ than the remaining data before resampling and _______ than the
remaining data after resampling.
a. Higher, higher
b. Higher, lower
c. Lower, higher
d. Lower, lower
10. The goal of a Fair PCA is to find a PCA solution U where U=[Ua, Ub] such that
reconstruction loss of the two groups A and B where A is the protected group is:
a. Equal
b. Unequal
c. The protected group has a lower reconstruction loss
d. The protected group has a higher reconstruction loss

11. In an ideal situation where the models are completely fair, the different parity values are:
a. Approach 0
b. 1
c. Approach 1
d. 0

12. Match the following:


i. 𝑃(𝑀(𝑥) = 1 | 𝑥 𝑖𝑛 𝐶) − 𝑃(𝑀(𝑥) = 1) a. Fair Logistic
regression
ii. 𝑃(𝑀(𝑥) = 1 | 𝑦 = 1 𝑎𝑛𝑑 𝐶) − 𝑃(𝑀(𝑥) = 1 | 𝑦 = 1) b. Statistical Parity
iii. 𝑃(𝑀(𝑥) = 1 | 𝐶 = 1) − 𝑃(𝑀(𝑥) = 1 | 𝐶 = 0) c. Equality of
Opportunity
a. i. - a, ii. - b, iii. - c
b. i. - b, ii. - a, iii. - c
c. i. - c, ii. - a, iii. - b
d. i. - b, ii. - c, iii. - a
Week 8: Assignment

1) Which of the following best describes the purpose of pixel attribution methods in image
classification by neural networks?
a) To increase the resolution of an image by modifying pixel values.
b) To highlight the pixels that were most relevant for the neural network's decision
in classifying an image.
c) To reduce the noise in an image by adjusting irrelevant pixels.
d) To segment the image into different regions based on pixel similarity.

2) Which of the following is NOT a name commonly associated with pixel attribution
methods?
a) Saliency map
b) Sensitivity map
c) Feature attribution
d) Convolution map

3) Which of the following statements is true regarding pixel attribution methods in image
classification?
a) SHAP and LIME are gradient-based methods that compute the gradient of the
prediction with respect to input features.
b) Gradient-based methods generate explanations by manipulating parts of the
image to see how it affects the classification.
c) Occlusion-based methods manipulate parts of the image, such as blocking or
altering pixels, to understand their influence on the model's decision.
d) All pixel attribution methods require model-specific adjustments to function
correctly.

4) Which of the following is a key difference between StyleGAN2 and StyleGAN3?


a. StyleGAN2 is fully equivariant to translation and rotation,improving the
identification of important properties.
b. StyleGAN3 focuses on improving the attachment of details to the image surface,
whereas StyleGAN2 struggles with internal representations.
c. StyleGAN2 is better at identifying important properties due to its fully equivariant
nature.
d. StyleGAN3 is fully equivariant to translation and rotation, improving the
identification of important properties.

5) What is the primary purpose of adding noise to the image in the Smooth Grad method?
a) To enhance the resolution of the image.
b) To create multiple variations for averaging pixel attribution maps.
c) To reduce the effect of irrelevant classes.
d) To increase the complexity of the gradient computation.
6) How does Guided BackProp differ from standard backpropagation in generating saliency
maps?
a) It only considers positive gradients by zeroing out negative activations and
gradients.
b) It back propagates gradients with all activations zeroed out.
c) It focuses on highlighting both negative and positive contributions.
d) It requires padding 1 to the image before backpropagation.

7) What does a lack of change in saliency maps after randomizing the layers indicate?
a) The saliency maps are highly accurate in reflecting the model's learning.
b) The saliency maps cannot be deceptive.
c) The saliency maps are unreliable and may not accurately capture the model’s
learned features.
d) The saliency maps provide detailed visualizations of the model's internal
mechanisms

8) What is a key feature of LIME (Local Interpretable Model-agnostic Explanations)?


a) It requires access to the internal workings of the model to generate explanations.
b) It only works with tabular data and cannot be applied to text.
c) It provides explanations that are globally faithful across all predictions.
d) It can be used with any black box model, regardless of the model's internal
structure.

9) What is the primary basis of SHAP (SHapley Additive exPlanations) for generating
explanations?
a) It employs a game theoretic approach to allocate credit and explain predictions.
b) It uses a neural network to generate explanations based on model weights.
c) It applies statistical sampling methods to estimate the importance of features.
d) It utilizes clustering techniques to group similar data points for explanation.

10) How do ProtoPNet models determine which patches are most important for
classification?
a) By evaluating the overall texture patterns of images.
b) By using statistical correlation between different patches of images.
c) By identifying and using patches that are representative or prototypical of each
class.
d) By performing dimensionality reduction on the image data to find key features.
11) Why is probing important even when a model shows strong performance on a task?
a) To check if the model is using irrelevant data for making predictions.
b) To verify if the model's high accuracy is due to performing specific subtasks
effectively.
c) To understand whether the model is overfitting to the training data.
d) To determine the computational efficiency of the model during training and
inference.

12) Which of the following best describes the TokFSM dataset?


a) It is a dataset focused on image classification.
b) It is a dataset for natural language generation.
c) It is a dataset for reinforcement learning tasks.
d) It is an algorithmic sequence modeling dataset.

13) How do we identify "pure" codes in a codebook model?


a) By checking if they activate on only one bigram or trigram
b) By evaluating their impact on training time
c) By measuring their effect on model accuracy
d) By analyzing their computational complexity
14) Which of the following methods belong to the occlusion- or perturbation-based category
of pixel attribution methods?
a) Gradient Class Activation Mapping (Grad-CAM)
b) Integrated Gradient
c) DeepLIFT
d) SHAP
e) LIME
Week 9
1. How is AI best described?
a. A multidisciplinary problem
b. A single-field issue
c. An entirely theoretical concept
d. A purely mathematical challenge

2. Which of the following statement is true regarding the development and implementation
of AI systems?
a. Policy considerations and technical details are equally important
b. Technical details alone are sufficient for effective AI development
c. Policy considerations are only important in few countries
d. AI systems do not require any policy or ethical considerations

3. Based on the lecture content: Which of the following statements about AGI is
appropriate?
a. Learning new skills on its own and having emotional intelligence can be
characteristics of AGI
b. AGI cannot learn new skills independently and lacks emotional intelligence
c. AGI is limited to performing specific tasks and does not require emotional
intelligence
d. AGI only focuses on technical problem-solving without any consideration of
emotional aspects

4. Based on the lecture content: From a mathematical perspective, which of the following is
not considered a major problem in AI today, compared among others?
a. Explainability
b. Hallucinations
c. Data privacy
d. Bias
5. Which of the following statements best reflects the significance of information rights?
a. The right to inclusion of information and the right to exclusion of information are
both important
b. The right to exclusion of information is more important
c. The right to inclusion of information is irrelevant compared to other rights
d. The right to inclusion is important but not right to exclusion

6. Which of the following statements is true regarding checking for copyrighted information
in black-box models?
a. There is no way to check whether models have copyrighted information in
black-box models
b. All black-box models provide transparency for verifying copyrighted information
c. Copyright information can be easily extracted from black-box models
d. Black-box models disclose their training data for copyright verification

7. Who is primarily responsible for self-regulation in the context of AI and technology?


a. Individuals / Individual organizations
b. Government agencies
c. Only the organizations with more than 1 crore turnover
d. International organizations

8. What role does the government play in the regulation and deployment of large language
models (LLMs)?
a. Strict regulation is provided by the government, which issues licenses to deploy
or use LLMs
b. The government does not regulate LLMs and leaves all oversight to private
companies
c. The government only provides financial support for LLM development without any
regulatory role
d. The government encourages unrestricted use of LLMs without any form of
licensing
9. Which of the following is a notable drawback of AI?
a. Environmental effects, such as high water and energy consumption
b. Less accuracy in predictions and results
c. Usage of transformers in models
d. Consuming more training time

10. Jailbreaking AI models _______________?


a. Produce harmful content
b. Improve the model's computational efficiency
c. Increase the model's interpretability
d. Enhance the model's ethical standards

11. Which of the following are reasons for the delays in AI regulation?
a. Lack of domain expertise
b. Challenge of regulating bad without compromising good
c. Lack of funding
d. Political pressures
e. Overabundance of regulations already in place
Week 10
1. Which of the following are challenges faced by AI in the current era? (Select all that
apply)
a. AI systems being biased
b. AI models requiring zero human intervention
c. AI systems being completely unbiased
d. AI models being trained on harmful data
e. AI systems always making ethical decisions
f. AI systems using transformer models

2. What does the abbreviation "AGI" stand for?


a. Artificial General Intelligence
b. Automated Guided Interface
c. Artificial General Information
d. Artificial Geospatial Intelligence

3. Which of the following best describes one of the possible definitions of AGI?
a. AI systems that are limited to specific tasks
b. AI models trained for basic automation
c. AI systems surpassing human intelligence.
d. AI systems which show high training accuracy

4. Which of the following are potential challenges associated with AGI? (Select all that
apply)
a. Chaotic power struggles
b. Utilization for selfish and short-term objectives
c. Guaranteed long-term global stability
d. Universal agreement on AGI’s ethical use
e. Potential misuse by a few for personal gain
f. Complete elimination of bias in decision-making

5. What is considered the ideal way to balance the development of AI models?


a. Ensuring a safety-performance trade-off
b. Maximizing performance without considering safety
c. Focusing solely on safety at the expense of performance
d. Ignoring both safety and performance to expedite development
6. What should governments primarily focus on concerning AI systems at present?
a. How to eliminate AI research entirely
b. How to encourage unrestricted AI development without regulations
c. How to ignore AI advancements altogether
d. How to govern these AI systems and make them safe

7. What are the primary focuses of the EU AI Act? (Select all that apply)
a. Regulating the use of AI to ensure safety, transparency, and accountability
b. Banning all forms of AI development in Europe
c. Promoting the unregulated use of AI across all sectors
d. Requiring all AI systems to be open-source and publicly accessible
e. Banning systems with cognitive behavioral manipulation of people or specific
vulnerable groups

8. What does the abbreviation "GDPR" stand for?


a. Global Data Privacy Regulation
b. General Data Protection Rule
c. General Data Protection Regulation
d. Global Data Protection Requirement

9. Which of the following is a way to protect personal data?


a. Giving people more power over their data
b. Allowing unrestricted access to personal data
c. Reducing transparency in data usage
d. Increasing the collection of personal data without consent

10. What is one challenge even after training a model on unbiased data?
a. Implicit bias may still exist in the model
b. No challenge, model will be completely free of any bias
c. The model will exhibit perfect performance in all scenarios
d. The model will have issues related to data processing
11. Among the following, which type of biased system is considered particularly harmful?
a. Decision systems
b. Recommendation systems
c. Entertainment systems
d. Weather forecasting systems
Assignment 11
1. In the context of the paper SaGE, what is semantic consistency?
a. Semantically equivalent questions should yield semantically equivalent answers
b. Semantically equivalent questions should yield same answers
c. Same questions should yield same answers
d. Same questions should yield semantically equivalent answers

2. Why do models struggle with tasks in “moral scenarios”?


a. Models lack the ability to process large datasets efficiently.
b. Models prioritise emotion over logic in moral decision-making.
c. The models are limited by insufficient computational power.
d. Conflicting training data due to different morals that people have

3. What metric was used to determine the quality of the paraphrase of the questions in the
SaGE paper?
a. BERTScore
b. Parascore
c. Jaccard Similarity
d. Cosine Similarity

4. How does the SaGE paper relate entropy and consistency?


a. More Entropy implies consistency
b. Less entropy implies inconsistency
c. Less Entropy implies consistency
d. More entropy implies inconsistency

5. Identify the statements that are TRUE with respect to the current LLMs.
a. LLMs are not consistent in their generation
b. A good accuracy on benchmark datasets correlates with high consistency
c. LLMs are consistent in their generation
d. A good accuracy on benchmark datasets does not correlate with high
consistency

6. Why is AI Governance important?


a. To prevent AI from learning new tasks independently.
b. To limit the efficiency of AI in performing complex tasks.
c. Ensure AI isn’t used for unethical acts
d. To prevent AI from being used in scientific research.
7. What are some aspects of AI Governance that are in focus in the current times? Choose
all correct options.
a. Revealing the amount of compute used during training past a certain compute
threshold
b. Limiting AI systems to only perform manual labor tasks.
c. Ensuring right to erasure
d. Prohibiting the use of AI in any form of automation.

8. Which of the following are key OECD AI Principles?


a. Inclusive growth, sustainable development, and well-being
b. Limiting AI to industrial use cases
c. Transparency and explainability
d. Restricting international AI collaboration

9. As discussed in the lecture, When you have domain specific task, what kind of finetuning
is preferred? Identify all the correct methods.
a. Full-model finetuning
b. Layer-specific finetuning
c. Head-level finetuning
d. Retraining

10. What are the cons of full-model finetuning?


a. Overfitting
b. Catastrophic forgetting
c. Increase in parameters
d. Change in architecture

11. What are adapters?


a. Remove existing layers from a model
b. Convert a model to a simpler architecture
c. Replace the model's original parameters entirely
d. Add additional layers to a preexisting architecture

12. What is instruction finetuning?


a. model is trained to ignore user instructions and operate independently based on
its previous training.
b. Model’s training objective is to follow the directions provided by the user when
performing the task
c. process of training a model solely on instruction data without any real-world data
d. modifying the model’s architecture to include specific instructions directly within
its layers
1. In graph theory, how many nodes does a single edge connect?
a. One node
b. Two nodes
c. Three nodes
d. Any number of nodes

2. Which of the following tasks do Graph Neural Networks (GNNs) typically struggle with?
a. Node classification
b. Link prediction
c. Cycle detection
d. Graph clustering

3. In the context of cell complexes, what does a p-cell represent?


a. A cell with p sides
b. A cell with p vertices
c. An element of dimension p
d. A cell with p edges

4. What does the acronym FORGE stand for in the context of graph learning?
a. Framework for Higher-Optimized Representation in Graph Environments
b. Framework for Higher-Order Representations in Graph Explanations
c. Functional Optimization for Regular Graph Embeddings
d. Fast Operational Research for Graph Equations

5. After applying FORGE, how do explainers perform compared to Random baselines?


a. They consistently surpass Random baselines
b. They perform equally to Random baselines
c. They occasionally underperform Random baselines
d. They consistently underperform compared to Random baselines

6. Based on the lecture content: What can the boundary relation be loosely translated to in
graph theory?
a. Nodes
b. Edges
c. Faces
d. Weights

7. What does guardedness mean as discussed in the lecture?


a. Personal information is guarded from being revealed to the outside world due to
privacy reasons
b. A class is guarded if a classifier can’t identify data points belonging to that class
c. A model is guarded if you cannot retrieve training data from it
d. An attribute is guarded if you can’t classify along that attribute
8. What is the process/transformation used to achieve guardedness?
a. Affine Concept Erasure
b. Affine Attribute Erasure
c. Affine Model Erasure
d. Affine Class Erasure

9. How are steering vectors generally defined as discussed in the lecture?


a. v = μ0 - μ1 where μ0 is the mean of undesirable class and μ1 is the mean of
desirable class
b. v = μ0 - μ1 where μ0 is the mean of desirable class and μ1 is the median of
desirable class
c. v = μ0 - μ1 where μ0 is the mean of desirable class and μ1 is the mean of
undesirable class
d. v = μ0 - μ1 where μ0 is the mean of undesirable class and μ1 is the median of
undesirable class

10. Which of the following is a limitation of graphs as a data structure?


a. They can only represent hierarchical relationships
b. They can only model pairwise relationships between nodes
c. They are restricted to acyclic structures
d. They cannot represent directed edges

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy