Adversarial ML
Last updated
Last updated
Bias in ML is when a model favors or discriminates against certain groups or outcomes due to flaws in the training data. Think of it like a teacher grading students unfairly because they have preconceived notions (e.g., “students from School X always get low scores”). In adversarial ML, attackers exploit or create this unfairness to harm the model’s credibility or manipulate its predictions.