From the course: CompTIA SecAI+ (CY0-001) Cert Prep
Unlock this course with a free trial
Join today to access over 25,500 courses taught by industry experts.
Model skewing
From the course: CompTIA SecAI+ (CY0-001) Cert Prep
Model skewing
Model skewing attacks occur when someone intentionally distorts an AI system's behavior over time by manipulating the data it consumes or the ways people use it. The attacker doesn't target the model's architecture or code. The attacker nudges the outputs in a specific direction. The change happens gradually and subtly and often avoids alerts. This attack works best in systems that update themselves with ongoing data such as feedback loops or reinforcement learning pipelines. the attacker doesn't need to poison the original training set. They feed the system enough biased, misleading, or carefully crafted malicious inputs so that the model's internal logic starts to drift. Consider a recommendation engine that learns from user clicks. An attacker repeatedly clicks on a specific type of content to push the system toward promoting a specific brand, ideology, or misinformation campaign. Over time, the model adapts to those signals and assumes that they reflect genuine interest. The…
Download courses and learn on the go
Watch courses on your mobile device without an internet connection. Download courses using your iOS or Android LinkedIn Learning app.
Contents
-
-
(Locked)
The AI lifecycle1m 39s
-
(Locked)
Business alignment in the AI lifecycle1m 43s
-
(Locked)
Data collection2m 20s
-
(Locked)
Data preparation3m 15s
-
(Locked)
Model development and selection2m 13s
-
(Locked)
Model evaluation and validation2m 29s
-
(Locked)
Model deployment and integration3m 25s
-
(Locked)
Monitoring and maintenance3m 19s
-
(Locked)
-
-
(Locked)
Manipulating application integrations4m 8s
-
(Locked)
AI supply chain attacks2m 4s
-
(Locked)
Insecure plug-in design2m 9s
-
(Locked)
Insecure output handling1m 23s
-
(Locked)
Output integrity attacks2m 8s
-
(Locked)
Model denial of service1m 31s
-
(Locked)
Excessive agency1m 33s
-
(Locked)
Overreliance1m 34s
-
(Locked)
AI hallucinations1m 4s
-
(Locked)
-
-
(Locked)
Monitoring prompts and responses2m 51s
-
(Locked)
Log monitoring4m 30s
-
(Locked)
Rate and cost monitoring5m 1s
-
(Locked)
Auditing for AI hallucinations3m 33s
-
(Locked)
Auditing for accuracy3m 29s
-
(Locked)
Auditing for bias and fairness4m 35s
-
(Locked)
Auditing access and security compliance3m 48s
-
(Locked)
-
-
(Locked)
Responsible AI5m 29s
-
(Locked)
AI risks2m 23s
-
(Locked)
Introduction of bias2m 37s
-
(Locked)
Accidental data leakage2m 53s
-
(Locked)
Reputational loss2m 11s
-
(Locked)
Accuracy and performance of the model2m 22s
-
(Locked)
Intellectual property risks3m 31s
-
(Locked)
Autonomous systems2m 27s
-
(Locked)
Shadow IT and shadow AI1m 48s
-
(Locked)
Awareness training2m 21s
-
(Locked)