-
Notifications
You must be signed in to change notification settings - Fork 54
Description
Following up from comment thread in PR #140
A few thoughts below, if we think this is worth adding. E1 is related but this feels like a sufficiently different question to consider, especially with everything that's already present in E1.
Possible wording:
E.5 Systematic effects
(could also be, say E.3 and push the other two to 4/5)
Have we considered risks posed by the scale, speed, or rigidity of the deployed model that aren't present in the equivalent human or prototype process (e.g., reinforced outcomes and feedback loops, ability to consider missing variables, societal impacts)?
Possible examples:
I think this example of feedback loop (formerly of "concept drift") actually fits in here; this is a result of using the model, not just a distribution that happens to shift on its own
- text: Sending police officers to areas of high predicted crime skews future training data collection as police are repeatedly sent back to the same neighborhoods regardless of the true crime rate.
url: https://www.smithsonianmag.com/innovation/artificial-intelligence-is-now-used-predict-crime-is-it-biased-180968337/
- text: -- Related academic study.
url: https://arxiv.org/abs/1706.09847
curious if folks have thoughts on other examples to represent broader systemic impacts that might not be captured elsewhere in the checklist.
possible example: speed of misinformation spread + twitter (Science article, Twitter's crisis misinformation policy to slow down viral tweets, etc.)