This commentary on Daniel Stader's recent article, “Algorithms Don't Have a Future: On the Relation of Judgement and Calculation” develops and complicates his argument by suggesting that algorithms ossify multiple kinds of prejudices, namely, the structural prejudices of the programmer and the exemplary prejudices of the dataset. This typology at once suggests that the goal of transparency may be impossible, but this impossibility enriches the possibilities for developing Stader's concept of reflected prejudiced use.