Predicting the past: a philosophical critique of predictive analytics
Article Sidebar
Google Scholar citations
Main Article Content
If we address this topic from a conceptual and critical point of view, we need to address three issues: 1) why predictions are too often right, 2) why, at the same time, they are so often mistaken, and 3) what consequences arise from the fact that our instruments for prediction ignore at least four realities that must be true about future forecasts or at least be conscious of their limits: a) that individuals cannot be fully subsumed into categories, b) that their future behaviour tends to have unpredictable dimensions, c) that propensity is not the same as causality and d) that democratic societies must make the desire to anticipate the future compatible with respect for the open nature of the future.
Article Details
This work is licensed under a Creative Commons Attribution-NoDerivatives 3.0 Unported License.
(c) Daniel Innerarity, 2023
Copyright
Contents published in IDP are subject to a Creative Commons Attribution-No Derivative Works 3.0 Spain licence, the full text of which can be consulted on http://creativecommons.org/licenses/by-nd/3.0/es/deed.en.
Thus, they may be copied, distributed and broadcast provided that the author and IDP are cited, as shown in the recommended citation that accompanies each article. Derivative works are not permitted.
Authors are responsible for obtaining the necessary permission to use copyrighted images.
Assignment of intellectual property rights
The author non exclusively transfers the rights to use (reproduce, distribute, publicly broadcast or transform) and market the work, in full or part, to the journal’s editors in all present and future formats and modalities, in all languages, for the lifetime of the work and worldwide.
The author must declare that he is the original author of the work. The editors shall thus not be held responsible for any obligation or legal action that may derive from the work submitted in terms of violation of third parties’ rights, whether intellectual property, trade secret or any other right.
Daniel Innerarity, University of the Basque Country / Euskal Herriko Unibertsitatea
Professor of political philosophy, Ikerbasque researcher at the University of the Basque Country and Chair of Artificial Intelligence and Democracy at the European University Institute. Former fellow of the Alexander von Humboldt Foundation at the University of Munich, visiting professor at the University of Paris 1-Sorbonne, visiting fellow at the London School of Economics, and at Georgetown University. His recent books in English include Ethics of hospitality (2017), The democracy in Europe (2018), Politics in the Times of Indignation (2019) and A Theory of Complex Democracy (2023). He has been awarded with different prizes, recently the National Research Prize of Human Sciences.
ABEBE, R.; KASY, M. (2021). “The means of prediction”. In: Acemoglu, D. Redesigning AI. Work, democracy, and justice in the age of automation, pp. 87-91. Cambridge: Boston Review.
ACCOTO, C. (2019). Il mondo ex machina. Cinque brevi lezioni di filosofia dell’automazion. Milan: Egea.
ADAMS, V.; MURPHY, M.; CLARKE, A. (2009). “Anticipation: Technoscience, life, affect, temporality”. Subjectivity, vol. 28, n.º 1, pp. 246-265. DOI: https://doi.org/10.1057/sub.2009.18
AGRAWAL, A.; GANS, J.; Goldfarb, A. (2018). Prediction Machines. The Simple Economics of Artificial Intelligence. Cambridge: Harvard University Press.
AMOORE, L.; PIOTUKH, V. (2015). “Life beyond big data: governing with little analytics”. Economy and Society, vol. 44, n.º 3, pp. 314-366. DOI: https://doi.org/10.1080/03085147.2015.1043793
ANDREJEVIC, M. (2013). Infoglut: How Too Much Information Is Changing the Way We Think and Know. New York: Routledge. DOI: https://doi.org/10.4324/9780203075319
ANGWIN, J.; LARSON, J. (2016, December). “Bias in Criminal Risk Scores is Mathematically Inevitable, Researches Say”. ProPublica [online]. Available at: https://www.propublica.org/article/bias-in-criminal-risk-scores-is-mathematical-inevitable-researches-say
ARENDT, H. (2017). Mensch und Politik. Stuttgart: Reclam.
BOELLSTORF, T. (2013). “Making big data, in theory”. First Monday, vol. 18, no. 10. DOI: https://doi.org/10.5210/fm.v18i10.4869
BRAMAN, S. (2009). Change of State: Information, Policy and Power. Cambridge: The MIT Press.
BRAYNE, S. (2020). Predict and Surveil: Data, Discretion, and the Future of Policing. Oxford University Press. DOI: https://doi.org/10.1093/oso/9780190684099.001.0001
BROUSSARD, M. (2018). Artificial Unintelligence: How Computers Misunderstand the World. Cambridge: The MIT Press. DOI: https://doi.org/10.7551/mitpress/11022.001.0001
DERRIDA, J. (1994). “Nietzsche and the Machine”. Journal of Nietzsche Studies, no. 7, pp. 7-65.
ESPOSITO, E. (2011). The Future of Futures: The Time of Money in Financing and Society. Edward Elgar. DOI: https://doi.org/10.4337/9781849809115
ESPOSITO, E. (2021). Artificial Communication: How Algorithms Produce Social Intelligence. Cambridge: The MIT Press. DOI: https://doi.org/10.7551/mitpress/14189.001.0001
EUROPEAN COMMISSION (EC) (2015). Evidence-Based Better Regulation. European Commission [online]. Available at: https://commission.europa.eu/law/law-making-process/planning-and-proposing-law/better-regulation/better-regulation-guidelines-and-toolbox_en
FEDERAL TRADE COMMISSION (2016, January). Big Data. A tool of Inclusion or Exclusion? Understanding the issues. United Sates of America: Federal Trade Commission [online]. Available at: https://www.ftc.gov/system/files/documents/reports/big-data-tool-inclusion-or-exclusion-understanding-issues/160106big-data-rpt.pdf
FOERSTER, H. Von (2003). Understanding Understanding: Essays on Cybernetics and Cognition. New York: Springer. DOI: https://doi.org/10.1007/b97451
HILDEBRANDT, M. (2006). “Privacy and identity”. In: Claes, E., Duff, A., Gurtwith, S. (eds.). Privacy and the Criminal Law. Antwerpen/Oxford: Intersentia, pp. 43-57. DOI: https://doi.org/10.1007/s11572-006-9006-x
KAPOOR, S.; NARAYANAN, A. (2022). “Leakage and the Reproducibility Crisis in ML-based Science”. Patterns, vol. 4, no. 9. DOI: https://doi.org/10.1016/j.patter.2023.100804
MACKENZIE, A. (2015). “The production of prediction: What does machine learning want?”. European Journal of Cultural Studies, vol. 18, no. 4-5, pp. 429-445. DOI: https://doi.org/10.1177/1367549415577384
MASSUMI, B. (2007). “Potential politics and the primacy of preemption”. Theory & Event, vol. 10, no. 2. DOI: https://doi.org/10.1353/tae.2007.0066
MATZNER, T. (2018). “Grasping the ethics and politics of algorithms”. In: Sætnan, A. R., Schneider, I., Green, N. (2018). The Politics of Big Data. Big Data, Big Brother, pp. 30-45. Oxford, New York: Routledge.
MAYER-SCHOENBERGER, V.; CUKIER, K. (2013). Big Data. A Revolution That Will Transform How We Live, Work, and Think. New York: Houghton.
MERTON, R. (1948). “The self-fulfilling prophecy”. The Antioch Review, vol. 8, no. 2, pp. 193-210. DOI: https://doi.org/10.2307/4609267
NOWOTNY, H. (2021). In AI we trust. Power, illusion and control of predictive algorithms. Cambridge: Polity Press.
PORTER, T. M. (1995). Trust in Numbers. The Pursuit of Objectivity in Science and Public Life. Princeton University Press. DOI: https://doi.org/10.1515/9780691210544
SCHNEIDER, I. (2018). “Bringing the state back in. Big Data-based capitalism, disruption, and novel regulatory approaches in Europe”. In: Sætnan, A. R., Schneider, I., Green, N. The Politics of Big Data. Big Data, Big Brother, pp. 129-175. Oxford, New York: Routledge.
STRAUSS, S. (2015). “Datafication and the Seductive Power of Uncertainty-A Critical Exploration of Big Data”. Information, no. 6, pp. 836-847. DOI: https://doi.org/10.3390/info6040836
TYLER, I. (2015). “Classificatory Struggles: Class, Culture and Inequality in Neoliberal Times”. The Sociological Review, vol. 63, no. 2, pp. 493-511. DOI: https://doi.org/10.1111/1467-954X.12296