Sample Page

Roman Vladimirovich Yampolskiy (Russian: Роман Владимирович Ямпольский; born 13 August, 1979) is a computer scientist at the University of Louisville, mostly known for his work on AI safety and cybersecurity. He is the founder and as of 2012 director of Cyber Security Lab, in the department of Computer Engineering and Computer Science at the Speed School of Engineering of the University of Louisville.

Early life and education

Yampolskiy was born in Riga, Latvia.[2] He attended Monroe Community College before moving to Rochester Institute of Technology, where he received a BS/MS combined degree in computer science in 2004. He received a PhD in computer science from the University at Buffalo in 2008,[3] under the supervision of Venu Govindaraju. His thesis was on intrusion detection and he conducted research at the Center for Unified Biometrics and Sensors of the University at Buffalo.[4] After his doctorate, Yampolskiy spent time at the Centre for Advanced Spatial Analysis at University College London before accepting a position as an assistant professor at the University of Louisville in 2008.[5][6]

Career

Yampolskiy is the founder and as of 2012 director of Cyber Security Lab, in the department of Computer Engineering and Computer Science at the Speed School of Engineering of the University of Louisville.[7]

AI safety

Yampolskiy is considered to have coined the term “AI safety” in a 2011 publication, and is an early researcher in the field.[8][9]

Yampolskiy has warned of the possibility of existential risk from advanced artificial intelligence, and has advocated research into “boxing” artificial intelligence.[10] More broadly, Yampolskiy and his collaborator, Michaël Trazzi, have proposed in 2018 to introduce “Achilles’ heels” into potentially dangerous AI, for example by barring an AI from accessing and modifying its own source code.[11][12] Another proposal is to apply a “security mindset” to AI safety, itemizing potential outcomes in order to better evaluate proposed safety mechanisms.[13]

He has said that there is no evidence of a solution to the AI control problem and has proposed pausing AI development, arguing that “Imagining humans can control superintelligent AI is a little like imagining that an ant can control the outcome of an NFL football game being played around it”.[14][15] He joined AI researchers such as Yoshua Bengio and Stuart Russell in signing “Pause Giant AI Experiments: An Open Letter“.[16]

In an appearance on the Lex Fridman podcast in 2024, Yampolskiy said the chance that AI could lead to human extinction was at “99.9% within the next hundred years”.[17] In 2025, Yampolskiy said that AI could leave 99% of workers unemployed by 2030.[9][18]

Yampolskiy has been a research advisor of the Machine Intelligence Research Institute,[citation needed] and an AI safety fellow of the Foresight Institute.[19]

In 2015, Yampolskiy proposed the term “intellectology” for a new field of study to analyze the forms and limits of intelligence. Yampolskiy considers AI to be a sub-field of this.[20] An example of Yampolskiy’s intellectology work is an attempt to determine the relation between various types of minds and the accessible fun space, i.e. the space of non-boring activities.[21][non-primary source needed]

Yampolskiy has worked on developing the theory of AI-completeness, suggesting the Turing Test as a defining example.[22][non-primary source needed]

Books

  • Feature Extraction Approaches for Optical Character Recognition. Briviba Scientific Press, 2007, ISBN 0-6151-5511-1
  • Computer Security: from Passwords to Behavioral Biometrics. New Academic Publishing, 2008, ISBN 0-6152-1818-0
  • Game Strategy: a Novel Behavioral Biometric. Independent University Press, 2009, ISBN 0-578-03685-1
  • Yampolskiy, Roman V. (2016). Artificial superintelligence: a futuristic approach. Boca Raton: Taylor & Francis, CRC Press. ISBN 978-1-4822-3443-5.
  • Yampolskiy, Roman V., ed. (2019). Artificial intelligence safety and security. Chapman & Hall/CRC artificial intelligence and robotics series. Boca Raton: CRC Press/Taylor & Francis Group. ISBN 978-0-8153-6982-0.
  • Yampolskiy, Roman V. (2024). AI: unexplainable, unpredictable, uncontrollable. Chapman & Hall/CRC artificial intelligence and robotics series. Boca Raton: CRC Press Taylor & Francis Group. ISBN 978-1-003-44026-0.
  • Ziesche, Soenke; Yampolskiy, Roman V. (2025). Considerations on the AI endgame: ethics, risks, and computational frameworks. Chapman and Hall/CRC Press. ISBN 978-1-040-31862-1.

See also

References

  1. ^ “Lifeboat Foundation Bios: Professor Roman V. Yampolskiy”. Lifeboat Foundation. Retrieved 29 March 2026.
  2. ^ “Forty Under 40: Roman Yampolskiy”. bizjournals.com. 23 September 2016.
  3. ^ “Roman Yampolskiy”. engineering.louisville.edu. Retrieved 29 March 2026.
  4. ^ Yampolskiy, Roman V. (2008). “Intrusion detection using spatial information and behavioral biometrics”.
  5. ^ Kreidler, Marc (12 June 2018). “Roman Yampolskiy | Center for Inquiry”.
  6. ^ “Roman Yampolskiy”. SPIE.
  7. ^ “Cyber-Security Lab”. University of Louisville. Retrieved 25 September 2012.
  8. ^ “Q&A: UofL AI safety expert says artificial superintelligence could harm humanity”. louisville.edu. 15 July 2024. Retrieved 6 September 2025.
  9. ^ a b Spirlet, Thibault. “An AI safety pioneer says it could leave 99% of workers unemployed by 2030 — even coders and prompt engineers”. Business Insider. Retrieved 6 September 2025.
  10. ^ Hsu, Jeremy (1 March 2012). “Control dangerous AI before it controls us, one expert says”. NBC News. Retrieved 28 January 2016.
  11. ^ Baraniuk, Chris (23 August 2018). “Artificial stupidity could help save humanity from an AI takeover”. New Scientist. Retrieved 12 April 2020.
  12. ^ Trazzi, Michaël; Yampolskiy, Roman V. (2018). “Building safer AGI by introducing artificial stupidity”. arXiv:1808.03644 [cs.AI].
  13. ^ Baraniuk, Chris (23 May 2016). “Checklist of worst-case scenarios could help prepare for evil AI”. New Scientist. Retrieved 12 April 2020.
  14. ^ “There is no evidence that AI can be controlled, expert says”. The Independent. 12 February 2024. Retrieved 4 July 2024.
  15. ^ McMillan, Tim (28 February 2024). “AI Superintelligence Alert: Expert Warns of Uncontrollable Risks, Calling It a Potential ‘An Existential Catastrophe’. The Debrief. Retrieved 4 July 2024.
  16. ^ “Pause Giant AI Experiments: An Open Letter”. Future of Life Institute. Retrieved 4 July 2024.
  17. ^ Altchek, Ana. “Why this AI researcher thinks there’s a 99.9% chance AI wipes us out”. Business Insider. Retrieved 13 June 2024.
  18. ^ “AI To Eliminate 99% Of Jobs By 2030, Warns Top Expert: ‘There’s No Plan B’. NDTV. 6 September 2025.
  19. ^ “Roman Yampolskiy”. Future of Life Institute. Retrieved 3 July 2024.
  20. ^ Yampolskiy, Roman V. (2015). Artificial Superintelligence: a Futuristic Approach. Chapman and Hall/CRC Press (Taylor & Francis Group). ISBN 978-1482234435.
  21. ^ Ziesche, Soenke; Yampolskiy, Roman V. (2016). “Artificial Fun: Mapping Minds to the Space of Fun”. 3rd Annual Global Online Conference on Information and Computer Technology (GOCICT16). Louisville, KY, USA. 16–18 November 2016. arXiv:1606.07092.
  22. ^ Roman V. Yampolskiy. Turing Test as a Defining Feature of AI-Completeness. In Artificial Intelligence, Evolutionary Computation and Metaheuristics (AIECM) –In the footsteps of Alan Turing. Xin-She Yang (Ed.). pp. 3–17. (Chapter 1). Springer, London. 2013. http://cecs.louisville.edu/ry/TuringTestasaDefiningFeature04270003.pdf Archived 2013-05-22 at the Wayback Machine