top of page

PRI blog: What do investors need to know about AI risks?

A blog post about a workshop on responsible investing I led at the UN's Principles for Responsible Investment (PRI) organization. The blog post is co-authored with Daram Pandian. You can read the full text here.


➤ AI is booming.


Most companies already employ AI and investments in AI are on the rise. It is likely that the vast majority of companies will use AI in the coming years.


➤ AI presents many environmental and social (ESG) risks.


Examples include exacerbating social inequalities, disrupting democratic processes, and high carbon emissions. Responsible AI is a part of ESG.


➤ Regulation efforts are picking up steam, and some are already in effect.


➤ Here are three approaches to evaluate the environmental and social risks that a particular AI system/company poses:


1. By application type - assign coarse grain risk levels based on the application type following the risk classification presented in the EU AI Act.


2. Evaluate the company's responsible AI maturity - companies that develop and deploy AI responsibly are more likely to detect AI ethics problems and fix them.


3. Third-party evaluation - for mature companies, consider using third-party auditors with relevant technical, ethical, and legal expertise.


➤ You can find a link to the full blog post in the comments.


➤ Thank you to Peter Dunbar, CFA who invited me to lead this workshop, to

Eline Sleurink who helped organize, and to the workshop's participants for a wonderful discussion.

Recent Posts

See All
bottom of page