Robeco logo

Disclaimer Robeco Switzerland Ltd.

The information contained on these pages is solely for marketing purposes.

Access to the funds is restricted to (i) Qualified Investors within the meaning of art. 10 para. 3 et sequ. of the Swiss Federal Act on Collective Investment Schemes (“CISA”), (ii) Institutional Investors within the meaning of art. 4 para. 3 and 4 of the Financial Services Act (“FinSA”) domiciled Switzerland and (iii) Professional Clients in accordance with Annex II of the Markets in Financial Instruments Directive II (“MiFID II”) domiciled in the European Union und European Economic Area with a license to distribute / promote financial instruments in such capacity or herewith requesting respective information on products and services in their capacity as Professional Clients.

The Funds are domiciled in Luxembourg and The Netherlands. ACOLIN Fund Services AG, postal address: Leutschenbachstrasse 50, CH-8050 Zürich, acts as the Swiss representative of the Fund(s). UBS Switzerland AG, Bahnhofstrasse 45, 8001 Zurich, postal address: Europastrasse 2, P.O. Box, CH-8152 Opfikon, acts as the Swiss paying agent.

The prospectus, the Key Investor Information Documents (KIIDs), the articles of association, the annual and semi-annual reports of the Fund(s) may be obtained, on simple request and free of charge, at the office of the Swiss representative ACOLIN Fund Services AG. The prospectuses are also available via the website https://www.robeco.com/ch.

Some funds about which information is shown on these pages may fall outside the scope of CISA and therefore do not (need to) have a license from or registration with the Swiss Financial Market Supervisory Authority (FINMA).

Some funds about which information is shown on this website may not be available in your domicile country. Please check the registration status in your respective domicile country. To view the Robeco Switzerland Ltd. products that are registered/available in your country, please go to the respective Fund Selector, which can be found on this website and select your country of domicile.

Neither information nor any opinion expressed on this website constitutes a solicitation, an offer or a recommendation to buy, sell or dispose of any investment, to engage in any other transaction or to provide any investment advice or service. An investment in a Robeco Switzerland Ltd. product should only be made after reading the related legal documents such as prospectuses, annual and semi-annual reports.

By clicking “I agree” you confirm that you/the company you represent falls under one of the above-mentioned categories of addressees and that you have read, understood and accept the terms of use for this website.

I Disagree

24-02-2023 · Insight

Mitigating the risks of Artificial Intelligence

Companies are becoming more aware of the risks that artificial intelligence (AI) can pose to society, but much still needs to be done. That was the outcome of three years of engagement by Robeco’s Active Ownership team with ten companies at the forefront of the technology.

    Authors

  • Daniëlle Essink-Zuiderwijk - Engagement Specialist

    Daniëlle Essink-Zuiderwijk

    Engagement Specialist

Summary

  1. Companies are developing responsible AI principles and ironing out risk

  2. Engagement successfully closed with four out of five companies

  3. Five other firms will be further engaged on their societal impact

AI can offer considerable advantages, from simple machine learning that seems to know what you meant to type, to more complex algorithms that can predict health care needs and detect patterns in climate change. It is now routinely used across the tech spectrum and often kicks in without the user even realizing that it is there.

However, it also poses significant threats to privacy, data management, and the prospect of machine ‘learning’ that leads to unwanted surveillance, racial profiling or discrimination. And it is hard to know the true state of affairs due to a lack of disclosure about companies’ AI activities.

This lack of information was one of the reasons why the Active Ownership team was only able to successfully conclude four out of five cases in the engagement program that ran from 2019 to 2022. The other five cases were transferred to the team’s SDG Engagement theme to further engage these companies on their societal impact.

Aligning practices

“Through our engagement, we learned that companies are gradually aligning internal practices to principles of responsible AI,” says engagement specialist Danielle Essink. “Many companies formalized AI principles that address topics like inclusion, fairness and transparency.”

“Additionally, companies are increasingly pursuing a collaborative approach by actively contributing to initiatives that aim to advance responsible governance and best practices. These types of initiatives play a decisive role in guaranteeing trustworthy AI across the industry.”

“However, ethical principles on their own do not ensure the responsible development and deployment of AI. Businesses require robust governance mechanisms to effectively implement their principles.”

Lack of disclosure

A major stumbling block is the lack of disclosure about what companies are actually doing to address concerns, along with their willingness to engage in the first place. Much of the AI technology and how it is implemented on different platforms is still shrouded in secrecy.

“In our engagement, we observed that transparency around governance and implementation remained low, as most companies’ public disclosures lacked clarity about how such principles translate into practice, and which checks and balances are in place,” says engagement specialist Claire Ahlborn, who also worked on the program.

“After talking to the companies, we learned about the specifics of the implementation, which then gave us the confidence to close some of the objectives successfully. The engagement results of this theme are, therefore, highly correlated with the company's willingness to set up constructive dialogues.”

As technology advances, so do the opportunities for quantitative investors. By incorporating more data and leveraging advanced modelling techniques, we can develop deeper insights and enhance decision-making.

Huge growth predicted

The International Data Corporation’s Worldwide Artificial Intelligence Software Forecast 2022 said the worldwide AI market is set to show compound annual growth of 18.6% from 2022 to 2026 alone.

Yet the potential benefits come with risks that are not yet fully explored, let alone understood, Essink says. To achieve the full potential of AI, companies need to manage the associated risks that come with the development and use of the technology, including human rights-related risks.

“Given the speed at which AI is being developed, there is no doubt that in the next few decades, this technology will transform our economy and society in ways we cannot imagine,” Essink says.

Positive changes

“This type of growth represents massive opportunities for AI to contribute to positive changes, such as detecting patterns in environmental data, or improving the analysis of health information.”

“At the same time, AI could cause new problems or aggravate existing ones if companies do not have enough understanding of the risks associated with these technologies. For example, using AI algorithms for profiling can have discriminatory effects, such as credit rating algorithms disfavoring people from certain ethnic backgrounds, or those living in certain areas.”

“Similarly, AI can be used for surveillance – in public spaces but also in the workplace – putting the right to privacy at risk. This shows a growing need for the responsible governance of AI systems to ensure that such systems conform to ethical values, norms, and the growing number of AI regulations.”

Upcoming regulation

Such regulatory moves and policy proposals have already been launched by governments, ethics committees, non-profit organizations, academics and the EU. In April 2021, the European Commission issued the AI Act. It sets out clear requirements and obligations regarding the specific uses of AI for developers, deployers and users.

The proposal identifies four regulatory categories based on the level of risk. At the high end, AI systems identified as high-risk, such as CV-scanning tools that rank job applicants, will be subject to strict obligations including enhanced risk management processes and human oversight. AI systems with limited risks will remain largely unregulated.

“This growing legislative pressure around AI could pose serious regulatory risks for companies that are not well prepared to conform with the rising obligations,” says Ahlborn.

Aligning engagement with the SDGs

“Meanwhile, the alignment of AI technologies with ethical values and principles will be critical to promote and protect human rights in society. As a result, we will continue our engagement work with a selection of companies in the tech sector under our ‘Sustainable Development Goals (SDG) engagement’ theme.”

“These dialogues have a strong focus on human rights and societal impact, and highlight topics like misinformation, content moderation and stakeholder collaboration. We will focus on how companies can contribute to SDG 10 (Reduced inequalities) and SDG 16 (Peace, justice and strong institutions) by safeguarding human rights in the development and use of AI and promoting social, economic and political inclusion.”

Read the full Q4 Active Ownership report here