Facebook and TUM create joint AI ethics research centre
The social network will contribute $7.5 million to the centre over a period of five years
Facebook has teamed up with the Technical University of Munich (TUM) to create an independent research centre focused on the study of AI ethics.
The Institute for Ethics in Artificial Intelligence will draw on the expertise of thought leaders and academics to research potential ethical issues related to the use of AI - such as safety, privacy, fairness, and transparency - as well as identifying possible with new use cases.
Facebook will contribute $7.5 million over five years and offer insight into how it's using AI and algorithms in initiatives such as its Fairness Flow that can determine unintended bias. Although TUM also plans to consider other funding sources, too.
"At Facebook, ensuring the responsible and thoughtful use of AI is foundational to everything we do from the data labels we use, to the individual algorithms we build, to the systems they are a part of," Joaquin Quionero Candela, director of Applied Machine Learning at Facebook, wrote in a post announcing the partnership.
"AI poses complex problems which industry alone cannot answer, and the independent academic contributions of the Institute will play a crucial role in furthering ethical research on these topics... The Institute will also benefit from Germany's position at the forefront of the conversation surrounding ethical frameworks for AI - including the creation of government-led ethical guidelines on autonomous driving - and its work with European institutions on these issues."
The Institute for Ethics in Artificial Intelligence will be led by Professor Dr. Christoph Ltge.
"At the TUM Institute for Ethics in Artificial Intelligence, we will explore the ethical issues of AI and develop ethical guidelines for the responsible use of the technology in society and the economy," Dr. Ltge said.
"Our evidence-based research will address issues that lie at the interface of technology and human values. Core questions arise around trust, privacy, fairness or inclusion, for example, when people leave data traces on the internet or receive certain information by way of algorithms.
"We will also deal with transparency and accountability, for example in medical treatment scenarios, or with rights and autonomy in human decision-making in situations of human-AI interaction."
BCDR buyer's guide for MSPs
How to choose a business continuity and disaster recovery solutionDownload now
The definitive guide to IT security
Protecting your MSP and your customersDownload now
Cost of a data breach report 2020
Find out what factors help mitigate breach costsDownload now
The complete guide to changing your phone system provider
Optimise your phone system for better business resultsDownload now