An AI-based computer system can gather data and use that data to make decisions or solve problems – using algorithms to perform tasks that, if done by a human, would be said to require intelligence. The benefits created by AI and machine learning (ML) systems for better health care, safer transportation, and greater efficiencies across the globe are already happening. But the increased amounts of data and computing power that enable sophisticated AI and ML models raise questions about the privacy impacts, ethical consequences, fairness, and real world harms if the systems are not designed and managed responsibly. FPF works with commercial, academic, and civil society supporters and partners to develop best practices for managing risk in AI and ML and assess whether historical data protection practices such as fairness, accountability, and transparency are sufficient to answer the ethical questions they raise.
Featured
India’s new Intermediary & Digital Media Rules: Expanding the Boundaries of Executive Power in Digital Regulation
The majority of these provisions were unanticipated, resulting in a raft of petitions filed in High Courts across the country challenging the validity of the various aspects of the Rules, including with regard to their constitutionality.
South Korea: The First Case Where the Personal Information Protection Act was Applied to an AI System
As AI regulation is being considered in the European Union, privacy commissioners and data protection authorities around the world are starting to apply existing comprehensive data protection laws against AI systems and how they process personal information. On April 28th, the South Korean Personal Information Protection Commission (PIPC) imposed sanctions and a fine of KRW […]
Takeaways from the Understanding Machine Learning Masterclass
Yesterday, the Future of Privacy Forum provided bespoke training on machine learning as a side event during the Computers, Privacy and Data Protection Conference (CPDP2020) in Brussels. The Understanding Machine Learning masterclass is a training aimed at policymakers, law scholars, social scientists and others who want to more deeply understand the data-driven technologies that are front of mind for data protection […]
Policy Brief: European Commission’s Strategy for AI, explained
The European Commission published a Communication on “Artificial Intelligence for Europe” on April 24th 2018. It highlights the transformative nature of AI technology for the world and it calls for the EU to lead the way in the approach of developing AI on a fundamental rights framework. AI for good and for all is the motto the Commission proposes. The Communication could be summed up as announcing concrete funding for research projects, clear social goals and more thinking about everything else.
Artificial Intelligence, Machine Learning, and Ethical Applications
On September 25, 2017, the Future of Privacy Forum and the Information Accountability Foundation will co-host an official side event at the International Conference of Data Protection Commissioners. The event follows IAF’s publication of Artificial Intelligence, Ethics and Enhanced Data Stewardship, and FPF’s curation of leading research highlighting the privacy challenges posed by artificial intelligence.
AI Ethics: The Privacy Challenge
The Future of Privacy Forum and the Brussels Privacy Hub of the Vrije Universiteit Brussel (VUB) are partnering with IEEE Security & Privacy in a call for papers focused on AI Ethics: The Privacy Challenge. Selected papers will be featured at The Brussels Privacy Symposium, an academic program jointly presented by the Brussels Privacy Hub of the VUB and FPF’s National Science Foundation supported Research Coordination Network.