Will the European Parliament ban face surveillance technologies used by police and private companies in public spaces?
Face recognition in public spaces, as well as the categorisation of people’s emotions, gender and ethnicity, erodes freedom of assembly, enables discrimination, and sets a precedent — opening the door to future human rights violations
Members of the European Parliament are considering banning some of these harmful practices
This campaign aims to ensure that NOBODY can use face surveillance technologies in public spaces
Not in real-time, not later on using public footage. Not using images scraped from the internet. Never: forbidden forever
We will run a few experiments to let European decision-makers better comprehend the issues they have the power to halt
What if the AI Act fails to ban face recognition and categorisation in public? Let’s see! We have collected all the MEPs’ faces and analysed them with software currently available to anyone
Similarly to how the surveillance data market examines individuals’ faces for ’enrichment,’ you can explore the Open Data Faces 🤩, access the openly released data, and apply AI attribution models to gain insights
It’s neither reliable nor accurate, and it’s likely to lead to discrimination. But that has not stopped companies and governments from using these systems against us
To effectively simulate the unethical data market, we’ve also launched a contest. The winner will be the wildest, scariest, or most original data ‘enrichment’
🥇 | 🥈 | 🥉 |
---|---|---|
200€ | 100€ | Honorable mention |
DeepFakes are also a modern problem. We can’t avoid that they might be produced, but if legitimized, massive biometry system like pimeyes
or clearview
would incorporate the false information to the victim profile. With no control or awareness for the affected citizen
It summarize why there is a second contest, about producing the most surprising MEP DeepFake
🥇 | 🥈 | 🥉 |
---|---|---|
200€ | 100€ | Honorable mention |