ESRs Fatma Dogan and Soumia El Mestari at Privacy Symposium 2023

ESRs Fatma Dogan and Soumia El Mestari attended the annual Privacy Symposium took place in Venice between 17-21 April. ESRs presented their paper which they have witten with Dr. Maria Botes from Centre for Medical Ethics and Law of Stellenbosch University.

Their paper titled “Technical and Legal Aspects Relating to the (Re-Use of Health Data when Repurposing Machine Learning Models in the EU” will be published in the conference book. In the study, they discussed technologies such as re-purposing machine learning models under EU data protection laws. The original point was that technologies like Al and loT are data greedy by their nature hence they need more datasets constanty. Machine learning technologies can reuse the existing machine learning models, also known as “knowledge transfer for other tasks”. However, this solution when examined under a legal lens becomes ambiguous because there is no exact equivalent of this term in current EU data protection laws. Even so, when this technology is used in health data tasks, the practice becomes even more complex. because health data qualifies as sensitive data which attracts stricter rules regarding its processing.

In the paper, they examined this topic from both a legal and a technical point of view. Their research considers the use of repurposing machine learning models and their application within the legal context of the secondary use and re-use of personal data. Their legal analysis includes the General Data Protection Regulation, Data Governance Act, and European Health Data Space proposal. The ESRs received questions and supportive feedback from the participants and also attended the
other enriching panels of the conference.

ESRs Fatma Dogan and Barbara Lazarotto at BILETA 2023

 

On April 13th 2023, ESRs Fatma Dogan and Barbara Lazarotto participated in the 38th Annual BILETA – British and Irish Law Education and Technology Association Conference. The Conference came back to The Netherlands after 22 years and had as a topic Cyberlaw: Finally getting its Act(s) together? held in Amsterdam Law and Technology Institute and online. (for more information consult the program of the conference). Fatma and Bárbara both presented their research at the Data Governance Panel, exploring the interaction between different EU Regulations.

Barbara presented her research titled The Right to Data Portability: An Holistic Analysis of GDPR, DMA and the Data Actin which she explored the interactions and overlaps when it comes to the right to data portability in the three Regulations and possible implications to data subjects’ rights. Barbara won the European Journal of Law and Technology best paper award with her paper.

  

Fatma presented her research titled Re-Use or Secondary Use: A Comparison between Data Governance Act and European Health Data Space, where she explored the re-use or secondary use of health data, making an analysis of the Data Governance Act and the European Health Data Space.

 

Both researchers received great questions and feedback from the participants which will enrich their research for future purposes.

Participation of ESR Aizhan Abdrassulova in two conferences in April 2023

Early Stage Researcher Aizhan Abdrassulova presented her research at III International Forum on Medical Law, which was held on April 6-7, 2023 in Ekaterinburg.  Her article’s title is «Some Aspects of medical ethics and Confidentiality in European Law». Aizhan outlined the basic principles of European medical ethics and its main directions of development. Participation in the conference was remote, materials will be published on the website and in print in June 2023.

Also on April 11-13, Aizhan was present offline at the 33rd Madrid International Conference on “Law, Education, Marketing and Management” (LEMM-23).

Aizhan’s topic “Data ownership: civil law aspect” is closely related to the topic of her dissertation. All papers of LEMM-23 will be published in the printed conference proceedings with a valid International ISBN number.

Each Paper will be assigned a unique Digital Object Identifier(DOI) from CROSSREF and the Proceedings of the Conference will be archived in DiRPUB’s Digital Library. Aizhan won the nomination for “Oral Best Paper Certificate”.

Clustering Workshop on Ethical and Legal Issues in Technology

The VALKYRIES project funded from the European Union´s Horizon 2020 research and innovation programme, conducted its third ethical and legal workshop titled “Dealing With Ethical Legal Issues Technology Development: The Current Approaches Adopted In Ongoing Projects” on March 3, 2023.

The workshop consisted of presentations by various EU projects such as VALKYRIES, 5GSOSIA, RESCUER, LeADS and FACILITATE, these presentations focused on the ethical and legal issues encountered and combatted during the course of their respective projects.

On behalf of the LeADS project, Prof. G. Comandé introduced the legality attentive data scientists project and discussed how the LeADS Project is aimed at bridging the gap between data science and law by training early-stage researchers who are involved in the project with technical prowess as well as legal expertise. Further, he mentioned that this has been made possible by the various partnerships with companies and government agencies that the LeADS Project has carved out for its early-stage researchers to create a balance between theoretical studies and practical implementations across sectors.

Prof. Paul De Hert and ESR Barbara Lazarotto at the Norface Governance Online Lecture Series

On 01 March 2023, Prof. Paul de Hert and ESR Barbara Lazarotto participated in the third NORFACE GOVERNANCE Online Lecture Series at the University of Luxembourg. They were joined by Prof. Jean-Bernard Auby, Emeritus Professor of Public Law, Sciences Po Paris – Former Director of “Mutations de l’Action Publique et du Droit Public” («Changes in Governance and Public Law»), Former Deputy Director of the Oxford Institute of European and Comparative Law. The online lecture series was a part of the dissemination practices for the governance program.

The lecture was titled Should the state be smart?“, and approached the concept of “smartness” as a buzzword that is sold by private companies to governments, which often adopt technology to micromanage individuals’ lives, confusing it with administrative effectiveness. The lecture explored possible alternative paradigms for governments for the future and their relationship with technology.

 

 

Going Beyond the Law or Back Again?

This blogpost was originally published on the Robotics & AI Law Society’s Blog at https://blog.ai-laws.org/going-beyond-the-law-or-back-again/

Abstract:

The relationship between law and ethics in the European Union’s regulatory policy of artificial intelligence remains unclear. Ambiguity persists specifically in whether ethics will be employed in a hard or soft fashion. Such an ambiguity creates fundamental problems in determining whether a given AI system is compliant with the proposed Artificial Intelligence Act (AIA), which is contrary to the objective of legal certainty cited by the AIA. This blog post is an attempt to uncover that problem which academic literature has not yet addressed.

The AIA is a product of the European strategy towards regulating AI which has developed over the past four years and has always embodied a peculiar relationship between law and ethics. In 2018, the European Commission released their Communication  outlining the beginnings of a European strategy for AI. Around that time, the High-Level Experts Group on AI (HLEG) was established with the goal of advising the Commission, and their 2019 Ethics Guidelines for Trustworthy AI , which took the fundamentals rights of the Charter to be the subject matter of both their ethical and legal approach, became central to the EU strategy. Then in 2020, the official White Paper on Artificial Intelligence was released, stating that the European approach “aims to promote Europe’s innovation capacity in the area of AI while supporting the development and uptake of ethical and trustworthy AI across the EU economy.” Later in 2020, the European Parliament formally requested the Commission to submit a proposal for a regulation specifically addressing ethical principles for the development, deployment, and use of artificial intelligence. And finally, in 2021, the Artificial Intelligence Act was proposed, and the Commission asserted that the Regulation ensures the protection of ethical principles.

The Distinction Between Hard and Soft Ethics

During those years, many AI ethics researchers had been concerned with the development of principles and struggled with their practical application to the problems of AI, while some others had taken the problem of application to be an irredeemable flaw with the principle-based approach and so called for a virtue ethics or question-based approach to be developed instead. That debate will continue to evolve and is not the direct subject matter of this post, nor are the ongoing debates about how best to operationalize ethics in SMEs or larger corporations. Instead, this blog is an effort to uncover the ambiguous relationship between law and ethics in the European Union’s regulatory policy of AI. That ambiguity can be traced back to a misapplication of Luciano Floridi’s “soft ethics” presented here, which asserted that governance should be composed of institutional governance, regulation, and ethics. He and others argued that ethics should play a role in governance because legal requirements are often necessary but not always sufficient to guide society toward a beneficial outcome, and that ethics can act as an anticipatory mechanism to identify and respond to the impacts of emerging technologies. According to his theory, digital ethics is constituted by hard ethics, which changes or creates law through the implementation of ethical values in legislation or judicial opinions; and soft ethics which considers what should be done “over and above the existing regulation, not against it, or despite its scope, or to change it, or to by-pass it.” In other words, while soft and hard ethics cover the same normative ground, the distinction between the two rests on their relationship to law.

Following that distinction between hard and soft ethics, compliance with a given law is equivalent to the compliance with the hard ethic that gave it shape, and thus the analysis used to determine whether some behavior is compliant is the legal method an attorney would use in a given jurisdiction. But, because ethics that are soft go “over and above” the law, a different methodology would presumably be required to determine whether some behavior is in alignment with a given set of values. When reading the forthcoming paragraphs keep this in mind: the relationship between hard ethics and law is characteristic of how scholars typically understand the cyclical nature of ethical norms being translated into regulation, similar to how the EU data protection acquis developed from shifts in the perceptions about and values of privacy in the public or how medical professional ethics developed in response to bioethics. The reader’s reaction then may be to assume that the exact same is happening in the EU regulatory approach to AI, that ethics is simply influencing regulators to act, and that they are embedding the values into the regulation (hard ethics). However, the policy outlined below suggests that compliance with the AIA may require a separate ethical analysis that takes the law as the starting point and goes beyond. It’s important to note that Floridi’s soft ethics approach was originally conceived as a “post compliance” tool (not a legal requirement), yet through the adoption of soft ethics in the regulatory policy, this blog will suggest that it may have (intentionally or unintentionally) taken on a new and confused life as a compliance mechanism.

The Cause of the Ambiguity: A Sketch of the EU Regulatory Development on AI

The consensus in the European Union has largely been that AI should meet both a legal and ethical requirement. In the HLEG’s Ethics Guidelines for Trustworthy AI, they state that AI should be legal, ethical, and technically and socially robust. The HLEG derived their ethical principles from the EU Charter of Fundamental Rights. Distinguishing between the legal and ethical requirements, the HLEG explained that after legal compliance is achieved, “ethical reflection can help us understand how the development, deployment and use of AI systems may implicate fundamental rights and their underlying values and can help provide more fine-grained guidance when seeking to identify what we should do rather than what we (currently) can do with technology.” The HLEG cited Floridi’s soft ethics approach, asserting that there “adherence to ethical principles goes beyond [emphasis added] legal compliance.” Yet, the question of which direction one should go beyond fundamental rights jurisprudence remains and is striking.

The demarcation between legal and ethical requirements carried on into the European Parliament’s Resolution C 404 , which distinguished between “legal obligations” and “ethical principles,” advising that high-risk AI systems be subject to the mandatory compliance of both. Yet, they also observed that “ethical principles are only efficient where they are also enshrined in law. . .” Note that if ethics is only efficient where it is enshrined in the law and compliance is mandatory with ethical principles by law, then the distinction between “legal obligations” and “ethical principles” collapses (hard ethics). However, the Parliament went on to call for common criteria to be developed for the granting of European certificates of ethical compliance, which suggests, again, that an ethical analysis (as opposed to a legal one) may need to be developed to satisfy ethical principles. Finally, the Parliament formally requested the Commission to submit a proposal for a regulation specifically addressing ethical principles for the development, deployment, and use of artificial intelligence.

The proposed AIA responded to that request and the Commission asserted that the Regulation ensures the protection of ethical principles. The general objective of the legislation was to “ensure the proper functioning of the single market by creating the conditions for the development and use of trustworthy artificial intelligence [emphasis added] in the Union.” The proposed minimum requirements for high-risk AI systems set out in Chapter II are based on HLEG’s work on Trustworthy AI. In the heart of the AIA, references to ethical principles are non-existent but references to fundamental rights abound. Remember that the ethical principles of Trustworthy AI are derived from fundamental rights.

Questions Going Forward

Thus, the question remains: does compliance of high-risk AI systems with Chapter II of the AIA require a separate ethical analysis, based on fundamental rights, that goes beyond legal compliance with fundamental rights law? Does the mandatory compliance mean that this creates a new philosophy for the interpretation of fundamental rights? And as inquired upon earlier, which direction does one go beyond to achieve compliance? Fundamental rights law must often strike a balance between rights, and the calculations in those decisions create the most controversial of legal judgements. Going beyond, especially obligatorily, has wide reaching consequences, and would likely be highly politically charged. Clarification and information for whether and how to go beyond must be given by policymakers.

What such an analysis might look like remains unclear. As mentioned at the top, AI ethics researchers are still debating how to best develop and apply principles or whether to even do so at all. The field is far from establishing consensus. Yet, underpinning AI ethics is the idea that law can benefit from ethics for various reasons. Yet, the opposite might also be true: that an ethical analysis or method could be modeled off the legal structure that allows for the application of normative propositions to specific sets of facts (perhaps something akin to Hart’s rule of recognition, change, and adjudication as a starting point).

LeADS Liaisons with the OpenMuse Consortium

The LIDER Lab at Sant’Anna School of Advanced Studies recently joined a consortium of music industry stakeholders and researchers from 12 European countries. The OpenMuse Consortium’s goal is to fill harmful data gaps in the European music ecosystem and provide artists with much needed data-driven tools to increase their success in the streaming environment.

ESR Robert Poe and post-doctoral researcher Pelin Turan recently represented Sant’Anna at the Kick-off meeting for this exciting new European project in Bratislava, Slovakia. Robert’s background in EU artificial intelligence law and AI ethics and Pelin’s background in cultural diversity and EU copyright law proved invaluable during the many coordination meetings that took place there. Sant’Anna will head up three deliverables for the Consortium including the development of a data management plan and policy recommendations involving fair machine learning techniques and topics in cultural diversity; research areas which present opportunities for the Legality Attentive Data Scientists to offer insightful contributions in the future. Sant’Anna is looking forward to a bright and productive collaboration between the two projects where there is sufficient overlap!

13th International Workshop on Socio-Technical Aspects in Security

LeADS Beneficiary University of Luxembourg is organising the 13th International Workshop on Socio-Technical Aspects in Security which will take place on 7 July 2023 as a hybrid event in Delft, The Netherlands. ESR Xengie Doan and Prof. Gabriele Lenzini are on the organising team of the event.

The objective of the workshop is to stimulate an exchange of ideas on how to design systems that are secure in the world where they interact with users of varying lived experiences and diverse needs. Bringing together experts working in various areas of computer security as well as in social and behavioral sciences.

The Call for Papers accept original papers in several formats:

  • Full Papers discussing original research, answering well-defined research questions, and presenting full and stable results;
  • Work in Progress describing original but unfinished piece of work, based on solid research questions or hypotheses.
  • Position Papers discussing existing challenges and introducing and motivating new research problems;

For more information check STAST 2023 website!

Brief presentation of research of Xengie Doan

Brief presentation of research of Xengie Doan (ESR9)

March 2nd 2023 at 1 p.m.

Room 1, Palazzo Toscanelli, Sant’Anna School of Advanced Studies (Pisa)

For further info

ESR Soumia Zohra El Mestari at 5th AAAI/ACM conference on Artificial Intelligence , Ethics and Society (AIES)

Early-Stage Researcher Soumia Zohra El Mestari attended the fifth AAAI/ACM conference on Artificial Intelligence, Ethics and Society (AIES) hosted at Keble college, university of Oxford in the United Kingdom from August 1st to 3rd 2022.

Soumia applied and was selected to attend the doctoral programme that was organized by the conference. As part of the student programme, Soumia participated in various workshops along with PhD students from different disciplines namely: Ethics, Law, Philosophy, and computer science.

The participants took part in many group working sessions along with mentors who are distinguished scientists in their fields. The workshops were followed by lunch roundtables with mentors.

During the last day of the conference, Soumia presented her PhD work around the usage of Privacy Enhancing Technologies(PETs) in machine learning pipelines as a tool to achieve the EU vision of trustworthy AI systems.

 

 

She first began by discussing the various conceptions of privacy in the context of data driven solutions, the different threat points throughout the machine learning pipeline and how these threat points can be exploited to maintain strong privacy attacks that can not only lead to heavy leakage bills but also may be invisible to detect. Then, the presentation also discussed the various PETs used to mitigate those attacks along with the pros and cons of each mitigation strategy.

Finally, Soumia concluded her presentation by stressing on the fact that the effective identification of the threat points can be a key element to solve some of the privacy issues. However, solving some privacy issues may come at the cost of other trustworthiness elements such as fairness and transparency. Hence, building a trustworthy AI pipeline is a challenging task since the acceptable tradeoff between those trustworthiness elements remains hard to measure.