Democratizing AI Governance: Balancing Expertise and Public Participation

post by Lucile Ter-Minassian (lucile-ter-minassian) · 2025-01-21T18:29:06.160Z · LW · GW · 0 comments

Contents

  1. The Case for Democratic AI Governance: A Stakeholder-Based Analysis
  1. A) Universal Stakeholdership and Access
  1. B) Democratic Legitimacy in an Era of Declining Trust
  1. C) Limitations of Expert-Centric Governance
  1. D) Economic and Civil Liberty Implications
  2. The Scientific Imperative: Why AI Governance Requires Expert Leadership
  2. A) Scientific Method and Democratic Process: Fundamental Incompatibilities
  2. B) Misinformation and Public Discourse
  2. C) Temporal Constraints and Decision-Making Efficiency
  3. Towards Balanced AI Governance: Integrating Democratic Participation with Expert Knowledge
  3. A) Accessibility of AI Governance
  3. B) Democratic Models in Practice: the French Citizens' Convention on Climate (CCC)
    Process and Methodology
    Challenges of the CCC
  3. C) Democratic Models in Practice: the Brazilian AI Framework
  3. D) Lessons for AI governance
    Inclusive Representation and Diversity
    Balancing Expertise and Public Engagement
    Challenges of Deliberative Processes
    Transparency, Accountability, and Iterative Processes
    Toward a Nuanced Hybrid Framework
  4. Framework Recommendations: Implementing Democratic AI Governance in the European Union
  4. A) Aligning Democratic Governance Models with EU AI Initiatives
  4.B) Temporal Considerations and Adaptive Governance
    Conclusion
    References
None
No comments

Technological innovation has historically emerged from concentrated centers of expertise, developed by a limited group of specialists with minimal public oversight or legal frameworks (e.g. printing press, nuclear weapons). The development of artificial intelligence—defined here as machine learning models trained on data that can be deployed for real-world applications, encompassing both supervised and unsupervised approaches, whether predictive or generative—follows this pattern, with development concentrated among a small number of private companies and research institutions [1].

However, AI systems are increasingly deployed in domains that affect fundamental aspects of human life, from healthcare and education to employment and criminal justice [2, 3, 4]. This broad societal impact, combined with AI's potential to reshape power structures and decision-making processes, raises critical questions about its governance. A stakeholder-based analysis suggests that those affected by AI systems should have a voice in shaping both their development and deployment.

Yet this proposition faces significant challenges. The technical complexity of AI systems creates a substantial knowledge gap between experts and the general public. This disconnect raises fundamental questions about the feasibility of democratic participation in AI governance: How can meaningful public input be incorporated into decisions that require deep technical understanding? What mechanisms could bridge the divide between scientific expertise and democratic legitimacy?

Democratic governance, defined here as decision-making processes that ensure representation, accountability, and participation from affected stakeholders, presents both opportunities and challenges in the context of AI. This analysis will examine competing schools of thought: first exploring arguments for democratic governance, then considering the case for expert-led oversight. Ultimately, we will interrogate the possibility of integrating direct democratic participation into AI governance frameworks. Drawing on case studies from France and Brazil, we propose recommendations for a balanced governance model tailored to European Union countries.

Key words: AI governance, expert-led governance, participatory democracy, deliberative democracy

 

 

1. The Case for Democratic AI Governance: A Stakeholder-Based Analysis

 

1. A) Universal Stakeholdership and Access

The democratization of AI technology represents a significant departure from historical patterns of technological innovation. Unlike previous transformative technologies such as nuclear weapons or early printing presses, which were characterized by centralized control and limited access, AI exhibits unprecedented openness through open-source initiatives.

This openness manifests across three distinct dimensions. First, access to AI applications has become widespread, with tools ranging from chatbots to image generators freely available to any internet user. Second, AI development appears accessible; while anyone with basic computational resources can fine-tune language models using open-source frameworks, the inherent complexity and opacity of AI systems preserve the traditional knowledge barrier between experts and the general public [6]. Third, AI's influence extends across critical domains, including healthcare, insurance, and media consumption, creating a unique context where every citizen is both an active and passive stakeholder in AI's deployment [2, 3, 4].

This three-tiered reality—universal access, technically-constrained development, and universal impact—creates a unique governance challenge. Despite widespread use and impact, decision-making processes remain concentrated among tech companies and policymakers, with limited public participation. The European Union's AI Act negotiations exemplify this democratic deficit, where civil society, despite representing the interests of affected populations, remains largely excluded from crucial policy formations [5]. This creates a concerning misalignment between AI's universal stakeholdership and its concentrated governance structures.

 

1. B) Democratic Legitimacy in an Era of Declining Trust

Contemporary democratic institutions face a crisis of legitimacy, evidenced by declining public trust across OECD nations. Empirical data indicates a roughly 20% decrease in confidence in governmental bodies over two decades, with Eurobarometer surveys showing trust in national governments falling below 35% in numerous EU member states [7]. Within this context, democratic AI governance becomes crucial for two reasons:

1. Legitimacy and Accountability: Public participation in AI governance ensures policy alignment with collective interests rather than narrow corporate or technical objectives. Arras and Braun [8] experimentally show that inclusive consultation processes can enhance regulatory legitimacy.

2. Transparency and Trust: AI's inherent opacity and pervasive influence necessitate democratic oversight to prevent public skepticism and resistance. O'Neil et al. demonstrate how opaque algorithmic systems can perpetuate inequality when developed without adequate public scrutiny [9].

 

1. C) Limitations of Expert-Centric Governance

Expert-only governance models present several critical limitations:

1. Representation Deficits: Exclusive technical governance risks overlooking broader ethical and social implications, potentially creating policies misaligned with public priorities. Zuboff's analysis of early internet governance illustrates how expert-centric approaches failed to anticipate crucial societal challenges like misinformation and privacy concerns [10].

2. Interdisciplinary Necessity: To effectively address the complex societal implications of AI, it is essential to adopt an interdisciplinary approach that bridges technical expertise with ethical and policy considerations. Coeckelbergh's concept of "democracy deficit" highlights how technocratic approaches often neglect the inherently political nature of AI-related decisions. [11] Cave et al. emphasize the importance of integrating perspectives from computer science, ethics, and policy to address AI's multifaceted challenges. [12]

 

1. D) Economic and Civil Liberty Implications

Since AI serves as a key driver of economic growth and a powerful tool with societal implications, democratic governance, characterized by its emphasis on inclusivity and collective decision-making, may offer several advantages for addressing economic and civil liberty concerns. First, it may promote economic equity by ensuring a more equitable distribution of AI-driven benefits across society, particularly supporting small and medium-sized enterprises [8]. Second, democratic frameworks may enhance rights protection by enabling public participation, which could help balance security requirements with civil liberties such as privacy and freedom of expression [13]. Finally, they may mitigate conflicts of interest by reducing the disproportionate influence of entities that primarily profit from AI development and deployment, fostering a more inclusive and balanced governance process.

 

 

2. The Scientific Imperative: Why AI Governance Requires Expert Leadership

 

2. A) Scientific Method and Democratic Process: Fundamental Incompatibilities

Klein's analysis in Le goût du vrai presents a fundamental critique of democratic approaches to scientific governance. [15] He articulates three critical tensions between scientific advancement and democratic processes:

  1. Epistemological Foundation: Scientific truth emerges through empirical validation and rigorous methodology, not through consensus or popular agreement. This epistemological framework stands in direct opposition to democratic principles of collective decision-making.
  2. Expertise Hierarchy: While democratic systems presuppose equality of voice, scientific progress necessitates the privileging of expert knowledge. This hierarchical structure of expertise, as elaborated by Collins and Evans in Rethinking Expertise [14], is essential for maintaining scientific rigor and preventing the dangerous flattening of specialized knowledge.
  3. Methodological Misalignment: Public interpretation of scientific processes often fails to grasp the necessity of uncertainty and debate within scientific methodology, potentially undermining crucial aspects of scientific advancement.

The COVID-19 pandemic response provides a compelling empirical example: epidemiological expertise necessarily superseded public opinion in implementing lockdown measures, demonstrating how scientific imperatives can override democratic processes in crisis situations.

 

2. B) Misinformation and Public Discourse

The proliferation of AI misinformation in public discourse presents a significant challenge to democratic governance models. Cave's analysis in Bridging near- and long-term concerns about AI [12] identifies two primary distortions that shape public understanding. On one hand, exaggerated threat narratives amplify potential risks beyond their empirical basis, fostering undue fear and misunderstanding. Conversely, popular media often promotes unrealistic optimism, presenting oversimplified or overly positive views of AI capabilities. Together, these distortions hinder informed decision-making and complicate efforts to address the societal implications of AI through democratic means. These distortions in public understanding can lead to policy decisions that fail to address actual technological challenges and risks. Furthermore, due to their inclusive nature, democratic processes may also be susceptible to populist pressures, potentially compromising essential strategic and ethical considerations in AI development.

 

2. C) Temporal Constraints and Decision-Making Efficiency

The rapid pace of AI development creates temporal pressures that conflict with democratic deliberation. The speed of technological advancement consistently outpaces democratic decision-making processes, leading to a persistent innovation-governance gap where regulatory responses lag behind new developments. Additionally, high-stakes decisions in AI development often demand immediate, expert-informed responses, which democratic processes are not always equipped to efficiently facilitate. These temporal and procedural constraints suggest that expert-led governance frameworks may be better suited to addressing the immediate challenges of AI development and deployment.

 

3. Towards Balanced AI Governance: Integrating Democratic Participation with Expert Knowledge

While technical expertise remains crucial, integrating democratic participation in AI governance can help ensure that policy decisions are aligned with societal values, fostering legitimacy and public trust in the outcomes. In this section, we explore two key models of democratic initiatives: participatory and deliberative democracy. Through the examination of two case studies—the French Citizens’ Convention on Climate (CCC) and the Brazilian AI Framework—we analyze how these approaches operate in practice. We conclude by drawing lessons from these examples and applying them to the context of AI governance, emphasizing how democratic engagement can enhance inclusivity, transparency, and accountability in addressing complex, technical challenges.

In the remainder of the paper, we will use the notions of participatory democracy and deliberative democracy. Participatory democracy emphasizes broad public engagement and direct decision-making, often through methods like voting, surveys, or public comments, aiming for maximum inclusion and citizen power. In contrast, deliberative democracy focuses on informed debate and consensus-building through structured dialogue, typically in smaller groups facilitated by trained moderators. While participatory democracy prioritizes the breadth of participation, deliberative democracy emphasizes depth, requiring more time and interaction with technical experts to achieve well-reasoned, collective decisions. These approaches differ in their goals: participatory democracy seeks widespread involvement, while deliberative democracy aims for thoughtful, consensus-driven outcomes.

 

3. A) Accessibility of AI Governance

The governance of AI systems often presents fewer technical barriers than commonly assumed. While some aspects undeniably require deep technical expertise, many critical governance decisions focus on universally relatable societal impacts. Dryzek [17] advocates for participatory and decentralized governance, combining scientific expertise with public

values and social considerations to address complex challenges, such as those posed by environmental issues. Similarly, Pasternak et al. [18] demonstrate how teaching core principles of complex scientific topics can empower policymakers to make informed decisions. They detail their experience in designing and delivering courses aimed at equipping policymakers to understand and apply scientific evidence effectively. Drawing from these perspectives, we argue that effective AI governance demands not only an understanding of core technical concepts, such as AI training methodologies, robustness, and explainability principles, but also a deep awareness of their societal implications for privacy, labor markets, and ethical considerations.

 

3. B) Democratic Models in Practice: the French Citizens' Convention on Climate (CCC)

Participatory democracy emphasizes broad public consultation and direct decision-making mechanisms, while deliberative democracy prioritizes informed debate through structured expert-citizen dialogue. The historical precedent of the Athenian Boule from 508-507 BCE provides an instructive model: a council of 500 citizens, chosen by lot annually, engaged with complex technical and administrative matters while consulting experts. The Boule met daily in specialized committees, demonstrating how ordinary citizens could effectively participate in technical governance when provided with appropriate structural support.

The French Citizens' Convention on Climate of 2019 [19, 21] (CCC) exemplifies modern application of these principles. The CCC was a democratic experiment that aimed to give citizens a direct voice in shaping climate policy. It brought together 150 citizens selected through sortition to represent the diversity of French society across six demographic dimensions: gender, age, socio-economic background, education level, location type, and province. Participants were tasked with defining measures to reduce greenhouse gas emissions by at least 40% by 2030, compared to 1990 levels, while considering social justice [19].

 

Process and Methodology

  1. Thematic Working Groups: The assembly was divided into five work groups covering transport, food, consumption, work and production, and housing.
  2. Expert Support: The convention was assisted by a governance committee, legal and technical experts, and professionals in citizen participation and deliberation.
  3. Deliberative Approach: Citizens learned about climate issues, debated, and prepared draft laws over seven sessions spanning several months.
  4. Public engagement through online contributions: Citizens could submit suggestions through an online platform. Over 1,000 contributions were synthesized and provided to the convention participants, though time constraints limited their thorough consideration. [22]

All participants received compensation as well as loss of professional income when relevant. Through their deliberations, participants produced 149 distinct policy proposals, showcasing citizens' capacity to meaningfully engage with complex issues and generate concrete solutions. The initiative gained particular significance when the President committed to submitting these proposals "without a filter" through various democratic channels, including referendum, parliamentary vote, or direct implementation. However, ultimately most proposals weren’t implemented in their original form. The government introduced the "Climate and Resilience Law" in 2021, which drew from the CCC’s recommendations but watered down several key proposals. [23] Beyond its immediate policy implications, the Convention fostered increased public awareness and engagement with climate issues, potentially catalyzing greater societal acceptance and support for climate action measures.

Challenges of the CCC

The Citizens' Convention on Climate faced challenges balancing deliberative independence and external engagement, as explained by Gaborit et. al [22]. Internally, tensions arose over expert influence, as participants sometimes perceived experts as steering discussions, particularly when their advice conflicted with citizen priorities. Externally, public contributions were synthesized but only partially utilized due to time constraints. Finally, activist groups lobbied participants near the convention venue, raising overlooked issues but risking undue influence on deliberations. Media interactions amplified public awareness but placed additional pressure on participants to act as both deliberators and public representatives, complicating the focus on internal decision-making. These dynamics highlight the difficulty of maintaining a clear boundary between the assembly's internal deliberations and its broader societal and institutional interactions. [22]

 

3. C) Democratic Models in Practice: the Brazilian AI Framework

The Brazilian AI Framework [20, 25, 26] aims to establish a comprehensive regulatory structure for the development, implementation, and use of artificial intelligence systems in Brazil, focusing on protecting fundamental rights, promoting responsible innovation, and ensuring safe and reliable AI systems for the benefit of society.

The framework was the result of a lengthy development, characterized by extensive deliberation, public engagement, and collaboration. Initially, it consolidated three separate bills (5.051/2019, 21/2020, and 872/2021). In March 2022, a commission was established to draft a new unified law, working over 240 days through meetings, seminars, and public hearings.

The Senate Temporary Committee on Artificial Intelligence (CTIA) played a key role, incorporating 85 amendments into the bill after extensive debates and multiple deadline extensions. Public engagement was central, involving hearings and contributions from stakeholders, such as recommendations from the Centre for Information Policy Leadership (CIPL).

Key features of the framework include:

●  Mandatory algorithmic impact assessments for high-risk AI systems developed by the companies. This ensures continuous evaluation of risks and benefits to fundamental rights. [26]

●  Public consultations for public sector AI: The Brazilian National Data Protection Authority (ANPD) initiated a public consultation to collect a broad range of contributions for shaping AI guidelines and regulations. This effort aimed to incorporate diverse perspectives from both technical experts and civil society, reflecting the democratic aspirations of the framework. While the consultation highlighted openness, the details of how contributions were synthesized remain limited. [25]

●  Regular dialogue between citizens and authorities: The creation of the National System for Artificial Intelligence Regulation and Governance (SIA) provides a structural mechanism for ongoing engagement. This system ensures that regulatory approaches to AI remain adaptive and aligned with societal needs, fostering trust and accountability. It signifies a commitment to maintaining citizen involvement as AI technologies evolve. [25]

●  Extensive public consultation during development: The process included public hearings, seminars, and stakeholder input. However, the depth and impact of these consultations on the final legislation are not fully transparent, raising questions about whether public feedback directly influenced the bill's provisions.

The bill ultimately passed with widespread support on December 10, 2024. [25]

The Brazilian AI Framework stands as a landmark in integrating citizen participation into AI governance, reflecting a balance between stakeholder input and legislative action. However, despite its significant public engagement, the absence of a citizen panel limited its scope as a fully participatory democratic initiative.

 

3. D) Lessons for AI governance

Lessons from the French Citizens' Convention on Climate (CCC) and the Brazilian AI Framework provide valuable insights for creating inclusive, effective AI governance. They show how structured democratic engagement can address complex, technical challenges while also exposing the inherent limits of such approaches.

Inclusive Representation and Diversity

Ensuring inclusive representation is crucial. The CCC selected participants by sortition to reflect France’s diversity in terms of gender, age, socio-economic background, and geography, ensuring varied perspectives in policy recommendations—an approach AI governance should emulate to capture diverse societal concerns. The Brazilian AI Framework, though it involved public consultation, lacked a mechanism for direct citizen representation. This underscores the importance of actively involving citizens beyond public hearings to empower their voices.

Balancing Expertise and Public Engagement

Effective governance must bridge knowledge gaps between experts and citizens. The CCC used thematic working groups and expert support so citizens could engage with complex issues like climate policy. Likewise, AI governance can benefit from structured deliberation, allowing citizens and AI experts to collaboratively address topics such as bias, transparency, and accountability. The Brazilian AI Framework faced time constraints and ambiguous feedback channels, highlighting the need for processes that collect and visibly integrate public input.

Challenges of Deliberative Processes

The CCC encountered difficulties balancing independence and external engagement, including perceived expert influence, lobbying, and media pressure. These challenges point to the need for boundaries and safeguards to protect citizen panels’ integrity. In contrast, the Brazilian AI Framework lacked a dedicated citizen panel, reducing such tensions but also limiting the depth of deliberation. Its reliance on public consultations and expert input made it more expert-driven, illustrating trade-offs between efficiency and democratic participation.

Transparency, Accountability, and Iterative Processes

Transparency and accountability are vital for building trust in governance. The CCC showed this through live-streamed sessions and published documentation, and AI governance should follow suit by providing public access to deliberations and clear accountability measures. The Brazilian AI Framework’s National System for Artificial Intelligence Regulation and Governance (SIA) facilitates ongoing dialogue between citizens and authorities, supporting iterative processes, continuous input, and periodic reviews of governance strategies as societal needs evolve.

Toward a Nuanced Hybrid Framework

These lessons suggest that democratic participation in AI governance is both possible and beneficial when structured carefully. The key is designing hybrid frameworks that blend expert knowledge with meaningful public input while managing tensions between technical complexity and democratic engagement.

Challenges from both the CCC and the Brazilian AI Framework show the need for intentional design to manage external pressures, incorporate public feedback, and balance deliberative independence with transparency. The next section offers specific recommendations for such hybrid frameworks in the European Union, building on existing structures to foster inclusive, transparent, and technically credible AI governance.

 

 

4. Framework Recommendations: Implementing Democratic AI Governance in the European Union

 

In this section, we provide recommendations for a comprehensive AI governance framework tailored for EU member states, aligning with the broader provisions of the EU AI Act. While not exhaustive, these recommendations aim to guide the development of governance models that integrate deliberative and participatory elements. The proposed framework is structured around key pillars: representative selection, adaptive governance mechanisms, and sustained public engagement. Together, these pillars offer a pathway for fostering inclusive, transparent, and effective governance of AI systems.

 

4. A) Aligning Democratic Governance Models with EU AI Initiatives

Participatory and deliberative approaches to AI governance at the country level within the EU may provide a robust mechanism for aligning national efforts with overarching EU frameworks, including the measures proposed in the First Draft of the General-Purpose AI Code of Practice [16]. These approaches may reinforce the principles of transparency, accountability, and inclusivity while ensuring the public's voice informs the governance of AI systems.

The First Draft of the General-Purpose AI Code of Practice [16] explicitly highlights the importance of involving civil society, academia, and other non-expert groups to address systemic risks posed by general-purpose AI. For instance, Measure 10.1 calls for the integration of participatory methods in evidence collection. EU member states could explore national initiatives such as citizen assemblies or deliberative panels to contribute to this evidence base, potentially offering diverse perspectives on AI risks and societal impacts.

Additionally, the Code’s emphasis on transparent documentation (Measures 1 and 2) may provide opportunities for participatory approaches to flourish. National governments might consider creating publicly accessible platforms for sharing technical documentation, including risk assessments and acceptable use policies (AUPs), as mandated by the EU AI Act. These platforms could include mechanisms for public feedback and input, allowing citizens to participate directly in the scrutiny and refinement of AI policies.

Incorporating public oversight measures outlined in the Code, such as serious incident reporting (Measure 18) and whistleblowing protections (Measure 19), could align with deliberative practices. Member states might consider establishing independent oversight bodies with citizen representation to review and respond to reported incidents. Similarly, strengthening protections for whistleblowers might enable individuals to safely highlight risks or ethical concerns, fostering trust in the governance process.

Finally, the post-deployment monitoring framework (Measure 11) proposed in the Code [16] could encourage sustained public engagement. Member states might implement participatory mechanisms for ongoing monitoring, such as community-led reviews or participatory audits of AI systems, to ensure that governance efforts remain adaptive to societal needs and technological evolution.

In conclusion, participatory and deliberative initiatives at the national level may not only align with the EU AI Act and the General-Purpose AI Code of Practice but also have the potential to enhance the inclusivity and effectiveness of AI governance. These approaches may help ensure that governance models reflect the diverse perspectives of EU citizens while adhering to the principles enshrined in the EU's broader AI regulatory framework.

 

4.B) Temporal Considerations and Adaptive Governance

A stepwise approach may reconcile rapid AI innovation with democratic deliberation. Drawing on Cowan’s [24] analysis of technological governance, the framework could initially rely on stronger expert oversight, then gradually integrate broader public input as societal understanding deepens. This transition might coincide with milestones like consensus on model explainability or baseline ethical standards.

To address AI’s rapid evolution potentially outpacing traditional policy-making, adaptive mechanisms are crucial. Iterative regulations, mixed expert-public advisory panels, and agile “sandboxes” allow incremental policy testing and refinement, aligning with technological advances and societal needs. Continuous public engagement can reflect shifting societal values and emerging concerns. Involving diverse stakeholders in ongoing discussions helps policies evolve alongside AI’s progress while staying aligned with public interests, bridging the gap between rapid technological change and slower policy processes.

In summary, adaptive and temporal considerations are pivotal for a governance framework that evolves over time. Gradual implementation, responsive regulations, and iterative engagement can help manage AI’s rapid development while preserving democratic principles.

 

Conclusion

This analysis explores the tension between democratic participation and expert governance in AI development. While scientific rigor demands specialized expertise, AI’s deep societal impact calls for public engagement. Real-world examples show that technical complexity does not have to exclude meaningful public input; rather than choosing between expertise and democracy, effective AI governance hinges on their careful integration.

The proposed EU framework demonstrates how structured deliberation and phased public participation can address time-sensitive AI challenges without compromising democratic legitimacy. By uniting random selection, expert guidance, and open engagement channels, it balances the urgency of AI innovation with the need for public trust. As AI continues to transform society, governance must preserve both technical excellence and collective values. Looking ahead, pilot initiatives and ongoing feedback loops can refine these democratic mechanisms, ensuring that AI governance remains inclusive, transparent, and adaptive. The future of AI governance lies in synthesizing expertise and public voice—an approach that strengthens accountability and positions AI as a truly shared endeavor.

 

 

 

References

[1] Jácome de Moura Jr., P., dos Santos Junior, C. D., Porto-Bellini, C. G., & Lima Dias Junior, J. J. (2024). The over-concentration of innovation and firm-specific knowledge in the artificial intelligence industry. Journal of the Knowledge Economy. Advance online publication. https://doi.org/10.1007/s13132-024-01974-1

[2] West, D. M., & Allen, J. R. (2018). How artificial intelligence is transforming the world. Brookings Institution.

[3] Obermeyer, Z., Powers, B., Vogeli, C., & Mullainathan, S. (2019). Dissecting racial bias in an algorithm used to manage the health of populations. Science.

[4] Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias: There's software used across the country to predict future criminals. ProPublica.

[5] Veale, M., & Borgesius, F. Z. (2021). Demystifying the Draft EU Artificial Intelligence Act—Analysing the good, the bad, and the unclear elements of the proposed approach. Computer Law Review International, 22(4), 97-112.

[6] Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., ... & Herrera, F. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82-115.

[7] OECD (2024), OECD Survey on Drivers of Trust in Public Institutions – 2024 Results: Building Trust in a Complex Policy Environment, OECD Publishing, Paris.

[8] Arras, S., & Braun, D. (2020). Stakeholder consultations and the legitimacy of regulatory decision-making: A survey experiment. Regulation & Governance, 14(3), 435-456.

[9] O'Neil, C. (2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown Publishing Group.

[10] Zuboff, S. (2019). The Age of Surveillance Capitalism: The Fight for a Human Future at the New Frontier of Power. PublicAffairs.

[11] Coeckelbergh, M. (2020). AI Ethics. MIT Press.

[12] Cave, S., Whittlestone, J., Nyrup, R., Ó hÉigeartaigh, S., & Calvo, R. A. (2019). Bridging near- and long-term concerns about AI. Nature Machine Intelligence, 1(1), 5-8.

[13] Cheong, I. Y., et al. (2024). Particip-AI: A Democratic Surveying Framework for Anticipating AI Risks and Benefits.

[14] Collins, H., & Evans, R. (2007). Rethinking Expertise. University of Chicago Press. [15] Klein, É. (2020). Le goût du vrai. Éditions Gallimard.

[16] First Draft General-Purpose AI Code of Practice. European Commission, November 2024. Available at: https://ec.europa.eu/futurium/en/ai-general-purpose

[17] Dryzek, J. S. (2019). The Politics of the Earth: Environmental Discourses (4th ed.). Oxford University Press.

[18] Pasternak, N., Almeida, P., Seixas, R. J., & Fonseca, E. M. (2024). Teaching scientific evidence and critical thinking for policy making. PLoS ONE, 19(4), e0293819.

[19] Pech, T., & Tubiana, L. (2019). Citizens' Convention on Climate: Governance and Procedure. Economic, Social and Environmental Council (ESEC), France.

[20] Zanatta, R. A. F., & Rielli, M. (2024). The Artificial Intelligence Legislation in Brazil: Technical Analysis of the Text to Be Voted on in the Federal Senate Plenary.

[21] Eymard, L., Fabre, A., & Strauss, S. (2022). Lessons from the French Citizens' Convention for Climate. HAL ENPC, hal-03119539v29.

[22] Gaborit, M., Jeanpierre, L. et Rozencwajg, R. (2022) . Les frontières négociées des assemblées citoyennes. Le cas de la Convention citoyenne pour le climat (2019-2020) Participations, N° 34(3), 173-204. https://doi.org/10.3917/parti.034.0173.

[23] KNOCA. (n.d.). French Citizens' Convention on the Climate (La Convention Citoyenne pour Le Climat). Knowledge Network on Climate Assemblies. Retrieved January 2, 2025, from https://www.knoca.eu/national-assemblies/french-citizens-convention-on-the-climate

[24] Cowan, R. S. (1997). A Social History of American Technology. Oxford University Press.

[25] Digital Policy Alert. (2024, November 6). Opened consultation on ANPD artificial intelligence and data protection regulatory project. Retrieved January 2nd 2025 from https://digitalpolicyalert.org/event/24272-opened-consultation-on-anpd-artificial-intelligence-a nd-data-protection-regulatory-project

[26] Data Privacy Brasil. (n.d.). The artificial intelligence legislation in Brazil: Technical analysis of the text to be voted on in the Federal Senate plenary. From January 2nd 2025 https://www.dataprivacybr.org/en/the-artificial-intelligence-legislation-in-brazil-technical-analy sis-of-the-text-to-be-voted-on-in-the-federal-senate-plenary/

0 comments

Comments sorted by top scores.